Publications

Literature
Artificial Intelligence and new technologies regulation
Mullingan D. K.; Bamberger K. A. (2019)
Procurement As Policy: Administrative Process for Machine Learning
At every level of government, officials contract for technical systems that employ machine learning—systems that perform tasks without using explicit instructions, relying on patterns and inference instead. These systems frequently displace discretion previously exercised by policymakers or individual front-end government employees with an opaque logic that bears no resemblance to the reasoning processes of agency personnel. However, because agencies acquire these systems through government procurement processes, they and the public have little input into—or even knowledge about—their design or how well that design aligns with public goals and values. This Article explains the ways that the decisions about goals, values, risk, and certainty, along with the elimination of case-by-case discretion, inherent in machine-learning system design create policies—not just once when they are designed, but over time as they adapt and change. When the adoption of these systems is governed by procurement, the policies they embed receive little or no agency or outside expertise beyond that provided by the vendor. Design decisions are left to private third-party developers. There is no public participation, no reasoned deliberation, and no factual record, which abdicates Government responsibility for policymaking. This Article then argues for a move from a procurement mindset to policymaking mindset. When policy decisions are made through system design, processes suitable for substantive administrative determinations should be used: processes that foster deliberation reflecting both technocratic demands for reason and rationality informed by expertise, and democratic demands for public participation and political accountability. Specifically, the Article proposes administrative law as the framework to guide the adoption of machine learning governance, describing specific ways that the policy choices embedded in machine-learning system design fail the prohibition against arbitrary and capricious agency actions absent a reasoned decision-making process that both enlists the expertise necessary for reasoned deliberation about, and justification for, such choices, and makes visible the political choices being made. Finally, this Article sketches models for machine-learning adoption processes that satisfy the prohibition against arbitrary and capricious agency actions. It explores processes by which agencies might garner technical expertise and overcome problems of system opacity, satisfying administrative law’s technocratic demand for reasoned expert deliberation. It further proposes both institutional and engineering design solutions to the challenge of policymaking opacity, offering process paradigms to ensure the “political visibility” required for public input and political oversight. In doing so, it also proposes the importance of using “contestable design”—design that exposes value-laden features and parameters and provides for iterative human involvement in system evolution and deployment. Together, these institutional and design approaches further both administrative law’s technocratic and democratic mandates.
Documents
Competition advocacy
Autorité de la concurrence and Bundeskartellamt (2019)
Algorithms and Competition
Algorithms are among the most important technological drivers of the ongoing digitalization process. They are becoming more and more important, enabling firms to be more innovative and efficient. However, debate has arisen on whether and to what extent algorithms might also have detrimental effects on the competitive functioning of markets. In their joint conceptual project – Algorithms and Competition – the Autorité de la concurrence and the Bundeskartellamt studied potential competitive risks that might be associated with algorithms. They elaborated on the concept of algorithm as well as on different types and fields of application. In their study, the two authorities focused in particular on pricing algorithms and collusion, but also considered potential interdependencies between algorithms and the market power of the companies using them as well as practical challenges when investigating algorithms. Isabelle de Silva, President of the Autorité de la concurrence: “Algorithms are used constantly in the digital economy, and are at the very core of how some fast growing businesses operate: online travel agencies, e-commerce, online advertising, to name only a few. It is essential that we look into how these algorithms work. We need to determine if there is a risk that algorithms might facilitate or permit behaviours that are contrary to competition law. With this joint study with the Bundeskartellamt we aim at reaching a common view on these matters and at starting a debate with stakeholders.” Andreas Mundt, President of the Bundeskartellamt: “The joint study is another proof of the continuing cooperation between our agencies. As digital markets keep evolving, we expand our expertise on algorithms in an exchange with each other. This is in line with our efforts to devote more resources to the digital economy with the clear-cut aim to enforce competition law also in the era of platform economy and digital business models.” Algorithms and competition are also the topic of an accompanying conference hosted by the Autorité de la concurrence and the Bundeskartellamt that is taking place in Paris today. Several renowned speakers, including business representatives, researchers and competition enforcers, are discussing potential business applications for algorithms, pricing algorithms and the risk of horizontal collusion, as well as ways to address the challenges raised by algorithms.
Literature
Better Regulation
Van Loo R. (2019)
The New Gatekeepers: Private Firms as Public Enforcers
The world’s largest businesses must routinely police other businesses. By public mandate, Facebook monitors app developers’ privacy safeguards, Citibank audits call centers for deceptive sales practices, and Exxon reviews offshore oil platforms’ environmental standards. Scholars have devoted significant attention to how policy makers deploy other private sector enforcers, such as certification bodies, accountants, lawyers, and other periphery “gatekeepers.” However, the literature has yet to explore the emerging regulatory conscription of large firms at the center of the economy. This Article examines the rise of the enforcer-firm through case studies of the industries that are home to the most valuable companies, in technology, banking, oil, and pharmaceuticals. Over the past two decades, administrative agencies have used legal rules, guidance documents, and court orders to mandate that private firms in these and other industries perform the duties of a public regulator. More specifically, firms must write rules in their contracts that reserve the right to inspect third parties. When they find violations, they must pressure or punish the wrongdoer. This form of governance has important intellectual and policy implications. It imposes more of a public duty on the firm, alters corporate governance, and may even reshape business organizations. It also gives resource-strapped regulators promising tools. If designed poorly, however, the enforcer-firm will create an expansive area of unaccountable authority. Any comprehensive account of the firm or regulation must give a prominent role to the administrative state’s newest gatekeepers.
Documents
Better Regulation
Finnish Government (2019)
Framework for innovation-friendly regulation
Radical innovations and break-through technologies are desperately needed in solving to-day’s difficult societal challenges, such as those created by climate change or ageing demographics. However, addressing complex societal challenges requires elaborate systemic planning, determined investments and often also, visionary and brave decisions by the legislators and regulators. While radical innovation may bring much needed economic benefits and solutions to pressing societal challenges, they can also generate new risks and ethical dilemmas. Hence, today’s legislators are faced with difficult questions in trying to foresee an optimal legal framework, which would sufficiently leave space for and encourage new solutions, but at the same time would ensure safe conditions and fair benefits to everyone. In light of the above, increased attention is paid to developing innovation-friendly regulatory approaches and practices. The introduction of European Commission’s Innovation Principle, as well as several national initiatives (such as regulatory sandboxes and regulation roadmaps), are good examples of such development. So far, there has not been a common definition, nor a comprehensive framework to grasp the different aspects of innovation-friendly regulation approaches and practices. Developing such framework has been one of the main objectives in Finnish government commissioned study on “Impacts of regulation on innovation and new markets”. This Policy Brief presents some first findings and introduces a draft framework for innovation-friendly regulation.
Documents
Artificial Intelligence and new technologies regulation
NESTA (2019)
Decision-making in the Age of the Algorithm
Frontline practitioners in the public sector – from social workers to police to custody officers – make important decisions every day about people’s lives. Operating in the context of a sector grappling with how to manage rising demand, coupled with diminishing resources, frontline practitioners are being asked to make very important decisions quickly and with limited information. To do this, public sector organisations are turning to new technologies to support decision-making, in particular, predictive analytics tools, which use machine learning algorithms to discover patterns in data and make predictions. While many guides exist around ethical AI design, there is little guidance on how to support a productive human-machine interaction in relation to AI. This report aims to fill this gap by focusing on the issue of human-machine interaction. How people are working with tools is significant because, simply put, for predictive analytics tools to be effective, frontline practitioners need to use them well. It encourages public sector organisations to think about how people feel about predictive analytics tools – what they’re fearful of, what they’re excited about, what they don’t understand. Based on insights drawn from an extensive literature review, interviews with frontline practitioners, and discussions with experts across a range of fields, the guide also identifies three key principles that play a significant role in supporting a constructive human-machine relationship: context, understanding, and agency.
Literature
Transparency
Coglianese C., Lehr D. (2019)
Transparency and Algorithmic Governance
Machine-learning algorithms are improving and automating important functions in medicine, transportation, and business. Government officials have also started to take notice of the accuracy and speed that such algorithms provide, increasingly relying on them to aid with consequential public-sector functions, including tax administration, regulatory oversight, and benefits administration. Despite machine-learning algorithms’ superior predictive power over conventional analytic tools, algorithmic forecasts are difficult to understand and explain. Machine learning’s “black-box” nature has thus raised concern: Can algorithmic governance be squared with legal principles of governmental transparency? We analyze this question and conclude that machine-learning algorithms’ relative inscrutability does not pose a legal barrier to their responsible use by governmental authorities. We distinguish between principles of “fishbowl transparency” and “reasoned transparency,” explaining how both are implicated by algorithmic governance but also showing that neither conception compels anything close to total transparency. Although machine learning’s black-box features distinctively implicate notions of reasoned transparency, legal demands for reason-giving can be satisfied by explaining an algorithm’s purpose, design, and basic functioning. Furthermore, new technical advances will only make machine-learning algorithms increasingly more explainable. Algorithmic governance can meet both legal and public demands for transparency while still enhancing accuracy, efficiency, and even potentially legitimacy in government.