Privacy in the Digital Age With Margot E. Kaminski
Associate Professor Margot E. Kaminski teaches, researches, and writes on law and technology. Her groundbreaking work has focused on privacy, speech, and online civil liberties, in addition to international intellectual property law and legal issues raised by artificial intelligence (AI) and robotics. She also serves as director of the Privacy Initiative at Silicon Flatirons.
What initially attracted you to this area of the law?
I worked in publishing after college, right around the time that e-books were becoming popular. That, along with the rise of social media, got me interested in how we arrange existing legal rights around new technologiesfrom speech to privacy to intellectual property rights.
Later, while in law school, I interned for a summer at the Electronic Frontier Foundation (EFF), a nonprofit in San Francisco that describes itself as defending civil liberties in the digital world. There was a team there working on national security surveillance litigation. That really opened my eyes to the central significance of human rights in the digital age. But Ive always been interested in the nitty gritty aspects of regulation, especially the role that transparency can play. I came back to law school and co-founded a clinic, the Media Freedom and Information Access (MFIA) clinic, which litigates government accountability and transparency cases, including cases involving new technologies. We found ourselves asking who should count as a journalist, and what role information technologies were playing in the face of existing power disparitiesboth as tools of surveillance and tools of accountability.
By the time I became a professor, I had really honed in on privacy. Its such a fast-moving policy space, with big implications for democracy and individual freedom. I spent a few years working on the privacy issues raised by unmanned aerial vehicles (UAVs), or drones, and then got the opportunity to go to Europe through a Fulbright grant. Thats when I turned to working increasingly on comparative data privacy law, with a focus on automated decisionmaking systems and AI.
Have legal issues raised by AI and robotics changed since you graduated from law school?
Gosh, yes. When I was in law school many of these questions felt ahead of the curve or hypothetical. Now, we have companies using facial recognition (a type of computer algorithm) to scan applicants faces to try to determine emotions and extrapolate personality traits, for example. We have government agencies using algorithms, including AI systems, to try to allocate benefits or catch fraud. The use of and investment in AI systems is everywhere, and lawmakers are taking the potential harms of AI systems seriously. Colorado, for example, just enacted a new law on facial recognition.
Youve written extensively about the role of AI algorithms in decision making. How do you see the future of balancing decisional authority between humans and machines?
This is a hard one. I dont see my role as predicting the future, necessarily. Im more interested in trying to figure out what the law can do to ensure that human values stay on the tablethat as we increasingly create and use new sociotechnical systems, we put in place whatevers necessary to make sure we dont lose sight of what matters. I definitely believe, for a number of reasons, that there are some decisional realms where we will always use humans. Legal decisions, for example, arent just about correctness and efficiency. Legitimacy, justification, accountability, even a respect for the dignitary rights of the person affected by a decisionthese are all reasons why law, at least as an ideal, isnt suited to automation.
The most interesting problems arent about whether to use a machine or a human, but about how to get them to work together. For example, Ive coauthored this recent article on humans in the loop, or the people involved in automated decisions. Often, well-intentioned lawmakers will look at a decision made by an AI system and try to solve some set of perceived problems by requiring that a human be involved. Its not that humans are worse decision makers than machinesin fact, humans still do a lot of things, like crossing contexts or dealing with edge cases, much better. But putting a human in the loop thoughtlessly actually creates new problems. Hybrid human-machine systems have known weaknesses and can be subject to complex failure cascades. So if were going to put a human in the loop of particularly significant automated decisions, we have to know why were putting her there, and set her (and the system) up to succeed.
You recently worked with Colorado Laws Samuelson-Glushko Technology Law & Policy Clinic to develop comments responding to the Attorney Generals Pre-Rulemaking Considerations for the Colorado Privacy Act. Tell us about that work.
I feel so lucky to be at a school with a tech law clinic! Professor [Blake] Reid 10 is a joy to work with, and his students (who are often also my students, from other classes) take their work very seriously and produce impressive and important output. This most recent project, responding to the Colorado AGs office on the Colorado Privacy Act, is a great example. Two clinic students worked to exhaustively identify aspects of the act that could benefit from focused rulemaking. They did an extraordinary amount of research, ranging from technical articles on how to best design an effective consent stream, to organizational literature on how to make an impact assessment successful. They also pointed the AGs office to resources on other privacy laws, both in Europe and in other states like California. This is particularly important as states like ours weigh the benefits of harmonization, which typically makes for lower compliance costs for businesses, with the appeal of being a policy leader in the consumer protection space.
Colorado Laws Tech Law and Policy program, along with the law schools Silicon Flatirons Center for Law, Technology, and Entrepreneurship, are nationally recognized. How would you like to see these programs evolve or grow?
Again, I feel very lucky to be at a school that has so many faculty members working in related spaces. Each of us does something slightly differentProfessor [Kristelia] Garcia works on copyright law, Professor [Brad] Bernthal '01 on entrepreneurship, Professor Reid on telecommunications and platform law, Professor [Harry] Surden on patent law and a different area of AIbut were able to collectively offer our students a depth of expertise and classes that arent really available elsewhere, except at a few very top law schools. There is, however, always room for growth. I would love to see us be able to offer our students more privacy courses, in particular. It would be amazing to be able to offer data privacy for practitioners or an international privacy course. We also, despite the expertise on our faculty, have yet to offer a class on law and AI!
What research themes or projects are you most looking forward to digging into in the coming year?
I have a few projects Im really excited about. This humans in the loop piece I already mentioned is a big one. So is a piece Ive been revising this summer called Regulating the Risks of AI. Most laws targeting AI have been risk regulationthe kind of thing we use, for example, in environmental law, or that companies use to try to mitigate risks. Risk regulation comes with a particular set of policy baggage. Its been funand challenging!to dig into how aspects of it do and dont work when its applied to algorithms and associated practices.
Im also really looking forward this fall to getting back into a piece Ive been calling Data as Speech Infrastructure, where Ill be looking at data privacy laws through the lens of the First Amendment. And theres a good chance something will come of all of the discussions Ive been having about the data privacy implications of the Supreme Courts decision revoking the right to abortion in Dobbs. In short, theres always something to do.
