AI

Women in AI: Claire Leibowicz, AI and media integrity expert at PAI

Comment

Women in AI Claire Leibowicz
Image Credits: TechCrunch

To give AI-focused women academics and others their well-deserved — and overdue — time in the spotlight, TechCrunch is launching a series of interviews focusing on remarkable women who’ve contributed to the AI revolution. We’ll publish several pieces throughout the year as the AI boom continues, highlighting key work that often goes unrecognized. Read more profiles here.

Claire Leibowicz is the head of the AI and media integrity program at the Partnership on AI (PAI), the industry group backed by Amazon, Meta, Google, Microsoft and others committed to the “responsible” deployment of AI tech. She also oversees PAI’s AI and media integrity steering committee.

In 2021, Leibowicz was a journalism fellow at Tablet Magazine, and in 2022, she was a fellow at the Rockefeller Foundation’s Bellagio Center focused on AI governance. Leibowicz — who holds a BA in psychology and computer science from Harvard and a master’s degree from Oxford — has advised companies, governments and nonprofit organizations on AI governance, generative media and digital information.

Q&A

Briefly, how did you get your start in AI? What attracted you to the field?

It may seem paradoxical, but I came to the AI field from an interest in human behavior. I grew up in New York, and I was always captivated by the many ways people there interact and how such a diverse society takes shape. I was curious about huge questions that affect truth and justice, like how do we choose to trust others? What prompts intergroup conflict? Why do people believe certain things to be true and not others? I started out exploring these questions in my academic life through cognitive science research, and I quickly realized that technology was affecting the answers to these questions. I also found it intriguing how artificial intelligence could be a metaphor for human intelligence.

That brought me into computer science classrooms where faculty — I have to shout out Professor Barbara Grosz, who is a trailblazer in natural language processing, and Professor Jim Waldo, who blended his philosophy and computer science background — underscored the importance of filling their classrooms with non-computer science and -engineering majors to focus on the social impact of technologies, including AI. And this was before “AI ethics” was a distinct and popular field. They made clear that, while technical understanding is beneficial, technology affects vast realms, including geopolitics, economics, social engagement and more, thereby requiring people from many disciplinary backgrounds to weigh in on seemingly technological questions.

Whether you’re an educator thinking about how generative AI tools affect pedagogy, a museum curator experimenting with a predictive route for an exhibit or a doctor investigating new image detection methods for reading lab reports, AI can impact your field. This reality, that AI touches many domains, intrigued me: There was intellectual variety inherent to working in the AI field, and this brought with it a chance to impact many facets of society.

What work are you most proud of in the AI field?

I’m proud of the work in AI that brings disparate perspectives together in a surprising and action-oriented way — that not only accommodates, but [also] encourages, disagreement. I joined the PAI as the organization’s second staff member six years ago and sensed right away the organization was trailblazing in its commitment to diverse perspectives. PAI saw such work as a vital prerequisite to AI governance that mitigates harm and leads to practical adoption and impact in the AI field. This has proven true, and I have been heartened to help shape PAI’s embrace of multidisciplinarity and watch the institution grow alongside the AI field.

Our work on synthetic media over the past six years started well before generative AI became part of the public consciousness, and exemplifies the possibilities of multistakeholder AI governance. In 2020, we worked with nine different organizations from civil society, industry and media to shape Facebook’s Deepfake Detection Challenge, a machine learning competition for building models to detect AI-generated media. These outside perspectives helped shape the fairness and goals of the winning models — showing how human rights experts and journalists can contribute to a seemingly technical question like deepfake detection. Last year, we published a normative set of guidance on responsible synthetic media — PAI’s Responsible Practices for Synthetic Media — that now has 18 supporters from extremely different backgrounds, ranging from OpenAI to TikTok to Code for Africa, Bumble, BBC and WITNESS. Being able to put pen to paper on actionable guidance that is informed by technical and social realities is one thing, but it’s another to actually get institutional support. In this case, institutions committed to providing transparency reports about how they navigate the synthetic media field. AI projects that feature tangible guidance, and show how to implement that guidance across institutions, are some of the most meaningful to me.

How do you navigate the challenges of the male-dominated tech industry and, by extension, the male-dominated AI industry?

I have had both wonderful male and female mentors throughout my career. Finding people who simultaneously support and challenge me is key to any growth I have experienced. I find that focusing on shared interests and discussing the questions that animate the field of AI can bring people with different backgrounds and perspectives together. Interestingly, PAI’s team is made up of more than half women, and many of the organizations working on AI and society or responsible AI questions have many women on staff. This is often in contrast to those working on engineering and AI research teams and is a step in the right direction for representation in the AI ecosystem.

What advice would you give to women seeking to enter the AI field?

As I touched on in the previous question, some of the primarily male-dominated spaces within AI that I have encountered have also been those that are the most technical. While we should not prioritize technical acumen over other forms of literacy in the AI field, I have found that having technical training has been a boon to both my confidence and effectiveness in such spaces. We need equal representation in technical roles and an openness to the expertise of folks who are experts in other fields like civil rights and politics that have more balanced representation. At the same time, equipping more women with technical literacy is key to balancing representation in the AI field.

I have also found it enormously meaningful to connect with women in the AI field who have navigated balancing family and professional life. Finding role models to talk to about big questions related to career and parenthood — and some of the unique challenges women still face at work — has made me feel better equipped to handle some those challenges as they arise.

What are some of the most pressing issues facing AI as it evolves?

The questions of truth and trust online — and offline — become increasingly tricky as AI evolves. As content ranging from images to videos to text can be AI-generated or modified, is seeing still believing? How can we rely on evidence if documents can easily and realistically be doctored? Can we have human-only spaces online if it’s extremely easy to imitate a real person? How do we navigate the trade-offs that AI presents between free expression and the possibility that AI systems can cause harm? More broadly, how do we ensure the information environment is not only shaped by a select few companies and those working for them but [also] incorporates the perspectives of stakeholders from around the world, including the public?

Alongside these specific questions, PAI has been involved in other facets of AI and society, including how we consider fairness and bias in an era of algorithmic decision-making, how labor impacts and is impacted by AI, how to navigate responsible deployment of AI systems and even how to make AI systems more reflective of myriad perspectives. At a structural level, we must consider how AI governance can navigate vast trade-offs by incorporating varied perspectives.

What are some issues AI users should be aware of?

First, AI users should know that if something sounds too good to be true, it probably is.

The generative AI boom over the past year has, of course, reflected enormous ingenuity and innovation, but it has also led to public messaging around AI that is often hyperbolic and inaccurate.

AI users should also understand that AI is not revolutionary, but exacerbating and augmenting existing problems and opportunities. This does not mean they should take AI less seriously, but rather use this knowledge as a helpful foundation for navigating an increasingly AI-infused world. For example, if you are concerned about the fact that people could miscontextualize a video before an election by changing the caption, you should be concerned about the speed and scale at which they can mislead using deepfake technology. If you are concerned about the use of surveillance in the workplace, you should also consider how AI will make such surveillance easier and more pervasive. Maintaining a healthy skepticism about the novelty of AI problems, while also being honest about what is distinct about the current moment, is a helpful frame for users to bring to their encounters with AI.

What is the best way to responsibly build AI?

Responsibly building AI requires us to broaden our notion of who plays a role in “building” AI. Of course, influencing technology companies and social media platforms is a key way to affect the impact of AI systems, and these institutions are vital to responsibly building technology. At the same time, we must acknowledge how diverse institutions from across civil society, industry, media, academia and the public must continue to be involved to build responsible AI that serves the public interest.

Take, for example, the responsible development and deployment of synthetic media.

While technology companies might be concerned about their responsibility when navigating how a synthetic video can influence users before an election, journalists may be worried about imposters creating synthetic videos that purport to come from their trusted news brand. Human rights defenders might consider responsibility related to how AI-generated media reduces the impact of videos as evidence of abuses. And artists might be excited by the opportunity to express themselves through generative media, while also being concerned about how their creations might be leveraged without their consent to train AI models that produce new media. These diverse considerations show how vital it is to involve different stakeholders in initiatives and efforts to responsibly build AI, and how myriad institutions are affected by — and affecting — the way AI is integrated into society.

How can investors better push for responsible AI?

Years ago, I heard DJ Patil, the former chief data scientist in the White House, describe a revision to the pervasive “move fast and break things” mantra of the early social media era that has stuck with me. He suggested the field “move purposefully and fix things.”

I loved this because it didn’t imply stagnation or an abandonment of innovation, but intentionality and the possibility that one could innovate while embracing responsibility. Investors should help induce this mentality — allowing more time and space for their portfolio companies to bake in responsible AI practices without stifling progress. Oftentimes, institutions describe limited time and tight deadlines as the limiting factor for doing the “right” thing, and investors can be a major catalyst for changing this dynamic.

The more I have worked in AI, the more I have found myself grappling with deeply humanistic questions. And these questions require all of us to answer them.

More TechCrunch

X will now allow users to post consensually produced NSFW content as long as it is prominently labeled as such.

X tweaks rules to formally allow adult content

Ashby consolidates existing talent acquisition tools and leans heavily on AI to automate the more repetitive steps in the recruitment pipeline.

Ashby injects recruiting with a dose of AI

Spotify has announced it’s hiking subscriptions for customers in the U.S., the second such price increase in the space of a year. The music-streaming giant reports that premium pricing will…

Spotify to increase premium pricing in the US to $11.99 per month

Monzo has announced its 2024 financial results, revealing its first full-year pre-tax profit. The company also confirmed that it’s in the early stages of expanding into the broader European market…

UK neobank Monzo reports first full (pre-tax) profit, prepares for EU expansion with Dublin hub

Featured Article

Inside Apple’s efforts to build a better recycling robot

Last week, TechCrunch paid a visit to Apple’s Austin, Texas manufacturing facilities. Since 2013, the company has built its Mac Pro desktop about 20 minutes north of downtown. The 400,000 square foot facility sits in a maze of industry parks, a quick trip south from the company’s in-progress corporate campus. In recent years, the capital…

2 hours ago
Inside Apple’s efforts to build a better recycling robot

Early attempts at making dedicated hardware to house artificial intelligence smarts have been criticized as, well, a bit rubbish. But here’s an AI gadget-in-the-making that’s all about rubbish, literally: Finnish…

Binit is bringing AI to trash

Temasek has previously invested in Lenskart, and this new funding follows a $500 million investment by the Abu Dhabi Investment Authority last year.

Temasek, Fidelity buy $200M stake in Lenskart at $5B valuation

Less than one year after its iOS launch, French startup ten ten has gone viral with a walkie talkie app that allows teens to send voice messages to their close…

French startup ten ten reinvents the walkie-talkie

Featured Article

Unicorn-rich VC Wesley Chan owes his success to a Craigslist job washing lab beakers

While all of Wesley Chan’s success has been well-documented over the years, his personal journey…not so much. Chan spoke to TechCrunch about the ways his life impacts how he invests in startups.

18 hours ago
Unicorn-rich VC Wesley Chan owes his success to a Craigslist job washing lab beakers

Presumptive Republican presidential nominee Donald Trump now has an account on the short-form video app that he once tried to ban. Trump’s TikTok account, which launched on Saturday night, features…

Trump takes off on TikTok

With fewer than 400,000 inhabitants, Iceland receives more than its fair share of tourists — and of venture capital.

Iceland’s startup scene is all about making the most of the country’s resources

Kobo put out a handful of new e-readers a few weeks back: color versions of the excellent Libra 2 and Clara, as well as an updated monochrome version of the…

Kobo’s new e-readers are a sidegrade most can skip (with one exception)

In an interview at his home near Reykjavík, the entrepreneur-turned-VC shared thoughts on his ventures and the journey that led him from Unity to climate tech, a homecoming of sorts.

Unity co-founder David Helgason’s next act: Gaming the climate crisis

Welcome back to TechCrunch’s Week in Review — TechCrunch’s newsletter recapping the week’s biggest news. Want it in your inbox every Saturday? Sign up here. Over the past eight years,…

Fisker collapsed under the weight of its founder’s promises

What is AI? We’ve put together this non-technical guide to give anyone a fighting chance to understand how and why today’s AI works.

WTF is AI?

President Joe Biden has vetoed H.J.Res. 109, a congressional resolution that would have overturned the Securities and Exchange Commission’s current approach to banks and crypto. Specifically, the resolution targeted the…

President Biden vetoes crypto custody bill

Featured Article

Industries may be ready for humanoid robots, but are the robots ready for them?

How large a role humanoids will play in that ecosystem is, perhaps, the biggest question on everyone’s mind at the moment.

2 days ago
Industries may be ready for humanoid robots, but are the robots ready for them?

VCs are clamoring to invest in hot AI companies, and willing to pay exorbitant share prices for coveted spots on their cap tables. Even so, most aren’t able to get…

VCs are selling shares of hot AI companies like Anthropic and xAI to small investors in a wild SPV market

The fashion industry has a huge problem: Despite many returned items being unworn or undamaged, a lot, if not the majority, end up in the trash. An estimated 9.5 billion…

Deal Dive: How (Re)vive grew 10x last year by helping retailers recycle and sell returned items

Tumblr officially shut down “Tips,” an opt-in feature where creators could receive one-time payments from their followers.  As of today, the tipping icon has automatically disappeared from all posts and…

You can no longer use Tumblr’s tipping feature 

Generative AI improvements are increasingly being made through data curation and collection — not architectural — improvements. Big Tech has an advantage.

AI training data has a price tag that only Big Tech can afford

Keeping up with an industry as fast-moving as AI is a tall order. So until an AI can do it for you, here’s a handy roundup of recent stories in the world…

This Week in AI: Can we (and could we ever) trust OpenAI?

Jasper Health, a cancer care platform startup, laid off a substantial part of its workforce, TechCrunch has learned.

General Catalyst-backed Jasper Health lays off staff

Featured Article

Live Nation confirms Ticketmaster was hacked, says personal information stolen in data breach

Live Nation says its Ticketmaster subsidiary was hacked. A hacker claims to be selling 560 million customer records.

3 days ago
Live Nation confirms Ticketmaster was hacked, says personal information stolen in data breach

Featured Article

Inside EV startup Fisker’s collapse: how the company crumbled under its founders’ whims

An autonomous pod. A solid-state battery-powered sports car. An electric pickup truck. A convertible grand tourer EV with up to 600 miles of range. A “fully connected mobility device” for young urban innovators to be built by Foxconn and priced under $30,000. The next Popemobile. Over the past eight years, famed vehicle designer Henrik Fisker…

3 days ago
Inside EV startup Fisker’s collapse: how the company crumbled under its founders’ whims

Late Friday afternoon, a time window companies usually reserve for unflattering disclosures, AI startup Hugging Face said that its security team earlier this week detected “unauthorized access” to Spaces, Hugging…

Hugging Face says it detected ‘unauthorized access’ to its AI model hosting platform

Featured Article

Hacked, leaked, exposed: Why you should never use stalkerware apps

Using stalkerware is creepy, unethical, potentially illegal, and puts your data and that of your loved ones in danger.

3 days ago
Hacked, leaked, exposed: Why you should never use stalkerware apps

The design brief was simple: each grind and dry cycle had to be completed before breakfast. Here’s how Mill made it happen.

Mill’s redesigned food waste bin really is faster and quieter than before

Google is embarrassed about its AI Overviews, too. After a deluge of dunks and memes over the past week, which cracked on the poor quality and outright misinformation that arose…

Google admits its AI Overviews need work, but we’re all helping it beta test

Welcome to Startups Weekly — Haje‘s weekly recap of everything you can’t miss from the world of startups. Sign up here to get it in your inbox every Friday. In…

Startups Weekly: Musk raises $6B for AI and the fintech dominoes are falling