5 questions for Fairplay AI’s Kareem Saleh

From: POLITICO's Digital Future Daily - Friday Apr 28,2023 08:02 pm
How the next wave of technology is upending the global economy and its power structures
Apr 28, 2023 View in browser
 
POLITICO's Digital Future Daily newsletter logo

By Mohar Chatterjee

With help from Derek Robertson

Fairplay AI's Kareem Saleh

Fairplay AI's Kareem Saleh.

This week, Digital Future Daily is focusing on the fast-moving landscape of generative AI and the conversation about how and whether to regulate it — from pop culture to China to the U.S. Congress. Read our full slate of coverage, from AI vs. Drake and The Weeknd to how large language models fit into the doctor's office.

...And welcome back to our regular Friday feature, The Future in Five Questions. To cap off the week in AI, we have Kareem Saleh, co-founder and CEO of FairPlay AI. The company uses AI to ensure that the AI and machine learning systems used by lenders don’t discriminate based on race, gender or other protected characteristics. Among their clients is the New York state’s Department of Financial Services, which is trying to prevent unlawful discrimination in the insurance sector. Before joining the startup ranks, Kareem served in the Obama administration, first as chief of staff to the State Department’s Special Envoy for Climate Change and later as a senior advisor, managing the U.S. government’s portfolio of emerging market investments.

Responses have been edited for length and clarity.

What’s one underrated big idea?

A fairness infrastructure for the internet that enables “Fairness-as-a-Service” — a set of on-demand tools, systems and protocols that can detect and correct biases in digital decisions in real-time. This is vital in the era of machine learning and AI, where algorithms increasingly influence everyday life — like whether you get a job interview, or a loan, or even a kidney transplant.

What's cool about the Fairness-as-a-Service approach is that it allows organizations of all sizes to adopt fairness in their decision-making without the need for extensive resources or expertise. And it's kind of our collective responsibility to prioritize ethical technology development.

We have all kinds of fairness protections and guardrails in physical space, right? Once upon a time, there were signs in shop windows that said, “No people of color allowed.” We had to adopt laws that made physical spaces more accessible to historically discriminated communities. It seems natural that as we transition to an increasingly digital existence, we will need a corresponding set of fairness safeguards.

We're in the early stages of creating those safeguards. It used to be that you would go to a bank and sit across the table from a loan officer. That loan officer would try to make a judgment about your creditworthiness based on whether he knew your kids from school or from church. Just as we had to prohibit human loan officers from taking into account characteristics like race, gender, age, marital status, etc., we need a similar set of governance rules that apply to machines making these decisions. You see this in finance, in consumer lending, in insurance, in criminal justice, and in predictive policing.

For those of us trying to make those machines safe for humanity, we need to put in place systems, processes and controls — just as we did in the physical world.

What’s a technology you think is overhyped? 

These days, the conventional wisdom is that large language models — like the one underpinning chatGPT — are overhyped. Large language models are trained on the open web. That can cause them to make silly mistakes — or worse, use discriminatory or abusive language. There's no question that large language models need to be fact-checked, de-biased and hardened against hacking attacks.

Even the most sophisticated AI companies struggle to get this stuff right. Just look at Google's roll-out of Bard. But the fact that these systems make silly mistakes today should not lead us to overly discount them. People forget that the iPhone was not terribly useful when it first launched. And Internet search engines often directed you to nonsense information in the early days. But both of those technologies spurred innovation that allowed imperfect tech to mature and eventually become essential to our lives.

So, are we in the midst of a generative AI hype cycle? Quite likely. But for all of their failings, generative large language models have the potential to transform how we interact, how we learn, how we create. When our head of data science encounters skepticism over generative AI, he says, “Look, even if these systems only make each of us 10% more productive — that’s 10%, more productivity across a wide swath of humanity, from the CEO to the janitor.” That's gonna have a profound effect on competitiveness and innovation.

What book most shaped your conception of the future?

More than any other book, “7 Powers” by Hamilton Helmer has given me a very useful framework for understanding trends in the technology landscape. He's referring to the seven sources of durable, competitive advantage. I’m super obsessed with this book — it's just a very useful framework by which I evaluate product ideas, new business opportunity investments.

What could government be doing regarding tech that it isn’t?

I think we’re starting to see a more serious consideration of what the regulatory regime for AI should look like. And there are a number of difficult questions to be answered here.

How should AI systems be governed? How do we assure ourselves of the accuracy of the data pipeline? In what industries and domains should AI systems be fairer?

Some domains — like financial services, healthcare, employment and labor — have good model governance regimes that can simply be updated for the AI era. But in other domains like social media, education and criminal justice, I suspect we're going to need a new regulatory body to ensure the safety and reliability of AI systems. Look at the criminal justice system, where we had the controversial COMPAS recidivism algorithm that was shown to be racially biased. Now imagine a thousand COMPAS systems. How do you rigorously analyze that sort of AI system at scale?

Traditionally, our economic system has been focused on innovation — letting a thousand flowers bloom. But what is really interesting is that even folks who are regulatory skeptics and industry participants who are quite knowledgeable about AI systems — say, Mr. Musk — are coming out and saying, “There's got to be some rules of the road here.”

And another common saying is that, “Oh, well, we shouldn't over-regulate, because if other jurisdictions don't apply these regulations, we are at a competitive disadvantage.” But a couple of weeks ago, China put out a set of very, very strict rules on AI governance. The Chinese cyberspace agency is saying they are concerned that AI will undermine their national unity. If you look closer at the statement from the Chinese, they are concerned about some of the same things that we've been talking about.

I think everybody is coming to the conclusion that these systems pose risks to the nation-state. We are likely to see a more global effort at harnessing the power of AI systems.

What has surprised you most this year?

We're seeing real advances in quantum computing hardware. Once, we thought we couldn’t build a 100 qubit system — now we're building 400+ qubit systems.

That is an entirely different computational model from anything humanity has ever experienced. Since the invention of the microprocessor in classical computing, we had Moore's law saying if you increase the number of bits by a factor of 10, the amount of information you can process increases by a factor of 10. But in quantum, if you increase the number of qubits by a factor of 10, it increases the computational power 1000-fold.

So computers of the future are going to be exponentially faster than the computers that exist today. Obviously, the big companies are focused on it, but the companies doing the most interesting work are still relatively small startups. The ability to process 1000x more information — that's gonna be world-changing. So I'm really excited about some of the new developments in quantum computing hardware.

 

GO INSIDE THE 2023 MILKEN INSTITUTE GLOBAL CONFERENCE: POLITICO is proud to partner with the Milken Institute to produce a special edition "Global Insider" newsletter featuring exclusive coverage, insider nuggets and unparalleled insights from the 2023 Global Conference, which will convene leaders in health, finance, politics, philanthropy and entertainment from April 30-May 3. This year’s theme, Advancing a Thriving World, will challenge and inspire attendees to lean into building an optimistic coalition capable of tackling the issues and inequities we collectively face. Don’t miss a thing — subscribe today for a front row seat.

 
 
crypto, con't

Sen. Kyrsten Sinema (D-Ariz.).

Sen. Kyrsten Sinema (D-Ariz.) during a hearing. | Bonnie Cash-Pool/Getty Images

It might not be as hot-button an issue as it was last Congress, but legislative and regulatory rulemaking around crypto is still rolling out apace.

POLITICO’s Morning Money newsletter today features a report on planned legislation from Sens. Kyrsten Sinema (I-Ariz.) and Cynthia Lummis (R-Wyo.) that will set advertising requirements for crypto, something Sinema’s office said would enforce “accurate advertising and clear disclosures on crypto products.”

MM also reported on the latest jurisdictional back-and-forth over crypto between the Securities and Exchange Commission and the Commodities Futures Trading Commission, something House Financial Services Chair Rep. Patrick McHenry (R-N.C.) told POLITICO’s Eleanor Mueller he’s hoping to split with a new bill. (Democrats, however, remain insistent that the SEC should continue to play the aggressive role it has in recent months.) — Derek Robertson

digs at the digital euro

Cheery news for the “digital euro”: “Banks don’t like it, ordinary people have never heard of it, and shopkeepers don’t want to pay for it.”

So writes POLITICO’s Bjarke Smith-Meyer in a report for Pro s yesterday, on the travails the European Union’s planned digital currency is facing as it crawls toward a planned 2026 launch. The problem, in short: None of the parties in question want to foot the bill for the transaction fees that will accompany the new financial regime, which in theory would make payments faster and reduce dependence on U.S. companies like Visa and MasterCard.

One potential solution is for the European Central Bank itself to pay those fees, Bjarke writes, but that plan faces its own form of resistance. The profits that banks make from printing banknotes (remember those?) would be the most likely means of doing so, but those profits fluctuate — and are also fill the coffers of individual national banks, who are loath to part with them.

As it turns out, it’s not easy to reinvent the wheel. “We need a digital euro that makes payments easier, not more complicated,” said one conservative German pol. “If this is not the case, it will be difficult to convince citizens to use it.” — Derek Robertson

Tweet of the Day

don't think of speeding up agi development and letting everyone have access before we solved alignment as 'summoning the demon' or an 'alien arrival' (scary, bad) think of it more like waving a magic wand (happy, fun, sparkly)

the future in 5 links

Stay in touch with the whole team: Ben Schreckinger (bschreckinger@politico.com); Derek Robertson (drobertson@politico.com); Mohar Chatterjee (mchatterjee@politico.com); Steve Heuser (sheuser@politico.com); and Benton Ives (bives@politico.com). Follow us @DigitalFuture on Twitter.

If you’ve had this newsletter forwarded to you, you can sign up and read our mission statement at the links provided.

 

GET READY FOR GLOBAL TECH DAY: Join POLITICO Live as we launch our first Global Tech Day alongside London Tech Week on Thursday, June 15. Register now for continuing updates and to be a part of this momentous and program-packed day! From the blockchain, to AI, and autonomous vehicles, technology is changing how power is exercised around the world, so who will write the rules? REGSITER HERE.

 
 
 

Follow us on Twitter

Ben Schreckinger @SchreckReports

Derek Robertson @afternoondelete

Steve Heuser @sfheuser

Benton Ives @BentonIves

 

Follow us

Follow us on Facebook Follow us on Twitter Follow us on Instagram Listen on Apple Podcast
 

To change your alert settings, please log in at https://www.politico.com/_login?base=https%3A%2F%2Fwww.politico.com/settings

This email was sent to by: POLITICO, LLC 1000 Wilson Blvd. Arlington, VA, 22209, USA

Please click here and follow the steps to .

More emails from POLITICO's Digital Future Daily

Apr 27,2023 08:02 pm - Thursday

The doctor is AI

Apr 26,2023 09:01 pm - Wednesday

A radical new idea for regulating AI

Apr 25,2023 08:28 pm - Tuesday

Can JARVIS hold a patent?

Apr 24,2023 08:15 pm - Monday

AI vs. the culture industry

Apr 21,2023 08:02 pm - Friday

5 questions for Kraken's Marco Santori

Apr 20,2023 08:02 pm - Thursday

Crypto, Miami and the future of tech hubs

Apr 19,2023 08:48 pm - Wednesday

What progress looks like to Elon