
In the wake of the firing of Timnit Gebru and other notable AI researchers at Google, Alphabetās circled the wagons and lawyered up. Reports flow out of Mountain View depicting teams of lawyers censoring scientific research and acting as unnamed collaborators and peer-reviewers.
Most recently, Business Insider managed to interview several researchers who painted a startling and bleak picture of what itās like to try and conduct research under such an anti-scientific regime.
Per the article, one researcher said:
Youāve got dozens of lawyers ā no doubt, highly trained lawyers ā who nonetheless actually know very little about this technology ā¦ and theyāre working their way through your research like English undergrads reading a poem.
The problem here is that Google isnāt censoring research to avoid, say, its secrets getting out. Its lawyers are targeting scientific research that makes the company look bad.
The person quoted above added that they were specifically talking about crossing out references to āfairnessā and ābiasā and scientists being told to change the results of their work. Itās not only unethical, itās incredibly dangerous.
The tea: Googleās AI is broken. It might be a trillion-dollar company and the most cutting-edge AI outfit on Earth, but its algorithms are biased. And thatās dangerous.
No matter how you slice it, Googleās AI doesnāt work as well for people who donāt look like the vast majority of Googleās employees (white dudes) as it does for people who do. From Searchās conflation of Black people and animals to the algorithms running the camera on the Pixel 6ās inability to properly process non-white skin tones, Googleās machine-learning woes are well-documented.
This is a big problem and it isnāt easy to fix. Imagine building a car that didnāt work as well for Black people and women as it did for white guys, selling 200 million, and then people slowly learning their automobiles were racist.
Thereād be a lot of feelings and emotions about what that would mean.
Googleās current situation is a lot like that. Its products are everywhere. It canāt just recall Search or put Google Ads on hold for a few days while it rethinks the entire world of deep learning to exclude bias. Why not fix world hunger and make puppies immortal while theyāre at it?
So what do you do when youāre one of the richest companies in the world and you come up against a truth so awful that its existence makes your model seem evil?
You do what big tobacco did. You find people willing to say whatās in your companyās best interests and you use them to stop the people telling the truth from sharing their research.
The National Institutes of Health released research in 2007 describing the role of lawyers during the big tobacco legal battles of the previous decades.
In the paper, which is titled āTobacco industry lawyers as a disease vector,ā the researchers attribute the spread of diseases associated with long-term tobacco use to the tactics employed by industry lawyers.
Some key takeaways from the paper include:
- Despite their obligation to do so, tobacco companies often failed to conduct product safety research or, when research was conducted, failed to disseminate the results to the medical community and to the public.
- Tobacco company lawyers have been involved in activities having little or nothing to do with the practice of law, including gauging and attempting to influence company scientistsā beliefs, vetting ināhouse scientific research, and instructing ināhouse scientists not to publish potentially damaging results.
- Additionally, company lawyers have taken steps to manufacture attorneyāclient privilege and workāproduct cover to assist their clients in protecting sensitive documents from disclosure, have been involved in the concealment of such documents, and have employed litigation tactics that have largely prevented successful lawsuits against their client companies.
And weāre seeing the same potential with Googleās approach. The companyās treating the scientific method as an optional component of research.
As researcher Jack Clark, formerly of OpenAI, pointed out on Twitter:
I like to collaborate with people in research and I do a huge amount of work on AI measurement/assessment/synthesis/analysis. Why would I try and collaborate with people at Google if I know that thereās some invisible group of people who will get inside our research paper?
Clarkās talking about legibility here, the idea that the researchers have their names on the papers but the censors and lawyers donāt.
See, down the road a few years, if Googleās inability to address bias or create algorithms that are fair turns out deadly at scale over time, no lawyers will be harmed in the proceeding lawsuits.
And thatās not fair. Billions of people put their trust in Google products every day. The AI we rely on is a part of our lives that influences our decisions. Whatever Googleās lawyers are hiding could hurt us all.
Get the TNW newsletter
Get the most important tech news in your inbox each week.