
IBMâs gone by just its initials for so long that many of us have to stop and think about what the letters stand for. International Business Machines.
I was reminded of the corporationâs singular focus last week during the TNW 2022 Conference when Seth Dobrin, IBMâs first chief AI officer, took the stage to talk about artificial intelligence.
As Dobrin put it, IBM âdoesnât do consumer AI.â You wonât be downloading IBMâs virtual assistant for your smart phone anytime soon. Big Blue wonât be getting into the selfie app AI filter game.
Simply put, IBMâs here to provide value for its clients and partners and to create AI models that make human lives easier, better, or both.
Thatâs all pretty easy to say. But how does a company thatâs not focused on creating products and services for the individual consumer actually walk that kind of talk?
According Dobrin, itâs not hard: care about how individual humans will be affected by the models you monetize:
Weâre very stringent about the type of data we will ingest and make money from.
During a discussion with the Financial Timesâ Tim Bradshaw during the conference, Dobrin used the example of large-parameter models such as GPT-3 and DALL-E 2 as a way to describe IBMâs approach.
He described those models as âtoys,â and for good reason: theyâre fun to play with, but theyâre ultimately not very useful. Theyâre prone to unpredictability in the form of nonsense, hate-speech, and the potential to output private personal information. This makes them dangerous to deploy outside of laboratories.
However, Dobrin told Bradshaw and the audience that IBM was also working on a similar system. He referred to these agents as âfoundational models,â meaning they can be used for multiple applications once developed and trained.
The IBM difference, however, is that the company is taking a human-centered approach to the development of its foundational models.
Under Dobrinâs leadership, the companyâs cherry-picking datasets from a variety of sources and then applying internal terms and conditions to them prior to their integration into models or systems.
Itâs one thing if GPT-3 accidentally spits out something offensive, these kinds of things are expected in laboratories. But itâs an entirely different situation when, as a hypothetical example, a bankâs production language model starts outputting nonsense or private information to customers.
Luckily, IBM (a company that works with corporations across a spectrum of industries including banking, transportation, and energy) doesnât believe in cramming a giant database of unchecked data into a model and hoping for the best.
Which brings us to whatâs perhaps the most interesting take away from Dobrinâs chat with Bradshaw: âbe ready for regulations.â
As the old saying goes: BS in, BS out. If youâre not in control of the data youâre training with, lifeâs going to get hard for your AI startup come regulation time.
And the Wild West of AI acquisitions is going to come to an end soon as more and more regulatory bodies seek to protect citizens from predatory AI companies and corporate overreach.
If your AI startup creates models that wonât or canât be compliant in time for use in the EU or US once the regulation hammers fall, your chances of selling them to or getting acquired by a corporation that does business internationally are slim to none.
No matter how you slice it, IBMâs an outlier. It and Dobrin apparently relish the idea of delivering compliance-ready solutions that help protect peopleâs privacy.
While the rest of big tech spends billions of dollars building eco-harming models that serve no purpose other than to pass arbitrary benchmarks, IBMâs more worried about outcomes than speculation.
And thatâs just weird. Thatâs not how the majority of the industry does business.
IBM and Dobrin are trying to redefine what big techâs position in the AI sector is. And, it turns out, when your bottom line isnât driven by advertising revenue, subscriber numbers, or future hype, you can build solutions that are as efficacious as they are ethical.
And that leaves the vast majority of people in the AI startup world with some questions to answer.
Is your startup ready for the future? Are you training models ethically, considering human outcomes, and able to explain the biases baked into your systems? Can your models be made GDPR, EU AI, and Illinois BIPA compliant?
If the current free-for-all dies out and VCs stop throwing money at prediction models and other vaporware or prestidigitation-based products, can your models still provide business value?
Thereâs probably still a little bit of money to be made for companies and startups who leap aboard the hype train, but thereâs arguably a whole lot more to be made for those whose products can actually withstand an AI winter.
Human-centered AI technologies arenât just a good idea because they make life better for humans, theyâre also the only machine learning applications worth betting on over the long haul.
When the dust settles, and weâre all less impressed by the prestidigitation and parlor tricks that big techâs spending billions of dollars on, IBM will still be out here using our planetâs limited energy resources to develop solutions with individual human outcomes in mind.
Thatâs the very definition of âsustainability,â and why IBMâs poised to become the defacto technological leader in the global artificial intelligence community under Dobrinâs so-far expert leadership.
Get the TNW newsletter
Get the most important tech news in your inbox each week.