AI

Salesforce is betting that its own content can bring more trust to generative AI

Comment

Illustration of two business people analyzing charts and graphs.
Image Credits: z_wei / Getty Images

It has become apparent in recent weeks that generative AI has the potential to transform how we interact with software, allowing us to describe what we want instead of clicking or tapping. That shift could have a profound impact on enterprise software. At the Salesforce World Tour NYC event last week, that vision was on full display.

Consider that during the 67-minute main keynote, it took less than five minutes for Salesforce CMO Sarah Franklin to introduce the subject of ChatGPT. The company then spent the next 40 minutes and several speakers talking about generative AI and the impact it would have across the entire platform. The final speaker talked about Data Cloud, an adjacent technology. It’s fair to say that other than a few minutes of introduction, it was all the company talked about.

That included discussions of EinsteinGPT, a tool for asking questions about Salesforce content, and SlackGPT, a tool for asking Slack questions about its content. In addition, the company talked about the ability to create landing pages on the fly, write sales emails (if that’s what you want) and write Apex code (Salesforce’s programming language) to programmatically trigger certain actions in a workflow, among other things.

When you think about the fact that generative AI wasn’t even really a thing people were talking about until OpenAI released ChatGPT at the end of last year, and events like this take months of planning, the company probably had to switch gears recently to focus its presentation so completely on this single subject.

Salesforce isn’t alone in its new focus on applying generative AI to its existing products and services. Over the past several months, we’ve seen many enterprise software companies announce plans to incorporate this technology into their stacks, even if overall most of these new tools are still a work in progress.

Just last week we had announcements from Zoho, Box and ServiceNow, while other companies too numerous to mention individually have made similar announcements in recent months.

A year after we saw the crypto and metaverse hype machines come crashing down, it’s fair to ask if these companies are moving too fast, chasing the next big shiny thing without considering some of the technology’s limitations, especially its well-documented hallucination problem. For this post, we are going to concentrate on Salesforce’s view of things and how it hopes to overcome some of those known issues when it comes to incorporating generative AI onto the platform.

Got 99 problems, but data ain’t one

Perhaps it’s unfair to put generative AI in the same category as other hyped technologies because we are only now seeing the direct impact of this approach. It took decades of research, development and technological shifts to get us to this point, said Juan Perez, Salesforce’s CIO, who is in charge of the company’s technology strategies.

“This is different, actually. First of all, it’s more real, and AI is not new. We’ve had decades and decades of advancement in AI,” Perez said. And he pointed out that it’s not new for Salesforce, either. It introduced its AI layer, Einstein, back in 2016, and has been refining it ever since.

Perez told TechCrunch+ that he actually uses Einstein AI to help generate reports to do his work, and the developments we are seeing with generative AI will only make the process easier. “With the advances of generative AI, with the compute power, the large-scale systems that can support these large language models, the game is entirely different,” he said.

One theme that Salesforce kept coming back to at the event was the notion of trust and that building AI solutions on top of Salesforce data could help develop more trusted AI. A more trustworthy underlying dataset could in turn help limit hallucination issues where the AI doesn’t actually know with certainty what the response should be and essentially makes one up.

But the company is working hard to make sure that the AI is giving the best answers possible with the understanding that nobody can guarantee that the generative AI won’t hallucinate answers at this point, according to Silvio Savarese, the company’s EVP and chief scientist.

“Good quality data is key for generating good quality outputs.Training or fine-tuning models using curated high-quality CRM data allows you to build trusted generative capabilities. However, even with high-quality data, LLMs can still generate hallucinations,” he said. It’s important to understand that as you implement the technology at your company.

Salesforce is working to mitigate the problem on several fronts, he said. By building its own models, the company can control for some factors that can cause the model to hallucinate. “We have full control of the learning procedure … can inject additional labeling/instruction capabilities and embed constitutional AI methods to mitigate hallucinations,” he said.

In addition, training can be ongoing rather than training once and deploying, as is sometimes the case with LLMs today, he said. “This is especially vital in the world of CRM, where data is constantly changing and freshness is mission critical. By keeping LLMs trained on the most up-to-date information, a common source of mistakes can be minimized.” It’s worth noting, however, that as customers build or bring their own LLMs, Salesforce will still supply the data but have less control over how it gets incorporated, managed and used in external models.

A matter of trust

By using a more constrained set of data for the LLMs that comes from a source like Salesforce, the company is operating on the theory that it will limit the hallucination problem. Vishal Sikka, CEO and founder at Vianai Systems, an MLOps startup told TechCrunch+ in a recent interview that it’s imperative to solve the hallucination issue before it can be used in mission-critical applications in enterprise settings.

“The first part is the safety issue because in the current state of the art, the scientists who have built this transformer technology don’t know how to make it produce good answers and not produce bad ones. They don’t know if it is even possible that it can be done,” he said.

That means that if you have a problem that requires a precise answer, you need total certainty, and we don’t have that yet.

But Ray Wang, founder and principal analyst at Constellation Research, told TechCrunch+ that there are business cases where you don’t need total accuracy to be useful.

“Generative AI ultimately requires massive amounts of data for high precision,” he said. “This requires removing false positives and false negatives with training and human augmentation. Areas where we need 100% accuracy will be hard to achieve, but if we can live with 70% or 80% accuracy, many tasks such as self-service customer care, or sales lead scoring, or campaign automation will become easier.”

Brent Hayward, CEO at Salesforce subsidiary Mulesoft, thinks that putting humans, who understand the data in the process could help tell the model when it’s right and when it’s not, what he calls “tuning for true.” That could help correct the AI when wrong and help improve models along the way.

“If the generative AI is helping create a workflow and generating code to help, the source of that code really matters,” Hayward said. “If the dataset we’ve trained the model on is all of our API’s, you can say the trust is quite high.”

He sees possibly developing a trust score based on where the data is coming from, and how much we can rely on the answers from a given set of data, an approach he thinks will be increasingly important.

People in fact remain a key part of Salesforce’s AI vision, Savarese said. “By enabling human-in-the-loop capabilities, users can verify the quality of the output of generative AI and intervene to fix hallucinations or other factual errors. This is both a powerful safety feature and an example of our core value at Salesforce AI, which is augmenting human talent rather than attempting to replace it,” he said.

Perez anticipates that part of his job, and that of all CIOs moving forward, will be ensuring that the company’s LLMs are using trusted data. “Remember the evolution of the CIO in the areas of security and privacy. We have had to really take a much stronger stance as CIOs to ensure that security is a priority, that privacy is priority. Well, now with generative AI, I think CIOs are going to have to also be like the guards of the castle and will have to ensure that there’s trusted data in support of AI,” he said.

It’s more than hallucinations

The hallucination issue is just one of the problems associated with generative AI. Another issue will be making sure that the generative AI doesn’t supply confidential company information or other sensitive data to people who aren’t supposed to see it.

Patrick Stokes, EVP and GM of platform at Salesforce, thinks that there will be limits put on what types of data can be put in the models to prevent this from happening. “Businesses and organizations like Salesforce are going to have to start to figure out what some of those swim lanes look like,” he said.

In practice that would mean looking at hiding certain fields from the model if it includes data you didn’t want unauthorized people seeing, but that’s still something that companies like Salesforce need to work out.

There’s also the issue of data ownership. For example, if you are creating a landing page on the fly, do you have permission to use the photos on that landing page (or the source of generated images)? These kinds of legal issues could slow enterprise enthusiasm for generative AI until there are clearer answers.

It’s going to be imperative to solve all of these problems, and others that are sure to arise, as we insert generative AI into more of our software. But of all the issues, limiting hallucinations is going to be paramount because everyone using the generative AI capabilities in Salesforce (and all enterprise software) is going to need to trust that the answers they are getting from the system are true and accurate and not putting the company at risk.

Salesforce is making a big bet that using its own data in LLMs will be the key to doing this. Time will tell if this is right, or at least, if it can help limit the problem.

More TechCrunch

Featured Article

CIOs’ concerns over generative AI echo those of the early days of cloud computing

CIOs trying to govern generative AI have the same concerns they had about cloud computing 15 years ago, but they’ve learned some things along the way.

2 hours ago
CIOs’ concerns over generative AI echo those of the early days of cloud computing

It sounds like the latest dispute between Apple and Fortnite-maker Epic Games isn’t over. Epic has been fighting Apple for years over the company’s revenue-sharing requirements in the App Store.…

Epic Games CEO promises to ‘fight’ Apple over ‘absurd’ changes

As deep-pocketed companies like Amazon, Google and Walmart invest in and experiment with drone delivery, a phenomenon reflective of this modern era has emerged. Drones, carrying snacks and other sundries,…

What happens if you shoot down a delivery drone?

A police officer pulled over a self-driving Waymo vehicle in Phoenix after it ran a red light and pulled into a lane of oncoming traffic, according to dispatch records. The…

Waymo robotaxi pulled over by Phoenix police after driving into the wrong lane

Welcome back to TechCrunch’s Week in Review — TechCrunch’s newsletter recapping the week’s biggest news. Want it in your inbox every Saturday? Sign up here. This week, Figma CEO Dylan…

Figma pauses its new AI feature after Apple controversy

We’ve created this guide to help parents navigate the controls offered by popular social media companies.

How to set up parental controls on Facebook, Snapchat, TikTok and more popular sites

Featured Article

You could learn a lot from a CIO with a $17B IT budget

Lori Beer’s work is a case study for every CIO out there, most of whom will never come close to JP Morgan Chase’s scale, but who can still learn from how it goes about its business.

24 hours ago
You could learn a lot from a CIO with a $17B IT budget

For the first time, Chinese government workers will be able to purchase Tesla’s Model Y for official use. Specifically, officials in eastern China’s Jiangsu province included the Model Y in…

Tesla makes it onto Chinese government purchase list

Generative AI models don’t process text the same way humans do. Understanding their “token”-based internal environments may help explain some of their strange behaviors — and stubborn limitations. Most models,…

Tokens are a big reason today’s generative AI falls short

After multiple rejections, Apple has approved Fortnite maker Epic Games’ third-party app marketplace for launch in the EU. As now permitted by the EU’s Digital Markets Act (DMA), Epic announced…

Apple approves Epic Games’ marketplace app after initial rejections

There’s no need to worry that your secret ChatGPT conversations were obtained in a recently reported breach of OpenAI’s systems. The hack itself, while troubling, appears to have been superficial…

OpenAI breach is a reminder that AI companies are treasure troves for hackers

Welcome to Startups Weekly — TechCrunch’s weekly recap of everything you can’t miss from the world of startups. Sign up here to get it in your inbox every Friday. Most…

Space for newcomers, biotech going mainstream, and more

Elon Musk’s X is exploring more ways to integrate xAI’s Grok into the social networking app. According to a series of recent discoveries, X is developing new features like the…

X plans to more deeply integrate Grok’s AI, app researcher finds

We’re about four months away from TechCrunch Disrupt 2024, taking place October 28 to 30 in San Francisco! We could not bring you this world-class event without our world-class partners…

Meet Brex, Google Cloud, Aerospace and more at Disrupt 2024

In its latest step targeting a major marketplace, the European Commission sent Amazon another request for information (RFI) Friday in relation to its compliance under the bloc’s rulebook for digital…

Amazon faces more EU scrutiny over recommender algorithms and ads transparency

Quantum Rise, a Chicago-based startup that does AI-driven automation for companies like dunnhumby (a retail analytics platform for the grocery industry), has raised a $15 million seed round from Erie…

Quantum Rise grabs $15M seed for its AI-driven ‘Consulting 2.0’ startup

On July 4, YouTube released an updated eraser tool for creators so they can easily remove any copyrighted music from their videos without affecting any other audio such as dialog…

YouTube’s updated eraser tool removes copyrighted music without impacting other audio

Airtel, India’s second-largest telecom operator, on Friday denied any breach of its systems following reports of an alleged security lapse that has caused concern among its customers. The telecom group,…

India’s Airtel dismisses data breach reports amid customer concerns

According to a recent Dealroom report on the Spanish tech ecosystem, the combined enterprise value of Spanish startups surpassed €100 billion in 2023. In the latest confirmation of this upward trend, Madrid-based…

Spain’s exposure to climate change helps Madrid-based VC Seaya close €300M climate tech fund

Forestay, an emerging VC based out of Geneva, Switzerland, has been busy. This week it closed its second fund, Forestay Capital II, at a hard cap of $220 million. The…

Forestay, Europe’s newest $220M growth-stage VC fund, will focus on AI

Threads, Meta’s alternative to Twitter, just celebrated its first birthday. After launching on July 5 last year, the social network has reached 175 million monthly active users — that’s a…

A year later, what Threads could learn from other social networks

J2 Ventures, a firm led mostly by U.S. military veterans, announced on Thursday that it has raised a $150 million second fund. The Boston-based firm invests in startups whose products…

J2 Ventures, focused on military healthcare, grabs $150M for its second fund

HealthEquity said in an 8-K filing with the SEC that it detected “anomalous behavior by a personal use device belonging to a business partner.”

HealthEquity says data breach is an ‘isolated incident’

Roll20 said that on June 29 it had detected that a “bad actor” gained access to an account on the company’s administrative website for one hour.

Roll20, an online tabletop role-playing game platform, discloses data breach

Fisker has a willing buyer for its remaining inventory of all-electric Ocean SUVs, and has asked the Delaware Bankruptcy Court judge overseeing its Chapter 11 case to approve the sale.…

Fisker asks bankruptcy court to sell its EVs at average of $14,000 each

Teddy Solomon just moved to a new house in Palo Alto, so he turned to the Stanford community on Fizz to furnish his room. “Every time I show up to…

Fizz, the anonymous Gen Z social app, adds a marketplace for college students

With increasing competition for what is, essentially, still a small number of hard tech and deep tech deals, Sidney Scott realized it would be a challenge for smaller funds like…

Why deep tech VC Driving Forces is shutting down

A guide to turn off reactions on your iPhone and Mac so you don’t get surprised by effects during work video calls.

How to turn off those silly video call reactions on iPhone and Mac

Amazon has decided to discontinue its Astro for Business device, a security robot for small- and medium-sized businesses, just seven months after launch.  In an email sent to customers and…

Amazon retires its Astro for Business security robot after only 7 months

Hiya, folks, and welcome to TechCrunch’s regular AI newsletter. This week in AI, the U.S. Supreme Court struck down “Chevron deference,” a 40-year-old ruling on federal agencies’ power that required…

This Week in AI: With Chevron’s demise, AI regulation seems dead in the water