SHARE
Facebook X Pinterest WhatsApp

Data Privacy Concerns Deter Enterprises From Commercial LLMs

thumbnail
Data Privacy Concerns Deter Enterprises From Commercial LLMs

Robotics and AI are being used to create artworks, with machines being trained to mimic the style of previous pieces. They can also generate creative visuals from text.

A new survey reveals insights on challenges, open-source alternatives, and the future of commercial LLMs.

Written By
thumbnail
David Curry
David Curry
Aug 31, 2023

Businesses of all shapes, sizes, and sectors are trying to figure out how to integrate large language models (LLMs) into their workflows, but so far, very few are willing to integrate commercial LLMs in production, primarily due to data privacy concerns. 

That’s according to a new survey published by low-code platform Predibase, which asked over 150 people in the artificial intelligence and machine learning industry about their company’s investments in LLMs, challenges, open-source interest, and customization efforts. 

SEE ALSO: Maximizing Software Quality with Artificial Intelligence

In the report, more than 85 percent of respondents said they had no plans to use commercial LLMs in the production stage, with 44 percent of that group using LLMs for experimentation, 27 percent expecting to start using LLMs in the next 12 months, and 15 percent having no immediate plans. 

Only 14 percent of respondents had LLMs in production, with one percent having more than two. 

“It is now open season for LLMs. Thanks to the widespread recognition of OpenAI’s ChatGPT, businesses are in an arms race to gain a competitive edge using the latest AI capabilities. Still, they require more customized LLMs to meet domain-specific use cases,” said Piero Molino, co-founder and CEO of Predibase, in a statement.

The most prominent reason for delaying or avoiding LLMs in production is giving up access to proprietary data. Some of the most popular LLMs are not open-source and by injecting proprietary data they may have to sacrifice ownership of that data to the LLM owner, which is then used to further train the model. According to Predibase, organizations can get over this roadblock by using open-source LLMs in a virtual private cloud, providing them with ways to maintain full ownership of their data. 

Another hurdle 30 percent of organizations face is customization and fine-tuning, as this often requires a high velocity of proprietary data and expertise in how to train LLMs. Outside of leading-edge businesses, most do not have the resources of talent to pull this off, which forces them to work with third-party providers or consultancy firms. There are open-source platforms coming available aimed at reducing the complexity of model training, which should enable more businesses to succeed. 

To train state-of-the-art LLMs organizations need to be willing to spend millions on cloud servers, AI, and other software costs. In the survey, 17 percent said this was an roadblock to adding LLMs in production. 

Other roadblocks included LLM hallucinations, which happen to all types of generative AI services. Even ChatGPT, which is seen by some as the golden standard for generative AI sophistication, is susceptible to hallucinations and can make up statistics and information, although OpenAI and others are embedding new techniques to reduce this from happening. Latency was the last concern mentioned by Predibase, as LLMs need a lot of compute resources to perform at sub-second timeframes. 

Even with these roadblocks, there is a keen awareness in the technology industry of the benefits which integrating LLMs into workflows can bring. The growth of open-source alternatives should make the market more accessible to organizations of all sizes, and it seems like the gap between proprietary solutions and open-source is closing every year. 

thumbnail
David Curry

David is a technology writer with several years experience covering all aspects of IoT, from technology to networks to security.

Recommended for you...

AI Agents Need Keys to Your Kingdom
The Rise of Autonomous BI: How AI Agents Are Transforming Data Discovery and Analysis
Why the Next Evolution in the C-Suite Is a Chief Data, Analytics, and AI Officer
Digital Twins in 2026: From Digital Replicas to Intelligent, AI-Driven Systems

Featured Resources from Cloud Data Insights

The Difficult Reality of Implementing Zero Trust Networking
Misbah Rehman
Jan 6, 2026
Cloud Evolution 2026: Strategic Imperatives for Chief Data Officers
Why Network Services Need Automation
The Shared Responsibility Model and Its Impact on Your Security Posture
RT Insights Logo

Analysis and market insights on real-time analytics including Big Data, the IoT, and cognitive computing. Business use cases and technologies are discussed.

Property of TechnologyAdvice. © 2026 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.