Should your company leverage a public large language model such as ChatGPT or your own private LLM? Understand the differences. Credit: Carsten Medom Madsen / Shutterstock Large language models (LLMs) continue to command a blazing bright spotlight, as the debut of ChatGPT captured the world’s imagination and made generative AI the most widely discussed technology in recent memory (apologies, metaverse). ChatGPT catapulted public LLMs onto the stage, and its iterations continue to rev up excitement—and more than a little apprehension—about the possibilities of generating content, or code, with little more than a few prompts. While individuals and smaller businesses consider how to brace for, and benefit from, the ubiquitous disruption that generative AI and LLMs promise, enterprises have concerns and a crucial decision to make all their own. Should enterprises opt to leverage a public LLM such as ChatGPT, or their own private one? Public vs. private training data ChatGPT is a public LLM, trained on vast troves of publicly available online data. By processing vast quantities of data sourced from far and wide, public LLMs offer mostly accurate—and frequently impressive—results for just about any query or content creation task a user puts to it. Those results are also constantly improving via machine learning processes. Even so, pulling source data from the wild internet means that public LLM results can sometimes be wildly off base, and dangerously so. The potential for generative AI “hallucinations,” where the technology simply says things that aren’t true, requires users to be savvy. Enterprises, in particular, need to recognize that using public LLMs could lead employees astray, resulting in severe operational issues or even legal consequences. As a contrasting option, enterprises can create private LLMs that they own themselves and train on their own private data. The resulting generative AI applications offer less breadth, but a greater depth and accuracy of specific knowledge, speaking to the enterprise’s particular areas of expertise. Challenges posed by public LLMs For many enterprises, unique data is an invaluable currency that sets them apart. Enterprises are, therefore, extremely (and rightfully) concerned about the risk that their own employees could expose sensitive corporate or customer data by submitting that data to ChatGPT or another public LLM. This concern is based in reality, as hackers now focus on exposing ChatGPT login credentials as one of their most popular targets. A hacked account can yield the entire history of an employee’s conversations with the generative AI application, including any trade secrets or personal customer data used in queries. Even in the absence of hacking, questions posed to public LLMs are harnessed in their iterative training, potentially resulting in future direct data exposure to anyone who asks. This is why companies including Google, Amazon, and Apple are now restricting employee access to ChatGPT and building out strict governance rules, in efforts to avoid the ire of regulators as well as customers themselves. Strategically, public LLMs confront enterprises with another challenge. How do you build a unique and valuable IP on top of the same public tool and level playing field as everyone else? The answer is that it’s very difficult. That’s another reason why turning to private LLMs and enterprise-grade solutions is a strategic focus for an increasing number of organizations. The emergence of private LLMs Enterprises should recognize the opportunity to leverage their own data in a private LLM tailored to the use cases and customer experiences at the heart of their business. For those that do, a market of supportive enterprise-grade tools is quickly emerging. For example, IBM’s Watson, one of the first big names in AI and in the public imagination since the days of its Jeopardy victory, has now evolved into the private LLM development platform watsonx. Enterprise solutions such as watsonx will need to draw the line between “public baseline general shared knowledge” and “enterprise-client specific knowledge,” and where they set that distinction will be important. That said, some very powerful capabilities should come to market with the arrival of these solutions. The decision for enterprises to try to govern the usage of public LLMs or build their own private LLMs will only loom larger as time progresses. Enterprises ready to build private LLMs—and harness AI engines specifically tuned to their own core reference data—will be laying a foundation they’ll continue to rely on well into the future. Brian Sathianathan is the co-founder and chief technology officer at Iterate.ai, where he leads the company’s enterprise AI solutions. Previously, Sathianathan worked at Apple on various emerging technology projects that included the Mac operating system and the first iPhone. — Generative AI Insights provides a venue for technology leaders—including vendors and other outside contributors—to explore and discuss the challenges and opportunities of generative artificial intelligence. The selection is wide-ranging, from technology deep dives to case studies to expert opinion, but also subjective, based on our judgment of which topics and treatments will best serve InfoWorld’s technically sophisticated audience. InfoWorld does not accept marketing collateral for publication and reserves the right to edit all contributed content. Contact doug_dineley@foundryco.com. Related content feature 14 great preprocessors for developers who love to code Sometimes it seems like the rules of programming are designed to make coding a chore. Here are 14 ways preprocessors can help make software development fun again. By Peter Wayner Nov 18, 2024 10 mins Development Tools Software Development feature Designing the APIs that accidentally power businesses Well-designed APIs, even those often-neglected internal APIs, make developers more productive and businesses more agile. By Jean Yang Nov 18, 2024 6 mins APIs Software Development news Spin 3.0 supports polyglot development using Wasm components Fermyon’s open source framework for building server-side WebAssembly apps allows developers to compose apps from components created with different languages. By Paul Krill Nov 18, 2024 2 mins Microservices Serverless Computing Development Libraries and Frameworks news Go language evolving for future hardware, AI workloads The Go team is working to adapt Go to large multicore systems, the latest hardware instructions, and the needs of developers of large-scale AI systems. By Paul Krill Nov 15, 2024 3 mins Google Go Generative AI Programming Languages Resources Videos