Even if you’re not closely following the advancements in AI and Large Language Models (LLMs), it’s highly likely that you’ve encountered ChatGPT, —whether in the news or through a friend. Instead of hearing, “Hey, let’s Google that”, you may now hear, “Hey, let’s ask ChatGPT”.
Let’s explore this technology and how it has been rapidly integrating into our daily lives, particularly in the workplace and across various activities within companies of all sizes. We’ll conclude with tips and advice on how to use it responsibly while maintaining realistic expectations. Your own workplace may already be taking initiatives to adopt tools or applications powered by LLMs.
First, let’s define what an LLM is: “It is a subset of Generative AI that refers to artificial intelligence systems capable of understanding and generating human-like language” (Chen, 2024, p.5). These generation capabilities are achieved through training on massive and diverse datasets over extended periods using deep learning — a branch of machine learning.
By leveraging layers of neural networks, which employ probability and other techniques to establish associations within unstructured data (e.g. text, images, videos, audio), LLMs generate the most likely response to a given query. Deep learning enables these models to mimic certain aspects of human cognition, processing vast amounts of information and making decisions in a way loosely analogous to how our brain maps memories, knowledge, and electric signals across billions of neurons.
Building on the concept of unstructured data, these models rely on it as the primary input and output for the services and tools increasingly integrated into our daily tasks. Text serves as the main interface for interacting with LLM applications. For instance, these models can summarize transcripts of virtual meetings, detailing what each participant shared, classifying topics discussed, and even suggesting action items— particularly helpful features for those of us who often forget to take notes.
In a broader, organization-wide use cases, LLMs assist with handling questions and answers by combining search functionalities to process queries efficiently. This often involves training a custom model using techniques like Retrieval-Augmented Generation (RAG), which equips the LLM to retain relevant context from an organization’s internal data. Such models can search through indexed company data in vector databases, delivering more accurate and tailored results.
The capabilities of multi-modal models that LLMs have recently improved, allowing the integration of more complex data and interactions. These models can now follow prompts and instructions at a higher level of programmability, enabling outputs such as customized image or video animations. Chatbots powered by LLMs can also generate highly realistic voice responses, making it increasingly difficult to distinguish between a bot and a real person.
All of this sounds fantastic—revolutionary, disruptive, and incredibly innovative. The hype surrounding LLMs remains high, fueled by the rapid pace of advancements in model performance, framework availability, and the emergence of new businesses and technologies. And then there’s the ambitious concept of Artificial General Intelligence (AGI), a goal that continues to captivate imaginations, though it carries significant risks and demands cautious exploration.
However, there are important factors to consider when adopting this technology. As end users, we must exercise caution when assigning tasks to any LLMs-powered service or application. One major limitation is the phenomenon of “hallucinations,” where models generate inaccurate or entirely fabricated responses. Even with well-tuned models, results can sometimes be unreliable. Therefore, it’s crucial not to trust these outputs blindly. Much like cross-checking information from multiple sources on Google, we should scrutinize LLM- generated responses and rely on our own domain knowledge and instincts.
To get the best results, turn interactions with LLMs into ongoing conversations rather than relying on one-shot exchanges (known as zero-shot prompts). Providing clear, specific instructions increases the likelihood of achieving accurate and useful outputs.
What Lies Ahead?
As mentioned earlier, the technological landscape around LLMs is evolving rapidly. Speaking of AGI, while it may be the ultimate aspiration for some, it’s a goal that requires careful and deliberate steps. When leveraging LLMs, we must prioritize quality, security, privacy and dependability in our professional and organizational activities. LLMs are powerful tools, but responsive use is key to unlocking their full potential.
Bibliography
- Chen, J. (2024). Demystifying Large Language Models. James Chen.
- Naveed, H., Khan, A. U., Qiu, S., Saqib, M., Anwar, S., Usman, M., Akhtar, N., Barnes, N., & Miani, A. (2024). A comprehensive overview of large language models. arXiv. https://arxiv.org/pdf/2307.06435
- Lu, C., Lu, C., Lange, R. T., Foerster, J., Clune, J., & Ha, D. (2024). The AI scientist: Towards fully automated open-ended scientific discovery. arXiv. https://arxiv.org/pdf/2408.06292
- Heaven, W. D. (2024). Large language models: Amazing, but nobody knows why. Technology Review. https://www.technologyreview.com/2024/03/04/1089403/large-language-models-amazing-but-nobody-knows-why/
- Stöffelbauer, A. (2023). How large language models work. Medium. https://medium.com/data-science-at-microsoft/how-large-language-models-work-91c362f5b78f
Written by:
Alejandro Castillo
FullStack Engineer
Country: Costa Rica