Whilst Artificial Intelligence (AI) has been around for a while, it has become ubiquitous over the past two years since the launch of ChatGPT. It is no longer just the realm of tech companies finding niche applications working in the background. With Meta’s AI embedded into Whatsapp chats, Google’s into Android phones and Microsoft’s into its Office suite, AI has become a part of the common man’s daily life. What changed? The specific technology behind Generative AI, the likes of ChatGPT is known as Large Language Model (LLM). Whilst most of us use the internet without knowing exactly how it works, knowing the science behind LLM, atleast in layperson terms, is likely to help us extract better from this technology. This blog does a great job of explaining it in simple English or rather Math, without getting into the complexities.

“…at their core, LLMs are nothing more than fancy probability engines.

Probability is a tool for quantifying randomness and uncertainty. Having a probabilistic nature is the source of an LLM’s power… and its unpredictability. It’s what makes it possible to generate novel, creative, actionable responses, and also makes LLMs very difficult to train or debug.

Whether you’re a layperson, practitioner, or researcher, the entire process of working with LLMs is an exercise in forming and manipulating their latent probability distributions into giving you the outputs you want.”

The blog gives a quick refresher on probability distributions, something that most of us would have studied in high school math, before going on to explain how LLM’s use probability of the occurrence of a word to generate text which is likely to make sense and give us the outputs we seek. Given this context, unsurprisingly, the most common application of LLM so far has been Chatbots. He gives an intuitive sense of how LLMs drive chat:

“Chat interfaces have become the dominant way for most people to interact with LLMs, capturing the public imagination and showcasing these models’ capabilities. But how do we get from generating individual words to engaging in full-fledged conversations? The answer lies in cleverly applying the principles we’ve discussed so far.

Here’s how it works:

A. When you start a chat, your initial message becomes the first piece of context.
B. The model generates a response based on this context, just as we described earlier.
C. For your next message, the model doesn’t just look at what you’ve just said. Instead, it considers everything that’s been said so far – your initial message, its first response, and your new message.
D. This process repeats for each turn of the conversation. The context grows longer, incorporating each new message and response.
This approach allows the model to maintain consistency and context throughout a conversation. It can refer back to earlier parts of the chat, answer follow-up questions, and generally behave in a way that feels more like a coherent dialogue than isolated text generation.

However, this method also introduces some challenges:

1. Context Length Limits: LLMs have a maximum amount of text they can process at once (often referred to as the “context window”). For very long conversations, the earliest parts might get cut off when this limit is reached.
2. Computational Cost: As the conversation grows, generating each new response requires processing more and more text, which can slow down the model’s responses and increase computational costs.
3. Consistency vs. Creativity: The model might become overly constrained by the conversation history, potentially leading to less diverse or creative responses over time.
Despite these challenges, this simple yet effective approach to chat is what powers the conversational AI interfaces we interact with daily. By treating the entire conversation as a growing context for probabilistic text generation, LLMs can engage in surprisingly coherent and context-aware dialogues.”

Whether you are a user of AI or an analyst trying to assess the disruptive potential of this technology, the blog is a compelling and easy read to understand what’s tipped to be the next big thing, as seen with the tens of billions of dollars already being invested by the world’s biggest technology companies.

“Microsoft, Meta and Alphabet (parent of Google) forms a part of Marcellus’ Global Compounders Portfolio, a strategy offered by GIFT City branch of Marcellus Investment Managers Private Limited. Hence, we as Marcellus, our immediate relatives and our clients may have interest and stakes in the mentioned stock. The stocks mentioned are for educational purposes only and not recommendatory.”

If you want to read our other published material, please visit https://marcellus.in/blog/

Note: The above material is neither investment research, nor financial advice. Marcellus does not seek payment for or business from this publication in any shape or form. The information provided is intended for educational purposes only. Marcellus Investment Managers is regulated by the Securities and Exchange Board of India (SEBI) and is also an FME (Non-Retail) with the International Financial Services Centres Authority (IFSCA) as a provider of Portfolio Management Services. Additionally, Marcellus is also registered with US Securities and Exchange Commission (“US SEC”) as an Investment Advisor.



2024 © | All rights reserved.

Privacy Policy | Terms and Conditions