Atharva Presentation
Atharva Presentation
LARGE
LANGUAGE
MODEL
Present by Atharva Virkar
FLOW
1 Introduction 6 What’s the Hype
RNN reads the sentence word by word and tries to remember what came
before.
It might forget the early context ("I wanted to follow up") and focus only on
EXAMPLE OF recent words like “we had last”.
Because it has a limited memory, by the time we reach "last...", the memory
WORKING of “meeting” is faint or lost. So, it might predict something vague like "time".
RNN fails when the important clue was far back in the sentence.
RNN & LSTM LSTMs take on it will be :
LSTM also reads word by word, but it has gates that help it remember that
the sentence started with “follow up on the meeting”.
It has 3 gates : Forget gate, Input gate, & Output gate. Which decides what
to ignore, what to remember and what to show to the next step.
So even if “meeting” was several words ago, it remembers that it’s important.
It is more likely to predict "week" or "Friday" because it knows we’re
referring to a meeting.
RNN Structure: Think of it as a Relay Race
RNN processes the sentence one word at a time.
Each word passes its meaning to the next like a baton in a relay race.
There's one shared memory (hidden state) passed along the chain.
Early information (like "meeting") has to travel through every step to reach the end.
The message gets weaker with each step — like a story passed from person to person. So by the time we
reach "last...", the memory of “meeting” is faint or lost.
To summarise it, we can say RNN is a one linear path, memory passed forward step-by-step, prone to
forgetting earlier parts.
STRUCTURE OF RNN
Maintains a separate memory cell, like a notepad, so it can hold onto key facts.
It remembers important words (like “meeting”) for a long time without fading. This helps it
make better decisions even many steps later.
To summarise it, we can say LSTM is a Step-by-step model with memory management, uses STRUCTURE OF LSTM
gates to decide what to remember or forget like a person jotting down key info.
Transformer doesn’t pass memory step by step.
Unlike earlier models, Transformers uses self-attention mechanism which can see
the whole sentence at once, making them better at understanding meaning.
While LLMs are the underlying engine trained to predict and generate
language, Gen AI refers to the broader suite of tools and systems using
these models to create new, original outputs.
Industry Transformation
Automating Complex Tasks: AI agents can autonomously
perform tasks that traditionally required human
HYPE ?
Enhancing Productivity: By handling routine and time-
consuming activities, AI agents free up human workers to
focus on strategic and creative endeavors.
Tools:
Google Sheets (read) – Agent reads data (e.g., get tasks or meeting
list)
Google Sheets1 (update) – Agent writes new info (e.g., log a
“This diagram represents an AI agent that acts like a smart digital assistant. When someone completed task)
sends a message, this assistant understands it, remembers useful information, and uses
external tools like Google Sheets or Google Calendar to take actions — automatically.” Google Calendar (create event) – Agent creates meetings
automatically
REAL LIFE ADAPTATION
& EXAMPLE
A platform by Jack Dorsey's Block that lets businesses build and Microsoft uses AutoGen (multi-agent) setup internally
deploy AI agents to automate tasks like customer support, and externally.
document analysis, and workflow management. One agent reads incoming email, another plans the task
Agents are powered by LLMs and integrated with tools (email, (e.g., schedule meeting, approve document).
web search, databases). Another executes the task or generates a reply and
Example: A "Legal AI Agent" that reads 20-page contracts, finally, a checker agent reviews and sends.
highlights key clauses, and sends compliance reports Enterprise Use: Auto-pilot mode for inboxes, especially
automatically. for executives and managers.
DnT Infotech
THANK YOU
for your time and attention