<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>LLM on Arshad Siddiqui</title><link>https://arshadhs.github.io/tags/llm/</link><description>Recent content in LLM on Arshad Siddiqui</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 15 Dec 2025 10:55:52 +0100</lastBuildDate><atom:link href="https://arshadhs.github.io/tags/llm/index.xml" rel="self" type="application/rss+xml"/><item><title>Generative AI</title><link>https://arshadhs.github.io/docs/ai/genai/</link><pubDate>Mon, 15 Dec 2025 10:55:52 +0100</pubDate><guid>https://arshadhs.github.io/docs/ai/genai/</guid><description>&lt;h1 id="generative-ai">
 Generative AI
 
 &lt;a class="anchor" href="#generative-ai">#&lt;/a>
 
&lt;/h1>
&lt;p>&lt;strong>Generative Artificial Intelligence (GenAI)&lt;/strong> refers to a class of AI systems that can &lt;strong>generate new content&lt;/strong> such as text, images, audio, video, or code, rather than only making predictions or classifications.&lt;/p>
&lt;p>GenAI systems learn &lt;strong>patterns and representations from large datasets&lt;/strong> and use them to produce &lt;strong>novel outputs&lt;/strong> that resemble the data they were trained on.&lt;/p>
&lt;hr>
&lt;h2 id="how-generative-ai-differs-from-traditional-ai">
 How Generative AI Differs from Traditional AI
 
 &lt;a class="anchor" href="#how-generative-ai-differs-from-traditional-ai">#&lt;/a>
 
&lt;/h2>
&lt;table>
 &lt;thead>
 &lt;tr>
 &lt;th>Traditional AI&lt;/th>
 &lt;th>Generative AI&lt;/th>
 &lt;/tr>
 &lt;/thead>
 &lt;tbody>
 &lt;tr>
 &lt;td>Predicts or classifies&lt;/td>
 &lt;td>Generates new content&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>Task-specific models&lt;/td>
 &lt;td>General-purpose models&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>Fixed outputs&lt;/td>
 &lt;td>Open-ended outputs&lt;/td>
 &lt;/tr>
 &lt;tr>
 &lt;td>Often rule-based&lt;/td>
 &lt;td>Data-driven and probabilistic&lt;/td>
 &lt;/tr>
 &lt;/tbody>
&lt;/table>
&lt;hr>
&lt;h2 id="core-idea-of-generative-ai">
 Core Idea of Generative AI
 
 &lt;a class="anchor" href="#core-idea-of-generative-ai">#&lt;/a>
 
&lt;/h2>

&lt;blockquote class='book-hint '>
 &lt;p>&lt;strong>Instead of learning “what label to assign”, Generative AI learns “how data is structured” and then creates new data following that structure.&lt;/strong>&lt;/p></description></item><item><title>Foundation Models</title><link>https://arshadhs.github.io/docs/ai/genai/foundation-model/</link><pubDate>Sun, 14 Dec 2025 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/genai/foundation-model/</guid><description>&lt;h1 id="foundation-model">
 Foundation Model
 
 &lt;a class="anchor" href="#foundation-model">#&lt;/a>
 
&lt;/h1>
&lt;p>AI models trained on massive datasets to perform a wide range of tasks with minimal fine-tuning.&lt;/p>
&lt;ul>
&lt;li>
&lt;p>are large deep learning neural networks&lt;/p>
&lt;/li>
&lt;li>
&lt;p>are large AI models trained on &lt;strong>massive and diverse datasets&lt;/strong> (text, images, audio, or multiple modalities).&lt;/p>
&lt;/li>
&lt;li>
&lt;p>Contain &lt;strong>millions or billions of parameters&lt;/strong>.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>designed to perform a &lt;strong>broad range of general tasks&lt;/strong>&lt;/p>
&lt;/li>
&lt;li>
&lt;p>designed for &lt;strong>general-purpose intelligence&lt;/strong>, not a single task.&lt;/p>
&lt;/li>
&lt;li>
&lt;p>acts as &lt;strong>base models&lt;/strong> for building specialised AI applications&lt;/p></description></item><item><title>LLM - Model</title><link>https://arshadhs.github.io/docs/ai/genai/llm/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/genai/llm/</guid><description>&lt;h1 id="llm--large-language-model">
 LLM – Large Language Model
 
 &lt;a class="anchor" href="#llm--large-language-model">#&lt;/a>
 
&lt;/h1>
&lt;p>Large Language Models (LLMs) are &lt;strong>advanced AI systems&lt;/strong> designed to process, understand, and generate &lt;strong>human-like text&lt;/strong>.&lt;/p>
&lt;p>They learn language by analysing &lt;strong>massive amounts of text data&lt;/strong>, discovering patterns in:&lt;/p>
&lt;ul>
&lt;li>
&lt;p>grammar&lt;/p>
&lt;/li>
&lt;li>
&lt;p>meaning&lt;/p>
&lt;/li>
&lt;li>
&lt;p>context&lt;/p>
&lt;/li>
&lt;li>
&lt;p>relationships between words and sentences&lt;/p>
&lt;/li>
&lt;li>
&lt;p>Built on &lt;strong>Deep Learning&lt;/strong>&lt;/p>
&lt;/li>
&lt;li>
&lt;p>Implemented using &lt;strong>Neural Networks&lt;/strong>&lt;/p>
&lt;/li>
&lt;li>
&lt;p>Based on &lt;strong>Transformers&lt;/strong>&lt;/p>
&lt;/li>
&lt;li>
&lt;p>Often combined with tools like:&lt;/p>
&lt;ul>
&lt;li>Retrieval (RAG)&lt;/li>
&lt;li>Agents&lt;/li>
&lt;li>External APIs&lt;/li>
&lt;li>Memory systems&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="what-makes-an-llm-special">
 What makes an LLM special?
 
 &lt;a class="anchor" href="#what-makes-an-llm-special">#&lt;/a>
 
&lt;/h2>
&lt;ul>
&lt;li>Built using &lt;strong>deep neural networks&lt;/strong>&lt;/li>
&lt;li>Trained on &lt;strong>very large datasets&lt;/strong> (books, articles, code, web text)&lt;/li>
&lt;li>Can perform many tasks &lt;strong>without task-specific training&lt;/strong>&lt;/li>
&lt;li>General-purpose language understanding, not single-task models&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="foundation-transformer-architecture">
 Foundation: Transformer Architecture
 
 &lt;a class="anchor" href="#foundation-transformer-architecture">#&lt;/a>
 
&lt;/h2>
&lt;p>LLMs are based on the &lt;strong>&lt;a href="https://arshadhs.github.io/docs/ai/deep-learning/transformer/">Transformer Architecture&lt;/a>&lt;/strong>, which allows models to understand &lt;strong>context and long-range dependencies&lt;/strong> in text.&lt;/p></description></item><item><title>Retrieval-Augmented Generation (RAG)</title><link>https://arshadhs.github.io/docs/ai/genai/rag/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://arshadhs.github.io/docs/ai/genai/rag/</guid><description>&lt;h1 id="retrieval-augmented-generation-rag">
 Retrieval-Augmented Generation (RAG)
 
 &lt;a class="anchor" href="#retrieval-augmented-generation-rag">#&lt;/a>
 
&lt;/h1>
&lt;p>&lt;strong>Retrieval-Augmented Generation (RAG)&lt;/strong> is a system design pattern that improves an LLM’s answers by:&lt;/p>
&lt;ol>
&lt;li>&lt;strong>Retrieving&lt;/strong> relevant information from an external knowledge source, and then&lt;/li>
&lt;li>&lt;strong>Augmenting&lt;/strong> the LLM prompt with that retrieved context before generating the final response.&lt;/li>
&lt;/ol>
&lt;p>RAG helps an LLM &lt;strong>look things up first&lt;/strong>, then &lt;strong>answer using evidence&lt;/strong>.&lt;/p>
&lt;hr>
&lt;h2 id="why-rag-is-useful">
 Why RAG is Useful
 
 &lt;a class="anchor" href="#why-rag-is-useful">#&lt;/a>
 
&lt;/h2>
&lt;p>RAG is commonly used when:&lt;/p>
&lt;ul>
&lt;li>Your knowledge is in &lt;strong>private documents&lt;/strong> (PDFs, policies, internal wiki)&lt;/li>
&lt;li>You need &lt;strong>up-to-date information&lt;/strong> (things not in the model’s training data)&lt;/li>
&lt;li>You want fewer &lt;strong>hallucinations&lt;/strong> by grounding answers in retrieved sources&lt;/li>
&lt;li>You want &lt;strong>traceability&lt;/strong> (show “where the answer came from”)&lt;/li>
&lt;/ul>
&lt;blockquote class="book-hint info">
&lt;p>RAG does not change the model weights.&lt;br>
It changes what the model &lt;em>sees&lt;/em> at inference time by adding retrieved context.&lt;/p></description></item></channel></rss>