Subjects
Activities
Tools
20 lessons ยท 7th Grade
Modern AI assistants use a pipeline: ASR for speech, an LLM for understanding and generation, and TTS for voice output. Each component uses different AI models.
AI agents can use tools โ browsing the web, running code, calling APIs. They break complex tasks into steps and execute them autonomously.
Strategies to reduce hallucinations include RAG, constrained generation, confidence scoring, and fact-checking pipelines that verify AI outputs.
LLMs have a context window limiting how much text they process at once. Techniques like sliding windows and summarization extend effective memory.
Evaluating AI quality uses metrics like BLEU (translation), perplexity (fluency), and human evaluation. No single metric captures everything.
Safety measures include content filters, output monitoring, rate limiting, and red-teaming where experts try to find vulnerabilities before public release.
Some AI models are open-source (anyone can use them) while others are closed (proprietary). This affects transparency, cost, and who controls AI.
Developers access AI through APIs โ interfaces that let programs send prompts and receive responses. This enables building custom AI applications.
Running LLMs requires expensive GPU clusters. Training GPT-4 cost over $100 million. Companies balance capability with computational cost.
Future assistants will be more personalized, proactive, and capable. They may anticipate needs, maintain long-term memory, and collaborate across tasks.
With APIs and frameworks, you can build simple AI tools. Start with prompt engineering, then learn about APIs, and eventually model fine-tuning.
LLMs like GPT are trained on billions of text tokens. They learn statistical patterns in language to predict and generate coherent text.
AI assistants use LLMs, ASR, and TTS in a pipeline. Understanding their architecture, limitations, and safety measures helps us use them effectively.
Before processing text, LLMs split it into tokens (word pieces) and convert them to number vectors called embeddings that capture meaning.
Attention lets the model weigh which words matter most for predicting the next token. This allows understanding long-range dependencies in text.
Training is when the model learns from data (expensive, slow). Inference is when it generates responses for users (fast). Most users only see inference.
After pre-training, models are fine-tuned with human feedback (RLHF). Humans rank responses, teaching the model to prefer helpful, safe outputs.
Advanced prompting includes few-shot examples, chain-of-thought reasoning, role-playing, and system prompts. Each technique guides the model differently.
RAG combines LLMs with external knowledge bases. The model retrieves relevant documents, then uses them to generate grounded, factual responses.
Modern AI can process text, images, audio, and video together. Multi-modal models understand a photo and discuss it, or transcribe and summarize a video.
Your cart is empty
Browse our shop to find activities your kids will love