Seattle Skeptics on AI

Evaluating Fine-Tuned LLMs: A Practical Guide to Measurement Protocols
Evaluating Fine-Tuned LLMs: A Practical Guide to Measurement Protocols

Tamara Weed, Apr, 7 2026

Learn how to measure the success of your fine-tuned LLMs. We cover ROUGE, LLM-as-a-Judge, HELM benchmarks, and practical protocols for safety and accuracy.

Categories:

AI Ethics Frameworks for Generative AI: A Practical Guide to Responsible AI
AI Ethics Frameworks for Generative AI: A Practical Guide to Responsible AI

Tamara Weed, Apr, 6 2026

Learn how to implement AI ethics frameworks for generative AI. Move from vague principles to technical practices, bias mitigation, and regulatory compliance.

Categories:

Privacy and Security Risks of Distilled LLMs: A Guide for Secure Deployment
Privacy and Security Risks of Distilled LLMs: A Guide for Secure Deployment

Tamara Weed, Apr, 5 2026

Explore the hidden privacy and security risks of distilled LLMs. Learn why model compression doesn't stop PII leaks and how to use Intel TDX to secure your AI deployment.

Categories:

Federated Learning for LLMs: How to Train AI Without Centralizing Data
Federated Learning for LLMs: How to Train AI Without Centralizing Data

Tamara Weed, Apr, 4 2026

Learn how Federated Learning enables training Large Language Models (LLMs) across decentralized data sources to ensure privacy and bypass data centralization.

Categories:

Emergent Capabilities in Generative AI: What Works and What Remains Unclear
Emergent Capabilities in Generative AI: What Works and What Remains Unclear

Tamara Weed, Apr, 1 2026

Exploring emergent capabilities in Generative AI: definition, examples like chain-of-thought, the 'mirage' debate, and safety implications for 2026.

Categories:

Layer Dropping and Early Exit Techniques for Faster Large Language Models
Layer Dropping and Early Exit Techniques for Faster Large Language Models

Tamara Weed, Mar, 31 2026

Explore how layer dropping and early exit techniques accelerate Large Language Model inference, reducing latency and costs without sacrificing accuracy.

Categories:

API Gateways and Service Meshes in Modern Microservices Architecture
API Gateways and Service Meshes in Modern Microservices Architecture

Tamara Weed, Mar, 30 2026

Explore the distinct roles of API Gateways and Service Meshes in modern microservices architecture, including performance comparisons and implementation strategies for 2026.

Categories:

Addressing Hallucinations in Generative AI: Practical Mitigation Strategies for 2026
Addressing Hallucinations in Generative AI: Practical Mitigation Strategies for 2026

Tamara Weed, Mar, 29 2026

Explore why AI hallucinations happen and learn practical strategies like RAG and RLHF to reduce factual errors in generative systems.

Categories:

Beyond BLEU and ROUGE: Semantic Metrics for LLM Output Quality
Beyond BLEU and ROUGE: Semantic Metrics for LLM Output Quality

Tamara Weed, Mar, 28 2026

Traditional metrics like BLEU fail to capture LLM meaning. Learn why semantic metrics like BERTScore and LLM-as-a-Judge provide accurate quality assessment for modern AI deployments.

Categories:

Global Teams Shipping Faster: Vibe Coding Use Cases in Distributed Organizations
Global Teams Shipping Faster: Vibe Coding Use Cases in Distributed Organizations

Tamara Weed, Mar, 27 2026

Discover how vibe coding transforms global team productivity by turning natural language into executable code. Learn about real-world use cases, velocity gains, and infrastructure needs.

Categories:

How Positional Information Enables Word Order Understanding in Large Language Models
How Positional Information Enables Word Order Understanding in Large Language Models

Tamara Weed, Mar, 26 2026

Learn how positional encoding solves the word order problem in Transformers. We explore absolute, relative, and rotary methods, recent research findings, and future trends.

Categories:

Enterprise Knowledge Management with LLMs: Building Internal Q&A Systems
Enterprise Knowledge Management with LLMs: Building Internal Q&A Systems

Tamara Weed, Mar, 25 2026

Explore how Large Language Models transform enterprise knowledge management by turning static documents into dynamic Q&A systems. Learn about RAG architecture, security challenges, and implementation costs.

Categories: