<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Transformer on Jeanphilo Blog</title><link>https://shio-chan-dev.github.io/jeanblog/zh/tags/transformer/</link><description>Recent content in Transformer on Jeanphilo Blog</description><generator>Hugo -- 0.159.2</generator><language>zh-cn</language><lastBuildDate>Sun, 25 Jan 2026 20:08:41 +0800</lastBuildDate><atom:link href="https://shio-chan-dev.github.io/jeanblog/zh/tags/transformer/index.xml" rel="self" type="application/rss+xml"/><item><title>Attention Is All You Need：Transformer 的核心算法与工程落地</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/attention-is-all-you-need/</link><pubDate>Sun, 25 Jan 2026 20:08:41 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/attention-is-all-you-need/</guid><description>从算法抽象、复杂度与工程约束出发，解释 Transformer 如何用注意力替代递归与卷积，并给出可运行示例与选型指南。</description></item><item><title>Self-Attention 计算公式与 Softmax 数值稳定：从推导到工程实现</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/self-attention-softmax-formula-and-stability/</link><pubDate>Sun, 25 Jan 2026 12:50:33 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/self-attention-softmax-formula-and-stability/</guid><description>用公式与可运行示例讲清 Self-Attention 的计算流程、softmax 的数值问题与工程实现要点。</description></item><item><title>CNN、RNN、LSTM 与 Transformer 的区别与适用场景</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/architecture/cnn-rnn-lstm-transformer-comparison/</link><pubDate>Sat, 24 Jan 2026 16:28:18 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/architecture/cnn-rnn-lstm-transformer-comparison/</guid><description>从依赖路径长度与资源复杂度两个核心概念出发，系统对比 CNN、RNN、LSTM 与 Transformer，并给出可运行示例与工程选型步骤。</description></item><item><title>ViT 结构描述：从 Patch Embedding 到 Transformer 编码器</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/vision/vit-architecture-overview/</link><pubDate>Sat, 24 Jan 2026 16:25:35 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/vision/vit-architecture-overview/</guid><description>系统讲清 ViT 的结构组件、工作流程与工程实践，并给出最小 PyTorch 示例。</description></item><item><title>Transformer 中可以用 BatchNorm 吗？</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/batchnorm-in-transformer/</link><pubDate>Sat, 24 Jan 2026 16:24:03 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/batchnorm-in-transformer/</guid><description>讨论 Transformer 使用 BatchNorm 的可行性、限制与工程取舍，并给出最小示例。</description></item><item><title>BN 与 LN 的区别：训练稳定性与工程取舍</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/batchnorm-vs-layernorm/</link><pubDate>Sat, 24 Jan 2026 16:23:47 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/batchnorm-vs-layernorm/</guid><description>对比 BatchNorm 与 LayerNorm 的原理、适用场景与工程代价，并提供最小 PyTorch 示例。</description></item><item><title>为什么注意力要除以 √(d_k)：从数值稳定到工程收益</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/why-scale-attention-by-sqrt-dk/</link><pubDate>Sat, 24 Jan 2026 16:22:25 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/why-scale-attention-by-sqrt-dk/</guid><description>解释注意力中 QK^T 为何需要除以 √(d_k)，并给出最小 PyTorch 示例与工程场景。</description></item><item><title>残差连接的作用：为什么深度网络离不开它</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/residual-connection-role/</link><pubDate>Sat, 24 Jan 2026 16:22:22 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/residual-connection-role/</guid><description>解释残差连接在深度网络中的作用与原理，并提供最小可运行示例。</description></item><item><title>Attention 的复杂度与为什么需要位置编码</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/attention-complexity-and-positional-encoding/</link><pubDate>Sat, 24 Jan 2026 16:21:51 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/attention-complexity-and-positional-encoding/</guid><description>解释注意力的时间/空间复杂度，并说明位置编码对序列建模的必要性，含最小示例。</description></item><item><title>为什么使用多头注意力机制：能力、稳定性与工程取舍</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/why-multi-head-attention/</link><pubDate>Sat, 24 Jan 2026 16:20:59 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/why-multi-head-attention/</guid><description>用 ACERS 框架解释多头注意力的必要性、核心原理与工程场景，并给出最小可运行示例。</description></item><item><title>Transformer 结构描述：从编码器到解码器</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/transformer-architecture-overview/</link><pubDate>Sat, 24 Jan 2026 16:18:19 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/transformer-architecture-overview/</guid><description>用 ACERS 框架讲清 Transformer 结构、模块职责与工程场景，并给出最小可运行示例。</description></item><item><title>为什么 GPT 是 Decoder-Only：自回归生成的最佳形态</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/why-gpt-decoder-only/</link><pubDate>Sat, 24 Jan 2026 16:15:34 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/why-gpt-decoder-only/</guid><description>解释 GPT 选择 decoder-only 结构的原因，并与 encoder-only / encoder-decoder 做工程对比。</description></item><item><title>LLaMA 中 RMSNorm 相比 LayerNorm 的优势</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/rmsnorm-vs-layernorm-llama/</link><pubDate>Sat, 24 Jan 2026 15:52:58 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/llm/rmsnorm-vs-layernorm-llama/</guid><description>从公式、复杂度与工程实践出发，解析 LLaMA 选择 RMSNorm 的原因，并给出最小 PyTorch 示例。</description></item><item><title>Self-Attention vs Cross-Attention：机制、差异与工程应用</title><link>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/self-attention-vs-cross-attention/</link><pubDate>Sat, 24 Jan 2026 15:44:12 +0800</pubDate><guid>https://shio-chan-dev.github.io/jeanblog/zh/ai/attention/self-attention-vs-cross-attention/</guid><description>用 ACERS 框架讲清 self-attention 与 cross-attention 的核心差异、公式与工程场景。</description></item></channel></rss>