MetaのDiT:Transformerアーキテクチャが拡散モデルの未来をどう変えるか

GitHub April 2026
⭐ 8516
Source: GitHubtransformer architectureArchive: April 2026
Metaのオープンソースプロジェクト「Diffusion Transformer(DiT)」は、生成AIにおける根本的なアーキテクチャの転換を意味します。拡散モデルの畳み込みU-Netバックボーンを純粋なTransformerに置き換えることで、DiTは前例のないスケーラビリティを示し、モデル性能が規模に応じて予測可能に向上します。
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The release of DiT by Meta's Fundamental AI Research (FAIR) team marks a pivotal moment in the evolution of generative image models. For years, the diffusion process for image synthesis has been dominated by U-Net architectures, a convolutional neural network design that excelled at capturing local spatial features. DiT challenges this orthodoxy by proving that a Transformer—the architecture that revolutionized natural language processing—can not only work for diffusion but can do so with superior scaling properties. The core innovation lies in its 'patchify' approach, where input images are broken into patches and treated as sequences of tokens, analogous to words in a sentence. This allows the model to leverage the global attention mechanism of Transformers, capturing long-range dependencies across the entire image.

The significance of DiT extends beyond a simple architecture swap. Its research paper, 'Scalable Diffusion Models with Transformers,' provides rigorous empirical evidence of what the AI community calls 'scaling laws.' As the model size (measured in parameters and attention heads) and training compute increase, the Fréchet Inception Distance (FID) score—a key metric for image quality—improves in a smooth, predictable manner. This predictability is a holy grail for AI research, as it allows for more confident investment in larger models. While the current public implementation is focused on class-conditional image generation on datasets like ImageNet, its design principles are being rapidly adopted and extended. The project's clean, modular PyTorch codebase has made it a favorite among researchers experimenting with next-generation video generation, 3D asset creation, and even audio synthesis, positioning DiT as a foundational codebase for the next wave of generative AI.

Technical Deep Dive

At its heart, DiT re-imagines the diffusion denoising process through the lens of sequence modeling. The traditional U-Net in models like Stable Diffusion operates on the noisy latent image directly, using convolutional layers to progressively refine it. DiT, however, first encodes the noisy input into a sequence. This is achieved through a patchification layer. For a 256x256 image and a patch size of 2, the model creates a sequence of 128x128 = 16,384 tokens. Each token is a small, flattened representation of a 2x2 pixel block. This sequence is then processed by a standard Transformer encoder.

The DiT block is augmented with two critical conditioning mechanisms that guide the generation:
1. Adaptive Layer Norm (adaLN): Instead of standard LayerNorm, DiT uses a conditional version where the scaling and shifting parameters are dynamically predicted by a small network based on the diffusion timestep `t`. This tells the Transformer "how noisy" the current input is.
2. Conditional Class Embeddings: For class-conditional generation, the class label is embedded and injected, often through cross-attention or additional modulation layers, steering the model to produce images of a specific category.

After processing by a stack of these modified Transformer blocks, the sequence is decoded back into a noise prediction (or image prediction, depending on the formulation) using a final linear layer that reconstructs the patches.

The most compelling data from the DiT paper concerns its scalability. The team trained models ranging from 600M to 6B parameters.

| Model Variant | Parameters (G) | GFLOPs (forward pass) | FID-50K (ImageNet 256x256) |
|---|---|---|---|
| DiT-XL/2 | ~3.0 | ~119 | 9.62 |
| DiT-XL/2 (cfg) | ~3.0 | ~119 | 2.27 |
| DiT-L/2 | ~1.2 | ~76 | 12.24 |
| U-Net (ADM) | ~0.7 | ~281 | 10.94 |

*Note: "cfg" denotes classifier-free guidance, a technique to sharpen generation quality. GFLOPs measured for a 256x256 image.*

Data Takeaway: The table reveals two critical insights. First, the largest DiT model (DiT-XL/2) with guidance achieves a state-of-the-art FID of 2.27, significantly outperforming the comparable U-Net-based ADM model. Second, and more importantly, DiT's computational footprint (GFLOPs) is more efficient for the Transformer-based forward pass compared to the convolutional U-Net for a similar parameter count, highlighting a potential efficiency advantage at scale.

The official `facebookresearch/dit` GitHub repository provides a well-documented codebase for training and inference. Key files include `models.py` containing the core DiT block definitions, and `train.py` with the essential training loop. The community has built upon this foundation; for instance, the `Projected_DiT` repository explores integrating text conditioning via a CLIP text encoder, bridging DiT with text-to-image models like Stable Diffusion.

Key Players & Case Studies

The development of DiT is part of a broader strategic competition to define the foundational architecture of generative AI. Meta's FAIR team, led by researchers like William Peebles (co-author of the DiT paper), is making a clear bet on the unifying power of the Transformer. This aligns with Meta's broader push for architectures like the Segment Anything Model (SAM) and its Llama language models, favoring scalable, general-purpose designs.

OpenAI's DALL-E 3 and the underlying technology of Sora, its video generation model, are also suspected to leverage Transformer-based diffusion or diffusion-like processes. While not open-sourced, the quality and coherence of their outputs suggest heavy investment in scalable, attention-based architectures for visual data. Stability AI, the company behind the U-Net-based Stable Diffusion, is now actively exploring Transformer integrations. Its Stable Diffusion 3 medium-tier model explicitly incorporates a "Multimodal Diffusion Transformer (MMDiT)," a direct acknowledgment of DiT's influence.

| Entity | Core Architecture | Key Product/Model | Strategic Position on DiT/Transformers |
|---|---|---|---|
| Meta (FAIR) | Transformer (DiT) | DiT codebase, Emu | Open-source research leader; betting on unified Transformer stack for all modalities. |
| OpenAI | Likely Transformer-hybrid | DALL-E 3, Sora | Closed, product-focused; scaling private models for commercial advantage. |
| Stability AI | U-Net → Transformer-hybrid | Stable Diffusion 3 | Pragmatic adapter; integrating Transformer ideas into established U-Net ecosystem for incremental improvement. |
| Google DeepMind | Diverse (U-Net, Transformer) | Imagen, VideoPoet | Research-heavy; explores multiple paths (e.g., Imagen uses T5 text + U-Net, VideoPoet uses language model backbone). |

Data Takeaway: The competitive landscape shows a clear architectural convergence towards Transformers. Meta is the most purist advocate, while others are taking a hybrid or transitional approach. This table underscores that DiT is not an isolated project but a central reference point in an industry-wide architectural debate.

Industry Impact & Market Dynamics

DiT's primary impact is lowering the research and development barrier for next-generation generative models. Its open-source nature and clean implementation have spawned a wave of innovation. Startups and academic labs can now experiment with Transformer-based diffusion without building the core infrastructure from scratch. This accelerates progress in niche areas like scientific imagery generation, architectural design, and personalized media creation.

The scalability proven by DiT directly influences investment decisions. Venture capital and corporate R&D budgets are increasingly directed towards projects that demonstrate clear scaling laws. DiT provides a blueprint for how to scale visual generative models, making large-scale training projects more justifiable. We are already seeing this in the funding landscape.

| Company/Project | Core Tech | Recent Funding/Inference | Relevance to DiT Trend |
|---|---|---|---|
| Stability AI | Stable Diffusion (U-Net/MMDiT) | Raised ~$100M in 2023 | Pivoting to incorporate Transformer scalability to maintain competitiveness. |
| Midjourney | Proprietary (likely hybrid) | ~$200M+ estimated annual revenue | Closed system; must internally develop or adopt scalable architectures to improve quality. |
| Runway ML | Gen-2 Video Model | Raised $50M+ Series C | Video generation is a prime target for scalable DiT-like architectures for longer, coherent sequences. |
| Various AI Chip Startups | Hardware (e.g., Groq, SambaNova) | Billions in collective funding | DiT's Transformer core aligns perfectly with hardware optimized for dense matrix operations and attention, unlike U-Net's mixed ops. |

Data Takeaway: The funding and strategic moves indicate that scalability is the new battleground. DiT's architecture favors hardware and software stacks designed for large-scale Transformers, creating a symbiotic relationship between AI model research and specialized hardware development. Companies that fail to architect for scalable training risk being left behind.

The model also pushes the industry towards multimodal unification. A pure Transformer backbone for vision creates a pathway to models where the same core architecture processes text, image, audio, and video tokens. This simplifies training infrastructure and could lead to more coherent cross-modal understanding and generation, a key step towards more general AI assistants.

Risks, Limitations & Open Questions

Despite its promise, DiT faces significant hurdles. The most immediate is astronomical training cost. Training the largest DiT models from scratch on high-resolution ImageNet requires thousands of high-end GPUs for weeks. This centralizes power in the hands of a few well-resourced corporations, potentially stifling open innovation. While the code is open, the compute is not.

Inference latency remains an open question. While the forward-pass GFLOPs may be favorable, the sequential nature of the diffusion process (often 50+ steps) with a large Transformer can be slower than optimized, distilled U-Net models in real-time applications. Techniques like latent consistency models or progressive distillation need to be effectively adapted to the DiT architecture.

The current DiT model is primarily class-conditional, not text-conditional. Bridging the gap to robust, creative text-to-image generation requires effectively integrating large language models or text encoders. Projects like `Projected_DiT` are steps in this direction, but achieving the prompt fidelity and compositional understanding of models like SDXL or DALL-E 3 with a pure DiT backbone is an ongoing challenge.

Ethically, DiT's scalability amplifies existing concerns about generative AI: misinformation through hyper-realistic synthetic media, copyright infringement via training data, and environmental impact from massive compute consumption. The architecture itself is neutral, but its efficiency at scale makes these societal impacts more potent and urgent to address.

AINews Verdict & Predictions

DiT is a foundational breakthrough, not merely an incremental improvement. It successfully transplants the scaling laws of language modeling into the domain of image generation, providing a clear roadmap for future progress. Its pure Transformer design is the correct long-term bet for the industry, as it paves the way for truly unified multimodal models.

We predict the following developments over the next 18-24 months:
1. The Demise of Pure U-Nets: New state-of-the-art image and video generation models will overwhelmingly adopt Transformer or Transformer-hybrid backbones. U-Nets will be relegated to specialized, efficiency-critical edge applications.
2. The Rise of "DiT-for-X": The core DiT design pattern will be successfully applied to 3D generation (3D DiT), video generation (Video DiT), and audio/music generation, with several open-source repos reaching prominence in each domain.
3. Text-to-Image Dominance: Within 12 months, an open-source, text-conditioned DiT model will match or exceed the quality of the current Stable Diffusion XL, becoming the new community standard for customizable image generation.
4. Hardware Co-design: The next generation of AI accelerators from Nvidia, AMD, and startups will feature architectural optimizations that further favor the DiT-style computation graph over traditional CNN/U-Net workloads.

The critical watchpoint is not if DiT will be widely adopted, but how quickly the ecosystem solves the text-conditioning and inference speed challenges. The organization or consortium that successfully deploys a scalable, fast, and open text-to-image DiT model will capture significant mindshare and dictate the pace of open-source generative AI development for the foreseeable future.

More from GitHub

Microsoft Data Formulator:自然言語はドラッグ&ドロップ分析に取って代わるか?Microsoft's Data Formulator, now available on GitHub with over 15,000 stars, represents a paradigm shift in how humans iAndrej Karpathy の GitHub スキルツリー:AI の信頼性を再定義する遊び心あふれる履歴書The GitHub repository 'vtroiswhite/andrej-karpathy-skills' has captured the AI community's imagination by presenting AndHotkey Helper: プラグイン設定の混乱を解決するObsidianプラグインObsidian's extensibility is its greatest strength, but also its Achilles' heel. As users accumulate plugins for tasks liOpen source hub1708 indexed articles from GitHub

Related topics

transformer architecture27 related articles

Archive

April 20263042 published articles

Further Reading

OpenAIの改良版DDPM:分散学習とノイズスケジューリングが拡散モデルを再定義する方法OpenAIは、改良版Denoising Diffusion Probabilistic Models(Improved DDPM)の公式実装をオープンソース化し、最先端の画像生成のための明確でプロダクショングレードのコードベースを提供しまMetaのImageBind、6つのモダリティに対応する汎用AI埋め込み空間を構築Meta AIのImageBindプロジェクトは、マルチモーダル人工知能におけるパラダイムシフトを象徴しています。画像、テキスト、オーディオ、深度、熱、IMUデータという6つの異なるデータタイプを結びつける単一の統合埋め込み空間を構築するこガーデンスキル:ConardLiのオープンソースAIツールキットが開発者ワークフローを変革ConardLiのGarden Skillsは急速に成長しているオープンソースリポジトリで、Webデザイン、知識検索、画像生成のためのモジュール式AIツールを提供しています。4,161のスターと1日あたり540の急増を記録し、その実用性で開EG3D:NVIDIAのトリプレーン革命が3D認識生成AIを再定義NVIDIA ResearchのEG3Dは、革新的なトリプレーン表現を活用し、純粋な3Dアプローチよりも劇的に低い計算コストで高解像度かつ視点一貫性のある合成を実現する、3D認識生成AIにおける重要なアーキテクチャとして登場しました。本記事

常见问题

GitHub 热点“Meta's DiT: How Transformer Architecture Is Reshaping the Future of Diffusion Models”主要讲了什么?

The release of DiT by Meta's Fundamental AI Research (FAIR) team marks a pivotal moment in the evolution of generative image models. For years, the diffusion process for image synt…

这个 GitHub 项目在“DiT vs Stable Diffusion architecture comparison”上为什么会引发关注?

At its heart, DiT re-imagines the diffusion denoising process through the lens of sequence modeling. The traditional U-Net in models like Stable Diffusion operates on the noisy latent image directly, using convolutional…

从“How to train a Diffusion Transformer from scratch”看,这个 GitHub 项目的热度表现如何?

当前相关 GitHub 项目总星标约为 8516,近一日增长约为 0,这说明它在开源社区具有较强讨论度和扩散能力。