we’re excited to share that you can now generate images directly with our Illustrious XL models on our official site: [illustrious-xl.ai]. We’ve launched a full image generation platform featuring high-res outputs, natural language prompting, and custom presets — plus, several exclusive models you won’t find on any other hub. Explore our updated model tiers and naming here: [Model Series]. Need help getting started? Check out our generation user guide: [ILXL Image Generation User Guide].
版本详情
Lumina 2
Illustrious-Lumina-v0.03
Before we dive into the details of 'Illustrious-Lumina-v0.03', we’re excited to share that you can now generate images directly with our Illustrious XL models on our official site: [illustrious-xl.ai](http://illustrious-xl.ai/).
We’ve launched a full image generation platform featuring high-res outputs, natural language prompting, and custom presets - plus, several exclusive models you won’t find on any other hub.
Explore our updated model tiers and naming here: [Model Series](https://www.illustrious-xl.ai/updates/20).
Need help getting started? Check out our generation user guide: [ILXL Image Generation User Guide](https://www.illustrious-xl.ai/updates/21).
## 1. Model Overview
- **Architecture**: **2 B parameters** DiT.
- **Text Encoder**: Pure LLM, **Gemma-2-2b **
- **Goal of this fork**: We test if the image backbone can learn illustration concepts **without** re‑training the LLM component.
**Illustrious-Lumina-v0.03** is experimental epoch of Lumina-2.0 based training session, to validate whether we would be able to achieve small DiT model just with LLM - to be trained as illustration-focused model.
The original model, is unfortunately bad at illustrations and lacked any of the knowledge - so the run focused on training abscent knowledges.
After 26,500 step, the model, Illustrious-Lumina-v0.03 has show successful fast adaptation toward the dataset.
However, please note that the original model is not good at illustrations, whileas our focus is only in illustrations - this would take a while to reach the certain level.
The examples are ready in [Blog post](https://www.illustrious-xl.ai/blog).
To test the model, please refer to the [huggingface space](https://huggingface.co/spaces/AngelBottomless/Lumina-Illustrious-v0.03)
If you prefer to run model locally, please use the **pth file** with [official installation guide](https://github.com/OnomaAI/Illustrious-Lumina).
**The safetensors file is meant to only "contain the weights" - for comfyui-compatible format, we will try to prepare it as soon as possible.**
## 2. Training Setup
| Item | Value |
|------|-------|
| Images Seen Total | 22 M image–text pairs |
| Steps | 26 500 |
| Global batch | 768 |
| Resolution | 1024, 256 |
| Checkpoint | `Illustrious_Lumina_2b_22100_ema_unified_fp32.safetensors` |
The model has seen 22M image-text pairs. To accelerate the training, multi-resolution training was utilized.
## 3. Inference Demo Code
If you prefer to run model locally, please use the **pth file** with [official installation guide](https://github.com/OnomaAI/Illustrious-Lumina).
The setup used for header image can be replicated with following setup:
## 4. Disclaimer
The model does not reflect any final product, and intended to be used for research analysis only. The model is not production-ready; use as own risk.
The model is in Proof Of Concept stage- supposedly, 3% of the compute required for full training, with only 22M samples seen with low-resolution joint training, with A6000 GPUs.
For training acceleration, please consider supporting us in [Support site](https://illustrious-xl.ai/model/17)!
项目权限
在吐司在线使用
在 吐司 作为在线训练的底模
使用时无需注明出处
用于模型融合
分享融合模型时使用不同的许可
使用权限
生成的内容用于商业用途
作为生成服务来商用
转售模型或出售融合模型
商用许可
相关帖子
描述想生成的画面,回车发送