From b5acef9e74282852bb1546f196bf0e26c3eb5711 Mon Sep 17 00:00:00 2001 From: mi804 <1576993271@qq.com> Date: Tue, 10 Feb 2026 17:42:57 +0800 Subject: [PATCH] test relative --- docs/en/conf.py | 4 ---- docs/zh/Pipeline_Usage/Model_Inference.md | 2 +- docs/zh/README.md | 2 +- docs/zh/conf.py | 4 ---- 4 files changed, 2 insertions(+), 10 deletions(-) diff --git a/docs/en/conf.py b/docs/en/conf.py index 9d2c32a..6f8d74a 100644 --- a/docs/en/conf.py +++ b/docs/en/conf.py @@ -121,7 +121,3 @@ myst_enable_extensions = [ 'dollarmath', 'colon_fence', ] - -linkcheck_ignore = [ - r'/docs/en/.*', -] diff --git a/docs/zh/Pipeline_Usage/Model_Inference.md b/docs/zh/Pipeline_Usage/Model_Inference.md index fef1481..d0cacf1 100644 --- a/docs/zh/Pipeline_Usage/Model_Inference.md +++ b/docs/zh/Pipeline_Usage/Model_Inference.md @@ -22,7 +22,7 @@ pipe = QwenImagePipeline.from_pretrained( ) ``` -其中 `torch_dtype` 和 `device` 是计算精度和计算设备(不是模型的精度和设备)。`model_configs` 可通过多种方式配置模型路径,关于本项目内部是如何加载模型的,请参考 [`diffsynth.core.loader`](/docs/zh/API_Reference/core/loader.md)。 +其中 `torch_dtype` 和 `device` 是计算精度和计算设备(不是模型的精度和设备)。`model_configs` 可通过多种方式配置模型路径,关于本项目内部是如何加载模型的,请参考 [`diffsynth.core.loader`](../API_Reference/core/loader.md)。
diff --git a/docs/zh/README.md b/docs/zh/README.md index c02665f..1463116 100644 --- a/docs/zh/README.md +++ b/docs/zh/README.md @@ -26,7 +26,7 @@ graph LR; 本节介绍 `DiffSynth-Studio` 的基本使用方式,包括如何启用显存管理从而在极低显存的 GPU 上进行推理,以及如何训练任意基础模型、LoRA、ControlNet 等模型。 -* [安装依赖](/docs/zh/Pipeline_Usage/Setup.md) +* [安装依赖](./Pipeline_Usage/Setup.md) * [模型推理](/docs/zh/Pipeline_Usage/Model_Inference.md) * [显存管理](/docs/zh/Pipeline_Usage/VRAM_management.md) * [模型训练](/docs/zh/Pipeline_Usage/Model_Training.md) diff --git a/docs/zh/conf.py b/docs/zh/conf.py index b6fcb00..2d12beb 100644 --- a/docs/zh/conf.py +++ b/docs/zh/conf.py @@ -121,7 +121,3 @@ myst_enable_extensions = [ 'dollarmath', 'colon_fence', ] - -linkcheck_ignore = [ - r'/docs/zh/.*', -]