From e655821e79be4a42cf7c88b17667cd485eb66045 Mon Sep 17 00:00:00 2001 From: Faxuan Cai <140777998+Layer-norm@users.noreply.github.com> Date: Sat, 9 Mar 2024 06:18:46 +0800 Subject: [PATCH] Update config.example.yaml fix Letter case --- config.example.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/config.example.yaml b/config.example.yaml index 509ba9f..ef38053 100644 --- a/config.example.yaml +++ b/config.example.yaml @@ -1,7 +1,7 @@ # this is an example for config.yaml file, you can rename it to config.yaml if you want to use it # ############################################################################################### # This path is for custom pressesor models base folder. default is "./ckpts" -# you can also use absolute paths like: "/root/ComfyUI/custom_nodes/comfyui_controlnet_aux/ckpts" or "D:\\comfyui\\custom_nodes\\comfyui_controlnet_aux\\ckpts" +# you can also use absolute paths like: "/root/ComfyUI/custom_nodes/comfyui_controlnet_aux/ckpts" or "D:\\ComfyUI\\custom_nodes\\comfyui_controlnet_aux\\ckpts" annotator_ckpts_path: "./ckpts" # ############################################################################################### # This path is for downloading temporary files. @@ -17,4 +17,4 @@ USE_SYMLINKS: False # empty list or only keep ["CPUExecutionProvider"] means you use cv2.dnn.readNetFromONNX to load onnx models # if your onnx models can only run on the CPU or have other issues, we recommend using pt model instead. # default value is ["CUDAExecutionProvider", "DirectMLExecutionProvider", "OpenVINOExecutionProvider", "ROCMExecutionProvider", "CPUExecutionProvider"] -EP_list: ["CUDAExecutionProvider", "DirectMLExecutionProvider", "OpenVINOExecutionProvider", "ROCMExecutionProvider", "CPUExecutionProvider"] \ No newline at end of file +EP_list: ["CUDAExecutionProvider", "DirectMLExecutionProvider", "OpenVINOExecutionProvider", "ROCMExecutionProvider", "CPUExecutionProvider"]