레이블이 ControlNet인 게시물을 표시합니다. 모든 게시물 표시
레이블이 ControlNet인 게시물을 표시합니다. 모든 게시물 표시

2025년 3월 12일 수요일

저사양 PC에서 활용할 수 있는 ControlNet 모델

아래는 저사양 PC에서 활용할 수 있는 ControlNet 모델을 정리한 표입니다:

ControlNet 이름

특징

적용 모델 (버전 포함)

다운로드 링크

ControlNet Canny

이미지의 외곽선을 기반으로 세부적인 제어 가능.

Stable Diffusion 1.5, 2.1

ControlNet Canny

ControlNet OpenPose

인체 자세를 감지하여 이미지 생성 시 활용 가능.

Stable Diffusion 1.5, 2.1

ControlNet OpenPose

ControlNet Depth

깊이 정보를 기반으로 이미지의 구조를 제어 가능.

Stable Diffusion 1.5, 2.1

ControlNet Depth

ControlNet Scribble

간단한 스케치를 기반으로 이미지를 생성.

Stable Diffusion 1.5, 2.1

ControlNet Scribble

ControlNet Segmentation

이미지의 세그먼트 정보를 활용하여 세부적인 제어 가능.

Stable Diffusion 1.5, 2.1

ControlNet Segmentation

2025년 3월 7일 금요일

5-1 Let's learn about ControlNet(Stable Diffusion Practical Guide Table of Contents)

>>>What is ControlNet?

ControlNet is a derivative extension of Stable Diffusion, which was published in the February 2023 paper, <Adding Conditional Control to Text-to-Image Diffusion Modes>. Using this, you can input an image or pose in addition to the prompt to generate complex compositions that cannot be indicated by the prompt alone, poses that are difficult to explain with text, and images that reproduce the characters in the original photo.

The person who created this technology is lllyasviel (zhang Lvmin), the main author of the paper, and as of March 2024, lllyasviel has released the details of the technology and the model on GitHub and Hugging Face.

Github - Lllyasviel / ControlNet-v1-1-nightly

https//github.com/lllyasviel/ControlNet-v1-1-nightly

Additionally, Mikubill has created an extension to allow ControlNet to be used in AUTOMATIC1111 and released it as open source.

Github - Mikubill / sd-webui-controlNet

https://github.com/Mikubill/sd-webui-controlnet

ControlNet is an artificial neural network technology that adds spatial condition control by the diffusion model of Stable Diffusion. It consists of several types of 'preprocessors' such as openpose, which extracts poses from images, and canny, which extracts outlines, and the extracted information is used as condition control for image creation using txt2img. If you use each preprocessor separately according to its purpose, you can control the composition or pose that was difficult to handle in the existing txt2img to create an image as you intended.

>>>Know the difference between img2img and ControlNet

If you think about generating a new image based on an image and a prompt, you might think of img2img, but img2img and ControlNet are completely different technologies.

While img2img identifies the features of the entire input image to generate an image, ControlNet pre-analyzes the information (image, border, pose, depth, etc.) on the input image using a pre-processor, and extracts only the features of the specific elements that each pre-processor is responsible for from the input image to generate an image. For example, you can reproduce only the pose from the input image as shown below.

Original image

Image source pexels

Pose of a character extracted with the openpose preprocessor

Original image

Pose of a character extracted with the openpose preprocessor

Recommended Posts

Love is taking responsibility till the end

  I created a campaign called "Don't Abandon Your Beloved Dog" with Stable Diffusion. I tried creating it several times, but i...