Skip to content

Commit 1040875

Browse files
yao531441pre-commit-ci[bot]
andauthoredDec 27, 2024··
Refactor image2image (#1076)
* Refactor image2image Signed-off-by: Yao, Qing <[email protected]> --------- Signed-off-by: Yao, Qing <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
1 parent a7888ab commit 1040875

File tree

11 files changed

+234
-129
lines changed

11 files changed

+234
-129
lines changed
 

‎.github/workflows/docker/compose/image2image-compose.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -5,9 +5,9 @@
55
services:
66
image2image:
77
build:
8-
dockerfile: comps/image2image/Dockerfile
8+
dockerfile: comps/image2image/src/Dockerfile
99
image: ${REGISTRY:-opea}/image2image:${TAG:-latest}
1010
image2image-gaudi:
1111
build:
12-
dockerfile: comps/image2image/Dockerfile.intel_hpu
12+
dockerfile: comps/image2image/src/Dockerfile.intel_hpu
1313
image: ${REGISTRY:-opea}/image2image-gaudi:${TAG:-latest}

‎comps/image2image/image2image.py

-117
This file was deleted.

‎comps/image2image/Dockerfile ‎comps/image2image/src/Dockerfile

+3-3
Original file line numberDiff line numberDiff line change
@@ -12,12 +12,12 @@ COPY comps /home/comps
1212

1313
RUN pip install --no-cache-dir --upgrade pip && \
1414
if [ ${ARCH} = "cpu" ]; then pip install torch torchvision --index-url https://download.pytorch.org/whl/cpu; fi && \
15-
pip install --no-cache-dir -r /home/comps/image2image/requirements.txt
15+
pip install --no-cache-dir -r /home/comps/image2image/src/requirements.txt
1616

1717
ENV PYTHONPATH=$PYTHONPATH:/home
1818

19-
WORKDIR /home/comps/image2image
19+
WORKDIR /home/comps/image2image/src
2020

21-
RUN echo python image2image.py --bf16 >> run.sh
21+
RUN echo python opea_image2image_microservice.py --bf16 >> run.sh
2222

2323
CMD bash run.sh

‎comps/image2image/Dockerfile.intel_hpu ‎comps/image2image/src/Dockerfile.intel_hpu

+3-3
Original file line numberDiff line numberDiff line change
@@ -19,11 +19,11 @@ ENV PYTHONPATH=/home/user:/usr/lib/habanalabs/:/home/user/optimum-habana
1919

2020
# Install requirements and optimum habana
2121
RUN pip install --no-cache-dir --upgrade pip && \
22-
pip install --no-cache-dir -r /home/user/comps/image2image/requirements.txt && \
22+
pip install --no-cache-dir -r /home/user/comps/image2image/src/requirements.txt && \
2323
pip install --no-cache-dir optimum[habana]
2424

25-
WORKDIR /home/user/comps/image2image
25+
WORKDIR /home/user/comps/image2image/src
2626

27-
RUN echo python image2image.py --device hpu --use_hpu_graphs --bf16 >> run.sh
27+
RUN echo python opea_image2image_microservice.py --device hpu --use_hpu_graphs --bf16 >> run.sh
2828

2929
CMD bash run.sh

‎comps/image2image/README.md ‎comps/image2image/src/README.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ export HF_TOKEN=<your huggingface token>
2828
Start the OPEA Microservice:
2929

3030
```bash
31-
python image2image.py --bf16 --model_name_or_path $MODEL --token $HF_TOKEN
31+
python opea_image2image_microservice.py --bf16 --model_name_or_path $MODEL --token $HF_TOKEN
3232
```
3333

3434
# 🚀2. Start Microservice with Docker (Option 2)
@@ -48,7 +48,7 @@ Build image-to-image service image on Xeon with below command:
4848

4949
```bash
5050
cd ../..
51-
docker build -t opea/image2image:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/image2image/Dockerfile .
51+
docker build -t opea/image2image:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/image2image/src/Dockerfile .
5252
```
5353

5454
### 2.1.2 Image-to-Image Service Image on Gaudi
@@ -57,7 +57,7 @@ Build image-to-image service image on Gaudi with below command:
5757

5858
```bash
5959
cd ../..
60-
docker build -t opea/image2image-gaudi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/image2image/Dockerfile.intel_hpu .
60+
docker build -t opea/image2image-gaudi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/image2image/src/Dockerfile.intel_hpu .
6161
```
6262

6363
## 2.2 Start Image-to-Image Service
File renamed without changes.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,2 @@
1+
# Copyright (C) 2024 Intel Corporation
2+
# SPDX-License-Identifier: Apache-2.0
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,139 @@
1+
# Copyright (C) 2024 Intel Corporation
2+
# SPDX-License-Identifier: Apache-2.0
3+
import base64
4+
import os
5+
import threading
6+
7+
from comps import CustomLogger, OpeaComponent, SDImg2ImgInputs, ServiceType
8+
9+
logger = CustomLogger("opea_imagetoimage")
10+
logflag = os.getenv("LOGFLAG", False)
11+
12+
import torch
13+
from diffusers import AutoPipelineForImage2Image
14+
from diffusers.utils import load_image
15+
16+
pipe = None
17+
args = None
18+
initialization_lock = threading.Lock()
19+
initialized = False
20+
21+
22+
def initialize(
23+
model_name_or_path="stabilityai/stable-diffusion-xl-refiner-1.0",
24+
device="cpu",
25+
token=None,
26+
bf16=True,
27+
use_hpu_graphs=False,
28+
):
29+
global pipe, args, initialized
30+
with initialization_lock:
31+
if not initialized:
32+
# initialize model and tokenizer
33+
if os.getenv("MODEL", None):
34+
model_name_or_path = os.getenv("MODEL")
35+
kwargs = {}
36+
if bf16:
37+
kwargs["torch_dtype"] = torch.bfloat16
38+
if not token:
39+
token = os.getenv("HF_TOKEN")
40+
if device == "hpu":
41+
kwargs(
42+
{
43+
"use_habana": True,
44+
"use_hpu_graphs": use_hpu_graphs,
45+
"gaudi_config": "Habana/stable-diffusion",
46+
"token": token,
47+
}
48+
)
49+
if "stable-diffusion-xl" in model_name_or_path:
50+
from optimum.habana.diffusers import GaudiStableDiffusionXLImg2ImgPipeline
51+
52+
pipe = GaudiStableDiffusionXLImg2ImgPipeline.from_pretrained(
53+
model_name_or_path,
54+
**kwargs,
55+
)
56+
else:
57+
raise NotImplementedError(
58+
"Only support stable-diffusion-xl now, " + f"model {model_name_or_path} not supported."
59+
)
60+
elif device == "cpu":
61+
pipe = AutoPipelineForImage2Image.from_pretrained(model_name_or_path, token=token, **kwargs)
62+
else:
63+
raise NotImplementedError(f"Only support cpu and hpu device now, device {device} not supported.")
64+
logger.info("Stable Diffusion model initialized.")
65+
initialized = True
66+
67+
68+
class OpeaImageToImage(OpeaComponent):
69+
"""A specialized ImageToImage component derived from OpeaComponent for Stable Diffusion model .
70+
71+
Attributes:
72+
model_name_or_path (str): The name of the Stable Diffusion model used.
73+
device (str): which device to use.
74+
token(str): Huggingface Token.
75+
bf16(bool): Is use bf16.
76+
use_hpu_graphs(bool): Is use hpu_graphs.
77+
"""
78+
79+
def __init__(
80+
self,
81+
name: str,
82+
description: str,
83+
config: dict = None,
84+
seed=42,
85+
model_name_or_path="stabilityai/stable-diffusion-xl-refiner-1.0",
86+
device="cpu",
87+
token=None,
88+
bf16=True,
89+
use_hpu_graphs=False,
90+
):
91+
super().__init__(name, ServiceType.IMAGE2IMAGE.name.lower(), description, config)
92+
initialize(
93+
model_name_or_path=model_name_or_path, device=device, token=token, bf16=bf16, use_hpu_graphs=use_hpu_graphs
94+
)
95+
self.pipe = pipe
96+
self.seed = seed
97+
98+
def invoke(self, input: SDImg2ImgInputs):
99+
"""Invokes the ImageToImage service to generate Images for the provided input.
100+
101+
Args:
102+
input (SDImg2ImgInputs): The input in SD images format.
103+
"""
104+
image = load_image(input.image).convert("RGB")
105+
prompt = input.prompt
106+
num_images_per_prompt = input.num_images_per_prompt
107+
108+
generator = torch.manual_seed(self.seed)
109+
images = pipe(
110+
image=image, prompt=prompt, generator=generator, num_images_per_prompt=num_images_per_prompt
111+
).images
112+
image_path = os.path.join(os.getcwd(), prompt.strip().replace(" ", "_").replace("/", ""))
113+
os.makedirs(image_path, exist_ok=True)
114+
results = []
115+
for i, image in enumerate(images):
116+
save_path = os.path.join(image_path, f"image_{i + 1}.png")
117+
image.save(save_path)
118+
with open(save_path, "rb") as f:
119+
bytes = f.read()
120+
b64_str = base64.b64encode(bytes).decode()
121+
results.append(b64_str)
122+
123+
return results
124+
125+
def check_health(self) -> bool:
126+
"""Checks the health of the ImageToImage service.
127+
128+
Returns:
129+
bool: True if the service is reachable and healthy, False otherwise.
130+
"""
131+
try:
132+
if self.pipe:
133+
return True
134+
else:
135+
return False
136+
except Exception as e:
137+
# Handle connection errors, timeouts, etc.
138+
logger.error(f"Health check failed: {e}")
139+
return False
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,81 @@
1+
# Copyright (C) 2024 Intel Corporation
2+
# SPDX-License-Identifier: Apache-2.0
3+
4+
import argparse
5+
import base64
6+
import os
7+
import time
8+
9+
from comps import (
10+
CustomLogger,
11+
OpeaComponentController,
12+
SDImg2ImgInputs,
13+
SDOutputs,
14+
ServiceType,
15+
opea_microservices,
16+
register_microservice,
17+
register_statistics,
18+
statistics_dict,
19+
)
20+
from comps.image2image.src.integration.opea_image2image_native import OpeaImageToImage
21+
22+
args = None
23+
24+
logger = CustomLogger("image2image")
25+
26+
27+
# Initialize OpeaComponentController
28+
controller = OpeaComponentController()
29+
30+
# Register components
31+
# try:
32+
33+
# except Exception as e:
34+
# logger.error(f"Failed to initialize components: {e}")
35+
36+
37+
@register_microservice(
38+
name="opea_service@image2image",
39+
service_type=ServiceType.IMAGE2IMAGE,
40+
endpoint="/v1/image2image",
41+
host="0.0.0.0",
42+
port=9389,
43+
input_datatype=SDImg2ImgInputs,
44+
output_datatype=SDOutputs,
45+
)
46+
@register_statistics(names=["opea_service@image2image"])
47+
def image2image(input: SDImg2ImgInputs):
48+
start = time.time()
49+
results = controller.invoke(input)
50+
statistics_dict["opea_service@image2image"].append_latency(time.time() - start, None)
51+
return SDOutputs(images=results)
52+
53+
54+
if __name__ == "__main__":
55+
parser = argparse.ArgumentParser()
56+
parser.add_argument("--model_name_or_path", type=str, default="stabilityai/stable-diffusion-xl-refiner-1.0")
57+
parser.add_argument("--use_hpu_graphs", default=False, action="store_true")
58+
parser.add_argument("--device", type=str, default="cpu")
59+
parser.add_argument("--token", type=str, default=None)
60+
parser.add_argument("--seed", type=int, default=42)
61+
parser.add_argument("--bf16", action="store_true")
62+
63+
args = parser.parse_args()
64+
# Instantiate Animation component and register it to controller
65+
opea_imagetoimage = OpeaImageToImage(
66+
name="OpeaImageToImage",
67+
description="OPEA Image To Image Service",
68+
seed=args.seed,
69+
model_name_or_path=args.model_name_or_path,
70+
device=args.device,
71+
token=args.token,
72+
bf16=args.bf16,
73+
use_hpu_graphs=args.use_hpu_graphs,
74+
)
75+
76+
controller.register(opea_imagetoimage)
77+
78+
# Discover and activate a healthy component
79+
controller.discover_and_activate()
80+
logger.info("Image2image server started.")
81+
opea_microservices["opea_service@image2image"].start()
File renamed without changes.

‎tests/image2image/test_image2image.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ip_address=$(hostname -I | awk '{print $1}')
1010
function build_docker_images() {
1111
cd $WORKPATH
1212
echo $(pwd)
13-
docker build --no-cache -t opea/image2image:latest -f comps/image2image/Dockerfile .
13+
docker build --no-cache -t opea/image2image:latest -f comps/image2image/src/Dockerfile .
1414
if [ $? -ne 0 ]; then
1515
echo "opea/image2image built fail"
1616
exit 1

0 commit comments

Comments
 (0)
Please sign in to comment.