Dear authors,
First of all, thank you very much for your excellent work and for making the source code publicly available. We truly appreciate your contribution.
We have been experimenting with StyleShot for medical image applications. Specifically, we tested the generation performance using styleshot_image_driven_demo.py and styleshot_controlnet_demo.py. However, the results were not as satisfactory as expected.
style:
content:

generate:
We also visualized the intermediate results (e.g., contour and lineart) and found that the edge extraction performance was not ideal either.
Lineart:
Contour:
Could you please share some suggestions on how to improve the performance in such scenarios?
Do you think retraining StyleShot on our medical dataset might help? Our dataset, however, does not contain any textual annotations — would it still be possible to train StyleShot under this condition?
Thank you again for your outstanding work and for taking the time to provide any insights!