Abstract
Garment line draft is the basis of clothing design. Automatic or semi-automatic colorization of garment line draft will improve the efficiency of fashion designers and reduce the drawing cost. In this paper, we present a garment line draft colorization method based on cGAN, which can support user interaction by adding scribbles to guide the colorization process. Due to the inadequacy of the garment line drafts, we construct a paired garment-line image dataset for training our colorization model. While existing methods for line art colorization are able to generate plausible colorized results, they tend to suffer from the color bleeding issue. We introduce a region segmentation fusion mechanism to aid colorization frameworks in avoiding color bleeding. The experimental results show that each module in the method can contribute to the final result. In addition, the comparison with the classical methods that our method can avoid large areas of leakage in the background and have cleaner garment details.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Goodfellow, I., Pouget-Abadie, J., Mirza, M., et al.: Generative adversarial nets. In: Neural Information Processing Systems. MIT Press, Cambridge (2014)
Mirza, M., Osindero, S.: Conditional generative adversarial nets. In: Computer Science, pp. 2672–2680 (2014)
Isola, P., et al.: Image-to-Image translation with conditional adversarial networks. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1125–1134 (2017)
Sangkloy, P., Burnell, N., Ham, C., et al.: The sketchy database: learning to retrieve badly drawn bunnies. ACM Trans. Graph. 35(4), 119 (2016)
Furusawa, C., Hiroshiba, K., Ogaki, K., Odagiri, Y.: Comicolorization: semi-automatic manga colorization. In: SIGGRAPH Asia 2017 Technical Briefs, vol. 12. ACM (2017)
Hensman, P., Aizawa, K.: cGAN-based manga colorization using a single training image. arXiv preprint arXiv:1706.06918 (2017)
Yu, Q., et al.: Sketch-a-Net: a deep neural network that beats humans. Int. J. Comput. Vision 122(3), 411–425 (2017)
Qian, Y., Feng, L., Song, Y.Z., et al.: Sketch me that shoe. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition(CVPR), pp. 799–807. IEEE, Piscataway (2016)
Xie, S., Girshick, R,, Dollár, P., et al.: Aggregated residual transformations for deep neural networks. IEEE (2016)
Hu, J., Shen, L., Albanie, S., et al.: Squeeze-and-excitation networks. arXiv preprint arXiv:1704.04861 (2017)
Liu, X., Wu, W., Li, C., Li, Y., Wu, H.: Reference-guided structure-aware deep sketch colorization for cartoons. Comput. Vis. Media 8(1), 135–148 (2022)
Liu, G., Dang, M., Liu, J., Xiang, R., Tian, Y., Luo, N.: True wide convolutional neural network for image denoising. Inf. Sci. 610, 171–184 (2022)
Li, P., Li, Z., Pang, X., Wang, H., Lin, W., Wentai, W.: Multi-scale residual denoising GAN model for producing super-resolution CTA images. Ambient Intell. Humaniz. Comput. 13(3), 1515–1524 (2022)
Xinyue, W., Chen, Z., Peng, C., Ye, X.: MMSRNet: pathological image super-resolution by multi-task and multi-scale learning. Biomed. Signal Process. Control 81, 104428 (2023)
Li, W., Zhou, K., Qi, L., Lu, L., Lu, J.: Best-buddy GANs for highly detailed image super-resolution. In: AAAI, pp. 1412–1420 (2022)
Lu, P., Yu, J., Peng, X., Zhao, Z., Wang, X.: Gray2colornet: transfer more colors from reference image. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 3210–3218 (2020)
Vitoria, P., Raad, L., Ballester, C.: Chromagan: adversarial picture colorization with semantic class distribution. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 2445–2454 (2020)
Li, Z., Geng, Z., Kang, Z., Chen, W., Yang, Y.: Eliminating gradient conflict in reference-based line-art colorization. In: ECCV 17, pp. 579–596 (2022)
Shu-Yu Chen, Jia-Qi Zhang, Lin Gao, Yue He, Shihong Xia, Min Shi, Fang-Lue Zhang.: Active Colorization for Cartoon Line Drawings. In: IEEE Trans. Vis. Comput. Graph. 28(2), 1198–1208 (2022)
Chen, W., Hays, J.: Sketchygan: towards diverse and realistic sketch to image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416–9425 (2018)
Ci, Y., Ma, X., Wang, Z., Li, H., Luo, Z.: User-guided deep anime line art colorization with conditional adversarial networks. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 1536–1544 (2018)
Zhang, L., Li, C., Wong, T.-T., Ji, Y., Liu, C.: Two-stage sketch colorization. ACM Trans. Graph. (TOG) 37(6), 1–14 (2018)
Zhang, L., Ji, Y., Lin, X., Liu, C.: Style transfer for anime sketches with enhanced residual u-net and auxiliary classifier gan. In: 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR), pp. 506–511. IEEE (2017)
Lee, J., Kim, E., Lee, Y., Kim, D., Chang, J., Choo, J.: Reference-based sketch image colorization using augmented-self reference and dense semantic correspondence. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5801–5810(2020)
Huang, X., Liu, M.-Y., Belongie, S., Kautz, J.: Multi-modal unsupervised image-to-image translation. In: Proceedings of the European conference on computer vision (ECCV), pp. 172–189 (2018)
Zhang, P., Zhang, B., Chen, D., Yuan, L., Wen, F.: Cross-domain correspondence learning for exemplar-based image translation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5143–5153 (2020)
Zou, C., Mo, H., Gao, C., Du, R., Fu, H.: Language-based colorization of scene sketches. ACM Trans. Graph. (TOG) 38(6), 1–16 (2019)
Kim, H., Jhoo, H. Y., Park, E., Yoo, S.: Tag2pix: line art colorization using text tag with secat and changing loss. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9056–9065 (2019)
Sarvadevabhatla, R.K., Dwivedi, I., Biswas, A., Manocha, S.: Sketchparse: towards rich descriptions for poorly drawn sketches using multi-task hierarchical deep networks. In: Proceedings of the 25th ACM international conference on Multimedia, pp. 10–18 (2017)
Zhang, L., Ji, Y., Liu, C.: DanbooRegion: an illustration region dataset. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12358, pp. 137–154. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58601-0_9
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
He, R., Yang, X., Huang, J. (2024). cGAN-Based Garment Line Draft Colorization Using a Garment-Line Dataset. In: Sheng, B., Bi, L., Kim, J., Magnenat-Thalmann, N., Thalmann, D. (eds) Advances in Computer Graphics. CGI 2023. Lecture Notes in Computer Science, vol 14496. Springer, Cham. https://doi.org/10.1007/978-3-031-50072-5_27
Download citation
DOI: https://doi.org/10.1007/978-3-031-50072-5_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-50071-8
Online ISBN: 978-3-031-50072-5
eBook Packages: Computer ScienceComputer Science (R0)