Page 24 - 2025S
P. 24

UEC Int’l Mini-Conference No.54                                                               17







            nator module. To the best of our research, no      [4] Fu, B., Yu, F., Liu, A., Wang, Z., Wen,
            previous studies had applied the approach of the      J., He, J., Qiao, Y.: Generate like experts:
            diffusion approach to the generation of Bengali       Multi-stage font generation by incorporat-
            fonts. Based on our experimental results, our         ing font transfer process into diffusion mod-
            proposed method effectively generates Bengali         els. In: Proc. of IEEE Computer Vision and
            fonts while maintaining their structural sound-       Pattern Recognition. pp. 6892–6901 (2024)
            ness. The implementation of our proposed dual
            aggregation cross-attention has been done suc-     [5] Fuad, M.M., Faiyaz, A., Arnob, N.M.K.,
            cessfully in terms of combining style information     Mridha, M.F., Saha, A.K., Aung, Z.:
            on the reference glyph with the content of the        Okkhor-diffusion: class guided generation
            source glyph. Since our proposed method works         of bangla isolated handwritten charac-
            on the digitally rendered fonts, we plan to ex-       ters using denoising diffusion probabilistic
            pand our model in the future to accommodate           model (ddpm). IEEE Access (2024)
            handwritten glyphs and to improve the perfor-
            mance of unseen fonts.                             [6] He, H., Chen, X., Wang, C., Liu, J., Du,
                                                                  B., Tao, D., Yu, Q.: Diff-font: Diffusion
                                                                  model for robust one-shot font generation.
            7    Acknowledgment                                   International Journal of Computer Vision
                                                                  132(11), 5372–5386 (2024)
            The Author was supported by a JASSO schol-
            arship.   The author thanks the Yanai Lab          [7] Kong, Y., Luo, C., Ma, W., Zhu, Q.,
            for their contributions, computational resources,     Zhu, S., Yuan, N., Jin, L.: Look closer
            and supportive research environment.   Their          to supervise better: One-shot font genera-
            continuous guidance and encouragement were in-        tion via component-based discriminator. In:
            strumental in completing this work.This work          Proc. of IEEE Computer Vision and Pat-
            was supported by JSPS KAKENHI Grant Num-              tern Recognition. pp. 13482–13491 (2022)
            ber, 22H00548, and JST CRONOS Grant Num-           [8] Liu, W., Liu, F., Ding, F., He, Q., Yi, Z.:
            ber JPMJCS24K4.                                       Xmp-font: Self-supervised cross-modality
                                                                  pre-training for few-shot font generation.
                                                                  In: Proc. of IEEE Computer Vision and
            References
                                                                  Pattern Recognition. pp. 7905–7914 (2022)
             [1] Abedin, M.M.h.z., Ghosh, T., Mehrub, T.,      [9] Lyu, P., Bai, X., Yao, C., Zhu, Z., Huang,
                 Yousuf, M.A.: Bangla printed character           T., Liu, W.: Auto-encoder guided gan for
                 generation from handwritten character us-        chinese calligraphy synthesis. In: 2017 14th
                 ing gan. In: Soft computing for data ana-        IAPR International Conference on Docu-
                 lytics, classification model, and control, pp.   ment Analysis and Recognition (ICDAR).
                 153–165. Springer (2022)                         vol. 1, pp. 1095–1100. IEEE (2017)


             [2] Azadi, S., Fisher, M., Kim, V.G., Wang, Z.,  [10] Park, S., Chun, S., Cha, J., Lee, B., Shim,
                 Shechtman, E., Darrell, T.: Multi-content        H.: Few-shot font generation with localized
                 gan for few-shot font style transfer. In:        style representations and factorization. In:
                 Proc. of IEEE Computer Vision and Pat-           Proc. of the AAAI Conference on Artificial
                 tern Recognition. pp. 7564–7573 (2018)           Intelligence. vol. 35, pp. 2393–2402 (2021)


             [3] Chang, J., Gu, Y., Zhang, Y., Wang, Y.F.,    [11] Saha, C., Faisal, R.H., Rahman, M.M.:
                 Innovation, C.: Chinese handwriting imita-       Bangla handwritten basic character recog-
                 tion with hierarchical generative adversarial    nition using deep convolutional neural net-
                 network. In: Proc. of the British Machine        work. In: 2019 Joint 8th International Con-
                 Vision Conference. p. 290 (2018)                 ference on Informatics, Electronics & Vision
   19   20   21   22   23   24   25   26   27   28   29