March 6, 2024, 5:45 a.m. | Sho Shimotsumagari, Shumpei Takezaki, Daichi Haraguchi, Seiichi Uchida

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.02919v1 Announce Type: new
Abstract: The purpose of this paper is to enable the conversion between machine-printed character images (i.e., font images) and handwritten character images through machine learning. For this purpose, we propose a novel unpaired image-to-image domain conversion method, CycleDM, which incorporates the concept of CycleGAN into the diffusion model. Specifically, CycleDM has two internal conversion models that bridge the denoising processes of two image domains. These conversion models are efficiently trained without explicit correspondence between the domains. …

abstract arxiv concept conversion cs.cv cyclegan diffusion diffusion model domain image images image-to-image machine machine learning novel paper through type

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne