Open questions

The following are the big questions the author is currently unable to answer. If you would have any ideas to contribute, please use the issues function of Github for this project.

The focus is on deep generative models for SVG images.

Open question #1: Is information in ‘raster space’ required or not?

One major open question for the author is if a generative model can work well just using SVG information – or if the model requires information from the ‘raster space’, such as via a differentiable renderer (diffvg). For us humans, it really matters how the image looks like rendered onto a display in the end.

Background

Some existing generative models for SVG are end-to-end and exclusively work in the ‘SVG space’. These models encode SVGs into a vector and solely train on and work with the vector image data representation. Examples are DeepSVG by Carlier et al. (2020) and Aoki and Aizawa (2022).

Other models include information from the ‘raster space’, e.g. Wang and Lian (2021). These models either feed in a raster image data representation in addition to the vector image data representation or they use a differentiable rasterizer to obtain a loss in the raster image space.

Observations

The current impression is that models that only work in the ‘SVG space’ perform less well. An exception appears to be the recent work by Aoki and Aizawa (2022) which appears to outperform DeepSVG and DeepVecFont.