ﻻ يوجد ملخص باللغة العربية
In this work, we present a non-parametric texture synthesis algorithm capable of producing plausible images without copying large tiles of the exemplar. We focus on a simple synthesis algorithm, where we explore two patch match heuristics; the well known Bidirectional Similarity (BS) measure and a heuristic that finds near permutations using the solution of an entropy regularized optimal transport (OT) problem. Innovative synthesis is achieved with a small patch size, where global plausibility relies on the qualities of the match. For OT, less entropic regularization also meant near permutations and more plausible images. We examine the tile maps of the synthesized images, showing that they are indeed novel superpositions of the input and contain few or no verbatim copies. Synthesis results are compared to a statistical method, namely a random convolutional network. We conclude by remarking simple algorithms using only the input image can synthesize textures decently well and call for more modest approaches in future algorithm design.
Search-based texture synthesis algorithms are sensitive to the order in which texture samples are generated; different synthesis orders yield different textures. Unfortunately, most polygon rasterizers and ray tracers do not guarantee the order with
Recently, deep generative adversarial networks for image generation have advanced rapidly; yet, only a small amount of research has focused on generative models for irregular structures, particularly meshes. Nonetheless, mesh generation and synthesis
A non-parametric interpretable texture synthesis method, called the NITES method, is proposed in this work. Although automatic synthesis of visually pleasant texture can be achieved by deep neural networks nowadays, the associated generation models a
Image generation from scene description is a cornerstone technique for the controlled generation, which is beneficial to applications such as content creation and image editing. In this work, we aim to synthesize images from scene description with re
Existing compression methods typically focus on the removal of signal-level redundancies, while the potential and versatility of decomposing visual data into compact conceptual components still lack further study. To this end, we propose a novel conc