Next Article in Journal
Analysis of the Mode Shapes of Kaplan Runners
Previous Article in Journal
A High-Level Representation of the Navigation Behavior of Website Visitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Variety of Choice Methods for Image-Based Artistic Rendering

1
Technology Talent Bechelor’s Program in Intelligent Maritime, National Kaohsiung University of Science and Technology, Kaohsiung City 813, Taiwan
2
Department of Information Management, R.O.C. Naval Academy, Kaohsiung City 813, Taiwan
3
Department of Telecommunication Engineering, National Kaohsiung University of Science and Technology, Kaohsiung City 813, Taiwan
4
Metal Industries Research & Development Centre (MIRDC), Taichung City 811, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6710; https://doi.org/10.3390/app12136710
Submission received: 16 June 2022 / Revised: 30 June 2022 / Accepted: 1 July 2022 / Published: 2 July 2022

Abstract

:
Neural style transfer (NST) is a technique based on the deep learning of a convolutional neural network (CNN) to create entertaining pictures by cleverly stylizing ordinary pictures with the predetermined visual art style. However, three issues must be carefully investigated during the generation of neural-stylized artwork: the color scheme, the strength of style of the strokes, and the adjustment of image contrast. To solve these problems and select image colorization based on personal preference, in this paper, we propose modified universal-style transfer (UST) method combined with the image fusion and color enhancement methods to design a good post-processing framework to tackle the three above-mentioned issues simultaneously. This work provides more visual effects for stylized images, and also can integrate into the UST method effectively. In addition, the proposed method is suitable for stylized images generated by any NST method, but it also works similarly to the Multi-Style Transfer (MST) method, which mixes two different stylized images. Finally, our proposed method successfully combined the modified UST method and post-processing method to meet personal preference.

1. Introduction

Automatic image colorization is one of the popular research topics in image processing. In general, image colorization is the process of adding colors to mono-chroma images to make them more aesthetically appealing and perceptually meaningful. However, the most crucial task in this paper is how to design a multiple-choice method depending on personal preference to define an appropriate color from the reference images for propagating to the target image. Among deep learning techniques, the convolutional neural networks (CNN) have been validated to be quite useful for image analysis and applications [1,2,3]. For example, CNN-based art style transfers have recently been employed to produce impressive artwork [4,5,6,7] to such an extent that a new market has been created for the mobile apps that can stylize user-provided images with a particular art style. The popularity of commercial apps such as Prisma [8] and Deepart.io [9] have been apparently increased. Neural style transfer (NST) is an optimization technique used to take two images—a content image C and a style reference image S (such as an artwork by a famous painter)—and blend them together so the output image looks like the content image, but “painted” in the style of the style reference image. In general, there are two main categories of style-transfer algorithms, namely Image-Optimization-Based Online Neural Methods (IOB-NST) and Model-Optimization-Based Offline Neural Methods (MOB-NST) [10]. According to the development process, there are two major approaches in the generative neural method, that is, Fast Style Transfer (FST) [4] and Multi-Style Transfer (MST) [4]. During the training phase in art education and industrial applications, the MST is advantageous in effectively and efficiently producing a variety of styles for the specified image. Further, Li et al. [11] proposed the seminal work that transfers an arbitrary visual style to content images, achieving so-called universal style transfer (UST). The main point of this paper is completely independent of the image style during the training phase and aims to make it “style learning-free”. Recent, Wang et al. [12] proposed a new knowledge distillation method to reduce the model size, which exploits the phenomenon that encoder–decoder pairs in UST construct an exclusive collaborative relationship, which is regarded as a new kind of knowledge for style transfer models.
Although the artistic effect of the NST-based methods is impressive, there are three issues worth noting. The first is the S C image copies the color distribution of the S image which might be undesirable for most artwork applications. In order to tackle this problem, one possible method is to use luminance-only transfer [13] by a 3×3 kernel to convert the C and S images from the RGB color space to the YIQ space, respectively. In the inverse conversion, the I and Q components of the C image are retained and the Y component of the C image is replaced by that of the S C image. Consequently, the color information of the C image can be preserved in the new image while the spatial details of the new image are inherited from the S C image. Second, in the control of the trade-off between content and style in the S C image, the predefined relative content-style weight plays a leading role. If the stylized effect is not as expected, one solution is to use the blending technique to adjust the mixing ratio between C and S C images [14]. Therefore, retraining is not required at all. The third concern is the use of batch or instance normalization in FST and MST. It is particularly useful for forwarding propagation when the number of CNN layers is too large. However, the normalization may significantly degrade the image contrast performance. To improve the visual effect, some effective contrast-enhancing methods have been proposed [15,16]. Gatys et al. [15] present intuitive ways to control NST. The authors proposed a method for transferring different styles to different spatial regions of the output image during stylization to substantially improve the quality and flexibility of the existing method. In addition, Jing et al. [16] propose a fine and flexible stroke size control approach for Fast Style Transfer. The core component of their algorithm is a Stroke Pyramid module, which learns different stroke sizes with adaptive receptive fields. By combining the proposed runtime control strategies, this work can achieve continuous changes in stroke sizes and produce distinct stroke sizes in different spatial regions within the same output image.
It is noted that almost all the implementations of the proposed algorithms are segmental. When the effect is not as expected, the style transfer procedures must be executed iteratively. In FST and MST, the style evaluation is very fast, and the most time-consuming operation is in the training phase, since what most of the users are concerned with is the style itself, not the style training. Thus, it will not be acceptable to spend too much time enhancing the S C image. In addition, to be effective, a good post-processing operation is also required to be fast and simple to use.
In this work, the image fusion algorithm can be applied to post-process the S C images. This work proposes a versatile post-processing framework that combines Brovery transform (BT) [17], blending technique, and intensity conservation direct decorrelation stretch (ICDDS) [18] methods to tackle the above-mentioned three issues simultaneously. This framework composed of six modules is implemented by a window-based program which is used to demonstrate the adjustment of the S C image. To validate the proposed techniques, several experiments are conducted with the stylized image produced from UST methods.
The remainder of the paper is organized as follows. Section 2 briefly introduces the related works. A detailed description of the proposed strategy for automatic colorization in neural stylized image is given in Section 3. The experimental results are discussed and described in Section 4, and Section 5 gives the conclusion.

2. Review of the Related Works

2.1. The Classical UST Method

A novel approach of image style transfer, namely UST, is proposed by Li et al. [11]. In this work, the style transfer is regarded as an image reconstruction process, accompanied by feature transform operations. The advantage of this method is that it can transform arbitrary visual style images without training on any predefined styles. As can be seen in Figure 1a, the UST employs the VGG-19 network as an encoder, and the decoder is designed as a symmetric network of VGG-19. This training is conducted by the Microsoft COCO dataset. UST extracts the image features from the five layers of the VGG-19 network and trains five corresponding decoders to achieve image reconstruction. The single UST module is shown in Figure 1b, the method is embedded with a pair of whitening and coloring transforms to an image reconstruction network. The steps of detail are shown as follows:
  • The C and S images first employ the encoder network (VGG-19) to get the feature maps f C and f S , and calculate their respective eigenvalues ( D C , D S ) and eigenvectors ( E C , E S ), respectively.
  • The output f ˜ C is whitening transform by f C , but the feature map f ˜ C S is coloring transform by the projection basis of f S , D S , and E S .
  • The new f ˜ C S is blending with f ˜ C S and f C , and using the parameter α to control. A simple blending technique is applied to trade-off between content and style in the output image.
  • The stylized image S C ( I x ) of a single UST module is reconstructed by the decoder network, as shown in Figure 1b. To obtain a more refined effect, the multi-module superposition from coarse to fine has been shown in Figure 1c.
  • The Figure 1c is the complete framework of UST. The X of the UST_X is the X levels being used in VGG19 features ( X = 1 ,   2 , ,   5 ). UST_5 means that it is only using the 5th modules. The UST_5,4 means that it is using the 5th and 4th modules without the 1st, 2nd, and 3rd module, and so on.
As a summary, the stylized image S C is controlled by parameter α and the number of modules in UST. Different from NST, FST, and MST, the UST method is the ideal combination between CNN networks and pattern recognition methods. A versatile implementation can be found in [11].

2.2. The Image Fusion and Color Enhancement for the Neural Stylized Imagery

The neural stylized effect is impressive, but it is limited in certain respects as described in Section 1. Although the problems can be solved separately, it may be quite time-consuming and cumbersome. Aimed at finding an effective solution, the image fusion and the color enhancement by ICDDS technique used in remote sensing community are applied. In the following, IHS, BT, and ICDDS methods will be briefly introduced.

2.2.1. Image Fusion by IHS and BT

Among various image fusion methods for remote sensing images, FIHS and BT are frequently used. They can be represented as [17,19]
[ R F I H S G F I H S B F I H S ] = [ R + ( P a n I ) G + ( P a n I ) B + ( P a n I ) ]   a n d   [ R B T G B T B B T ] = P a n I · [ R G B ] ,
where [ R F I H S   G F I H S   B F I H S ] T and [ R B T   G B T   B B T ] T are the fused image by the FIHS and BT, respectively, I = ( R + G + B ) / 3 is the intensity component of the low-resolution RGB image, and P a n is the high-resolution panchromatic image.
The fused images of various IHS-like and BT-like approaches can be evaluated in the IHS color space by the method reported in [19]. The performances for FIHS and BT methods are:
I F I H S = I + ( P a n I ) = P a n   vs .   I B T = P a n I × I = P a n
φ F I H S = φ H F I H S = H   vs .   H B T = H
S F I H S S = I I F I H S = I P a n   vs .   S B T S = I B T I = P a n I
where S = 1 3 min ( R , G , B ) R + G + B = I a I , and a = min ( R , G , B ) , which is the definition of saturation in the nonlinear RGB-IHS conversion model [20]. From Equations (2)–(4), it can be seen that the hue value of the fused image is the same as the original RGB image, the intensity and saturation values will be altered. Moreover, it can be seen that the saturation value is the main difference in color distortion between FIHS and BT methods. The value is compressed in FIHS when P a n is larger than I, but the value is stretched in BT. This means that the fused images by FIHS and BT suffer from the color distortion problems. Although, the saturation distortion is an important issue in image fusion, it seems negligible in NST because the saturation only affects the contrast of the stylized image without affecting the style in the image. From the color theory point of view, the replacement of the image will inevitably lead to color distortion. Using FIHS or BT to replace the luminance-only transfer is an effective method because they do not require coordinate conversion.

2.2.2. Color Enhance by ICDDS

As mentioned in Section 1, the S C images produced by the FST or MST are usually with lower contrast. The remote sensing images also have the similar problem. The multispectral bands are often highly correlated and thus unable to be adequately shown on RGB displays by independent contrast stretching, which is caused by the limit of sensor bandwidth. To solve this problem, the Intensity Conservation Direct Decorrelation Stretch (ICDDS) technique has been presented in [18]. ICDDS method is represented as
[ R I C D D S G I C D D S B I C D D S ] = I I k · a [ R k · a G k · a B k · a ]
Obviously, the ICDDS approach has a BT-like form, in which an IHS-like operation is included. The following properties can be derived from Equations (2)–(4),
I I C D D S = I
H I C D D S = H
S I C D D S S = S I C D D S S · S S = I I C D D S I · I I = ( I I k · a ) 2
It is important to recognize that both intensity and hue are preserved in the process and that only the saturation component is changed by controlling the k parameter. Evidently, ICDDS is a suitable tool for color enhancement.

3. The Proposed Methods

3.1. The Modified UST Method

Figure 2a is the C image, and the upper left corner is the S image, the Figure 2b is the S C image using complete framework of UST with α = 0.5 , and the Figure 2d is the S C image using the UST_5,4 with α = 1 (no blending). In comparison to Figure 2b,d, there is not much difference from the viewpoint of vision. However, Figure 2c is the S C image only using the single module (UST_5) with α = 1 . Although the stylized effect of Figure 2c is impressive, the image has lower contrast. To solve this problem, first of all, we would like to describe the whitening and coloring transforms of UST. Figure 3a is the Gaussian distribution of two-dimensional data. After principal component analysis (PCA), the distribution of Figure 3a is transformed to the distribution of Figure 3b, and this is then transformed to Figure 3c by whitening transform. Then, the coloring transform is an inverse transform module to transform Figure 3c to Figure 3b, and the inverse PCA transformation is used to transform Figure 3b to Figure 3a.
As shown in Figure 3c, when the coloring transform is used to transform to Figure 3b, the eigen matrix Λ is multiplied by the constant k (k > 1), and conduced to the Figure 3d distribution being extended. We refer to this phenomenon as enhance coloring transform. Then, the distribution of inverse PCA transformation is also extended, as shown in Figure 3e. As we can see in Figure 1b, the single UST module can be rewritten as follows:
f ^ C S = E S ( k D S ) 1 2 E S T f ^ C
E ( f ^ C S f ^ C S T ) = E S k 1 2 D S 1 2 E S T E ( f ^ C f ^ C T ) E S k 1 2 D S 1 2 E S T
= k E S D S 1 2 E S T I E S D S 1 2 E S T = k E ( f S f S T )
Equation (9) states that the amount of variance in the original digital image becomes larger, which indicates the image has higher contrast. The S C image only used the single module (UST_5) with α = 1 and k = 2 to obtain the desired results, as shown in Figure 4. The original UST architecture used the five UST modules to generate the style image. However, each module cannot reserve the full features of VGG because of the blending method. The proposed method only used one module (UST_5) without blending, the style image contains the whole features of VGG and obtains the higher contrast. In addition, the proposed method can be practiced in bigger-size images due to the lower computation cost and GPU memory.

3.2. The Post-Processing Framework

For NST, users hope to improve three issues: the color scheme, the strength of the style of the strokes, and the adjustment of image contrast. To satisfy the demands of users, the S C image from modified UST method (UST_5) combined with the image fusion and color enhancement methods to design a good post-processing framework to tackle the above-mentioned three issues simultaneously. Equation (10) can use the blending technique to re-adjust the mixing ratio between the C and S C images. A possible approach is to integrate the blending technique as shown below
[ R G B ] = k 1 · I B + ( 1 k 1 ) · I A I B k 2 · a B · [ R B k 2 · a B G B k 2 · a B B B k 2 · a B ]
where a B = min ( R B , G B , B B ) , the parameter k 1 ranging from 0 to 1 is for adjusting the blend ratio between the C and S C images to adjust the stroke strength, k 2 ranging from −1 to 1 is for controlling the degree of saturation compression or stretch. I B and I A are the C or S C images to solve the color scheme. Based on Equation (10), six different modules are as follows
  • The content-preserved ( C P ) module when k 1 = k 2 = 0 ,   I A is the content image, and I B is the stylized image.
  • The style-preserved ( S P ) module when k 1 = k 2 = 0 , I A is the stylized image, and I B is the content image.
  • The S P module with content/style blending when 0   k 1   1 , k 2 = 0 , I A is the stylized image, and I B is the content image.
  • The C P module with content/style blending when 0 k 1   1 , k 2 = 0 , I A is the content image, and I B is the stylized image.
  • The S P module with content/style blending and the contrast adjusting when 0 k 1 1 , 1 < k 2 < 1 , I A is the stylized image, and I A is the content image.
  • The C P module with content/style blending and the contrast adjusting when 0 k 1 1 , 1 < k 2 < 1 , I A is the content image, and I B is the stylized image.
The two most important modules of the proposed approach are (5) and (6). Modules (5) and (6) can not only combine the spatial details of the style/content image with the color information of the content/style image, but also control the compression or stretch of the image saturation. By making some adjustments the value of k 2 , we can solve the problem of low contrast. Thus, they can provide multiple-choice methods for automatic colorization in the new stylized image which will be shown in the next section.

4. Experiment Results and Analysis

The proposed versatile framework has been implemented by a window-based program, as shown in Figure 5. Based on the concept of “What You See Is What You Get” (WYSIWYG), the window consists of five sub-windows which are called the main window and four Thumbnail Previews, respectively. Initially, the S C and C images are read and displayed on the upper left and upper right sides of the main window. Next, select the content- or the style-preserved module and execute. The resultant C P or S P images will be displayed at the bottom left or bottom right side of the main window. The default is module (1), however, the user can easily change the module and adjust the parameters k 1 and k 2 to find the most satisfactory module. The processing results are obtained and analyzed in real time. In the following, several experiments are conducted.

Effect of Different Modules of the Post-Processing Framework

The first experiment demonstrates the operations and corresponding results of the proposed versatile framework. The C image is a digital painting image, and its S C image is generated by the UST_5. The S C , C, and the resultant C P and S P images are shown in Figure 5, respectively. Compared to the S C and C images, both the C P and S P images achieve their intended functionality. By the default settings in modules (1), we have k 1 = k 2 = 0 . Thus, the C P image can keep the spatial details of the C image and maintain the color information of the S C image while the S P image is the result of the opposite function provided by modules (2).
To evaluate modules (3) and (4), the parameter k 1 is set between 0 and 1 to obtain an image with different mixing effects. The C P images with k 1 = 0.2 ,   0.5 ,   and   0.8 from left to right in sequence are displayed on the upper row in Figure 6. The bottom row is the S P images with the same k 1 values as C P . Compared to the original S C and C images in Figure 6, the style effect in the C P images are increased gradually by increasing the k 1 value. On the contrary, the style effects in the S P images are then decreased gradually. As expected, both modules (3) and (4) achieve the design goals, this allows the user to flexibly choose his favorite composition of the C and S C images.
In order to make the image contrast more variable, the modules (5) and (6), extended from the modules (3) and (4), can perform contrast reduction or enhancement. First, set the value of k 1 by choosing a favorite type of mixing, and then adjust the k 2 value to control the required saturation. When k 2 < 0 , the contrast is reduced, the smaller the k 2 is, the lower the contrast of the image, but when k 2 > 0 , the effect is just the opposite, that is, the contrast is enhanced. The resultant images with different k 2 values are demonstrated in Figure 7. The first row is displayed the reduced-contrast C P images with a fixed k 1 = 0.5 and k 2 = 0.8 , 0.5 ,   and 0.2 , from left to right in sequence. The second row shows the enhanced-contrast C P with the same k 1 value, but k 2 = 0.2 ,   0.5 ,   and   0.8 , respectively. The third and fourth rows are the contrast-reduced and -enhanced S P images with the same k 1 and k 2 values as the first and second rows, respectively. As can be seen in Figure 7, the saturation of each image has reached the desired effect. This demonstrates that Equation (3) can execute the functions of replacement, mixing, and contrast adjustment to improve the neural stylized image.
Artists often need to render different colors for sketches or monotonous textures which is time-consuming and tedious. In the second experiment, we will show that the S P module (5) is able to provide the color re-rending effect. In addition, we used modified UST features (UST_5) and shown the corresponding stylized results in the third row of Figure 8. In this case, it clearly shows that the UST_5 features capture more complicated local structure. However, the C P images not only generate better style quality but also preserve content shape. Because the parameters are fixed, i.e., k 1   = 0 and k 2 = 0.8, module (6) can be executed by a batch operation so that this re-coloring work will be efficient.
The third experiment will show the results of a real digital painting processed by the UST_5 and the proposed modules (5) and (6). Figure 9 shows the results of C, S C , C P , and S P images. The parameters are k 1 = 0.4, k 2 = 0.2 for module (5), and k 1 = 0.5, k 2 = −0.4 for module (6). As can be seen in Figure 9, the original C image is simple and elegant, showing a relaxed and pleasant style. In contrast, the S C image processed by the UST_5 has a larger contrast in color and a clearly visible texture, showing a cool style. By using modules (5) and (6), the C P , and S P images are created by balancing the visual effects. They provide better style quality in terms of preserving content shape and keeping long style coherence.
In order to prove the validity of general landscape photos, we used the stylized image produced by UST_5,4 (described in Section 3.1) with Equation (10) in the following examples. As it can be seen in Figure 10a,e are two pairs of images including the C and S images. Figure 10b,f are the S C images produced by the UST_5,4 method. The two stylized images have extracted enough features from the original C and S images and the effects are very abstract. By using Equation (6), the corresponding C P and S P images (both of k 1 = 0.5, k 2 = 0) are shown in Figure 10c,d,g,h, respectively. Finally, those stylized images have a more visual-pleasing effect. Generally speaking, the quality of the artistic style image varies from person to person. For the stylized image, some people prefer more content-preserved images, and some others prefer more style-preserved images. Herein, our proposed method successfully combined the modified UST method and post-processing method to meet personal preference.
In addition, most NST methods allow us to recompose the content of general natural image in another style, but this is dissatisfied with the portrait images. Figure 11a is the original portrait image, and Figure 11b–d show the different styles of the original portrait image by using the different values of k 1 and k 2 . In this case, we used the stylized image produced by UST_3,2,1 (UST_5,4 are unused). As we can see in Figure 11b–d, the three portrait images have outstanding performance in the color rendering, the style strokes, and the image contrast. For NST application, modified UST is a superior method. Based on VGG architecture, different applications should use different encoder/decoder pairs rather than using the original UST method.

5. Conclusions

More recently, NST-based style transfer algorithms have demonstrated very exciting results that have attracted academic and industrial attention. To obtain more visual effects, researchers have been using more complex algorithms to improve CNN’s learning or network architecture. However, the effect of these methods seems not satisfactory. Since the qualitative assessment is highly subjective, there is no metric to evaluate artistic style transfer effects. To give users more choices, a versatile post-processing framework has been proposed. In this framework, the blending technique, BT, and ICDDS methods have been fully integrated and implemented by a window-based program. It is easy to use and able to provide more visual effects for the stylized images, and also can integrate into the UST method effectively. Especially in the second experiment, we provide art workers a good solution for processing massive color rendering. When the deep learning technology is limited inherently, traditional methods such as this work may sometimes be able to solve the problem. Our proposed method is suitable for stylized images generated by any NST method, but it also works similarly to the MST method, which mixes two different stylized images.
This paper has investigated MST, and UST, and proposes a modified UST method to achieve more visual effects for the stylized images, which makes our method more applicable to real-world image generation scenarios where style needs to be defined by the user. However, the proposed method would fall short due to massive memory cost when processing ultra-high-resolution images. Therefore, how to avoid the memory problem caused by ultra-high-resolution images will be the subject of our future research work.

Author Contributions

Data curation, H.-C.H.; Formal analysis, C.-H.C.; Methodology, C.-C.L. and T.-M.T.; Resources, C.-B.H.; Software, J.-C.L. and T.-M.T.; Supervision, T.-M.T. and H.-C.H.; Writing—original draft, J.-C.L.; Writing—review & editing, C.-H.C. and H.-C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, X.; Yang, Y.; Li, Z.; Ning, X.; Qin, Y.; Cai, W. An Improved Encoder-Decoder Network Based on Strip Pool Method Applied to Segmentation of Farmland Vacancy Field. Entropy 2021, 23, 435. [Google Scholar] [CrossRef] [PubMed]
  2. Lopac, N.; Hržić, F.; Vuksanović, I.P.; Lerga, J. Detection of Non-Stationary GW Signals in High Noise From Cohen’s Class of Time–Frequency Representations Using Deep Learning. IEEE Access 2022, 10, 2408–2428. [Google Scholar] [CrossRef]
  3. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient Detection of Knee Anterior Cruciate Ligament from Magnetic Resonance Imaging Using Deep Learning Approach. Diagnostics 2021, 11, 105. [Google Scholar] [CrossRef] [PubMed]
  4. Gatys, L.A.; Ecker, A.S.; Bethge, M. A neural algorithm of artistic style. arXiv 2015, arXiv:1508.06576. [Google Scholar] [CrossRef]
  5. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
  6. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
  7. Dumoulin, V.; Shlens, J.; Kudlur, M. A learned representation for artistic style. arXiv 2016, arXiv:1610.07629. [Google Scholar]
  8. Prisma App. Available online: https://prisma-ai.com/ (accessed on 8 July 2016).
  9. Deepart.io. Available online: https://alternativeto.net/software/deepart-io/ (accessed on 26 April 2022).
  10. Jing, Y.C.; Yang, Y.H.; Feng, Z.L.; Ye, J.W.; Song, M.L. Neural Style Transfer: A Review. arXiv 2017, arXiv:1705.04058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Li, Y.; Fang, C.; Yang, J.; Wang, Z.; Lu, X.; Yang, M.H. Universal Style Transfer via Feature Transforms. Adv. Neural Inf. Process. Syst. (NIPS) 2017, 30, 1–11. [Google Scholar]
  12. Wang, H.; Li, Y.; Wang, Y.; Hu, H.; Yang, M.H. Collaborative Distillation for Ultra-Resolution Universal Style Transfer. arXiv 2020, arXiv:2003.08436. [Google Scholar]
  13. Gatys, L.A.; Bethge, M.; Hertzmann, A.; Shechtman, E. Preserving Color in Neural Artistic Style Transfer. arXiv 2016, arXiv:1606.05897. [Google Scholar]
  14. Lee, S.H. A Fast Neural Style Transfer Implement with Keras 2. Available online: https://github.com/misgod/fast-neural-style-keras (accessed on 6 December 2019).
  15. Gatys, L.A.; Ecker, A.S.; Bethge, M.; Hertzmann, A.; Shechtman, E. Controlling perceptual factors in neural style transfer. arXiv 2017, arXiv:1611.07865. [Google Scholar]
  16. Jing, Y.C.; Liu, Y.; Yang, Y.; Feng, Z.; Yu, Y.; Tao, D.; Song, M. Stroke Controllable Fast Style Transfer with Adaptive Receptive Fields. arXiv 2018, arXiv:1802.07101. [Google Scholar]
  17. Tu, T.M.; Huang, P.S.; Hung, C.L.; Chang, C.P. A fast intensity-hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE GeoSci. Remote Sens. Lett. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  18. Liu, J.G.; Moore, J.M. Direct decorrelation Stretch Technique for RGB colour composition. Int. J. Remote Sens. 1996, 17, 1005–1018. [Google Scholar]
  19. Tu, T.M.; Hsu, C.L.; Tu, P.Y.; Lee, C.H. An adjustable pan-sharpening approach for IKONOS/QuickBird/GeoEye-1/WorldView-2 imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 125–134. [Google Scholar] [CrossRef]
  20. Ledley, R.S.; Buas, M.; Golab, T.J. Fundamentals of true-color image processing. Intl. Conf. Pattern Recognit. 1990, 1, 791–795. [Google Scholar]
Figure 1. (a) UST used the VGG network as an encoder/decoder pair to achieve image reconstruction; (b) the single UST module; (c) the complete framework of UST.
Figure 1. (a) UST used the VGG network as an encoder/decoder pair to achieve image reconstruction; (b) the single UST module; (c) the complete framework of UST.
Applsci 12 06710 g001
Figure 2. (a) is the C image, and the upper left corner is the S image, (b) is the S C image using the UST_5~1 and α = 0.5 , (c) is the S C image using the UST_5 and α = 1 , (d) is the S C image using the UST_5,4 and α = 1 .
Figure 2. (a) is the C image, and the upper left corner is the S image, (b) is the S C image using the UST_5~1 and α = 0.5 , (c) is the S C image using the UST_5 and α = 1 , (d) is the S C image using the UST_5,4 and α = 1 .
Applsci 12 06710 g002
Figure 3. (a) A 2D Gaussian distribution, (b) PCA transform, (c) whitening transform, (d) enhance coloring transform, (e) 2D Gaussian distribution reconstruction.
Figure 3. (a) A 2D Gaussian distribution, (b) PCA transform, (c) whitening transform, (d) enhance coloring transform, (e) 2D Gaussian distribution reconstruction.
Applsci 12 06710 g003
Figure 4. The S C image used the UST_5 module with α = 1 and k = 2 .
Figure 4. The S C image used the UST_5 module with α = 1 and k = 2 .
Applsci 12 06710 g004
Figure 5. Upper left—the S C image; upper right—the C image; bottom left—the C P image; bottom right—the S P image; and middle—Main window.
Figure 5. Upper left—the S C image; upper right—the C image; bottom left—the C P image; bottom right—the S P image; and middle—Main window.
Applsci 12 06710 g005
Figure 6. Upper row: the mixed effects for the C P images with k 1 = 0.2 ,   0.5 ,   and   0.8 , from left to right. Bottom row: the S P images with k 1 = 0.2 ,   0.5 ,   and   0.8 .
Figure 6. Upper row: the mixed effects for the C P images with k 1 = 0.2 ,   0.5 ,   and   0.8 , from left to right. Bottom row: the S P images with k 1 = 0.2 ,   0.5 ,   and   0.8 .
Applsci 12 06710 g006
Figure 7. First row: the contrast-reduced C P images with a fixed k 1 = 0.5 and k 2 = −0.8, −0.5, and −0.2, from left to right. Second row: the contrast-enhanced C P images with k 1 = 0.5 and k 2 = 0.2, 0.5, and 0.8, respectively. Third row: the contrast-reduced S P images with the same k 1 and k 2   as the first row. Fourth row: the contrast-enhanced S P images with the same k 1 and k 2 as the second row.
Figure 7. First row: the contrast-reduced C P images with a fixed k 1 = 0.5 and k 2 = −0.8, −0.5, and −0.2, from left to right. Second row: the contrast-enhanced C P images with k 1 = 0.5 and k 2 = 0.2, 0.5, and 0.8, respectively. Third row: the contrast-reduced S P images with the same k 1 and k 2   as the first row. Fourth row: the contrast-enhanced S P images with the same k 1 and k 2 as the second row.
Applsci 12 06710 g007
Figure 8. The middle of first row: the C image. Second row: three S C images. Third row: the three corresponding C P images with k 1 = 0 and k 2 = 0.8.
Figure 8. The middle of first row: the C image. Second row: three S C images. Third row: the three corresponding C P images with k 1 = 0 and k 2 = 0.8.
Applsci 12 06710 g008
Figure 9. (a)the C image, (b) the S C image, (c) the C P image ( k 1 = 0.4, k 2 = 0.2), and (d) the S P   image ( k 1 = 0.5, k 2 = −0.4).
Figure 9. (a)the C image, (b) the S C image, (c) the C P image ( k 1 = 0.4, k 2 = 0.2), and (d) the S P   image ( k 1 = 0.5, k 2 = −0.4).
Applsci 12 06710 g009
Figure 10. (a,e): two image pairs include the C and S images. (b,f): the two stylized images produced by the UST method. (c,d,g,h): the C P and S P images produced by our method.
Figure 10. (a,e): two image pairs include the C and S images. (b,f): the two stylized images produced by the UST method. (c,d,g,h): the C P and S P images produced by our method.
Applsci 12 06710 g010
Figure 11. (a) is the original portrait image, (bd) are the three different stylized images with different values of k 1 and k 2 .
Figure 11. (a) is the original portrait image, (bd) are the three different stylized images with different values of k 1 and k 2 .
Applsci 12 06710 g011
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, C.-C.; Hsu, C.-B.; Lee, J.-C.; Chen, C.-H.; Tu, T.-M.; Huang, H.-C. A Variety of Choice Methods for Image-Based Artistic Rendering. Appl. Sci. 2022, 12, 6710. https://doi.org/10.3390/app12136710

AMA Style

Lin C-C, Hsu C-B, Lee J-C, Chen C-H, Tu T-M, Huang H-C. A Variety of Choice Methods for Image-Based Artistic Rendering. Applied Sciences. 2022; 12(13):6710. https://doi.org/10.3390/app12136710

Chicago/Turabian Style

Lin, Chiu-Chin, Chih-Bin Hsu, Jen-Chun Lee, Chung-Hsien Chen, Te-Ming Tu, and Huang-Chu Huang. 2022. "A Variety of Choice Methods for Image-Based Artistic Rendering" Applied Sciences 12, no. 13: 6710. https://doi.org/10.3390/app12136710

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop