• No results found

Deep Colorization with Facial Gender Recognition

N/A
N/A
Protected

Academic year: 2021

Share "Deep Colorization with Facial Gender Recognition"

Copied!
8
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Deep Colorization with Facial Gender Recognition

Bachelor’s Project Thesis

Jonathan Hogervorst, s2357712, jonathan@hogervorst.info Supervisors: Emmanuel Okafor & dr. Marco Wiering

Abstract. Recent research suggests that colorization models have the capability of generating plausible color versions from grayscale images. In this paper, we investigate whether colorization prior to gender classification improves classification performance on the FERET grayscale face dataset. For this, we colorize the images using an existing Lab colorization model, both with and without class rebalancing, and our novel HSV colorization model without class rebalancing. Then we construct gender classification models on the grayscale and colorized datasets using a reduced GoogLeNet convolutional neural network. Several models are trained using different loss func- tions (cross entropy loss, hinge loss) and gradient optimization solvers (Nesterov’s Accelerated Gradient Descent, Stochastic Gradient Descent), initialized using both random and pre-trained weights. Finally, we compare the gender classification accuracies of the models when applied to the face image color variants. The best performances are obtained by models initialized using pre-trained weights, and models using colorization without class rebalancing.

1 Introduction

Gender recognition is an interesting problem with various applications, like video surveillance and au- thentication. Furthermore, gender plays an impor- tant role in human interaction. Gender recognition can therefore also be a useful method of improv- ing human-computer interaction [34]. While gen- der recognition is a simple task for humans, it is difficult to create a system that can perform this task with sufficient performance. Additionally, face images may only be available in grayscale. For ex- ample, simple surveillance cameras and night vision cameras often do not provide color images, and his- toric photo and video material is also not available in color. Gender recognition may be more difficult on grayscale images because color images contain more information that may aid recognition.

Much research has been done into face gender recognition systems, applying various kinds of tech- niques [4, 20, 25, 40]. A successful approach is the use of support vector machines on face images [13, 18, 22, 31, 34]. Recently, convolutional neural networks (CNNs) have been tried as an approach to gender classification as well [16, 21, 34].

Research has examined the development of sys- tems that can predict color information given a

grayscale image using different approaches. One such approach is the transfer of color from a source image [5, 10, 12, 28, 36], requiring one or more color source images featuring a scene sim- ilar to the image to be colorized. Another ap- proach is colorization of grayscale images using color scribbles [17, 19, 27, 30, 37]. While such sys- tems do not need similar source images, they re- quire the user to scribble appropriate colors onto the grayscale image. Recent research yielded col- orization models which use large datasets of color images [3, 6–8, 11, 15, 39]. These systems colorize images automatically, without any user input.

Contributions. In this paper, we investigate the influence of colorization prior to gender classi- fication on the FERET grayscale face dataset [26].

The images are colorized using a colorization CNN [39] in Caffe [14]. We use the provided Lab col- orization model, which predicts CIE Lab color val- ues given a grayscale input image. We use the pro- vided models with and without class rebalancing, which compensates for the fact that some color ar- eas in the ab output space occur less often in pho- tos. Furthermore, we adapt the CNN to predict HSV color values and train our own HSV coloriza- tion model without class rebalancing on the Im- ageNet dataset [29]. We then train gender clas-

(2)

sification models on the various face image color variants. For this, we use a reduced version of the GoogLeNet CNN [32] in Caffe. We compare two classification loss functions: cross entropy loss and hinge loss. Additionally, we compare two gradient optimization solvers: Nesterov’s Accelerated Gra- dient Descent (NAGD) [23] and Stochastic Gradi- ent Descent (SGD). Gender classification models are trained both from scratch, using random Xavier initialization [9], and fine-tuned from a pre-trained GoogLeNet model. We compare classification ac- curacies of the eight resulting gender model vari- ants for the four image color variants using ten-fold Monte Carlo cross-validation (MCCV).

Outline. In Section 2, we present our method:

first, we train an HSV colorization model in addi- tion to the existing Lab colorization models (Sec- tion 2.1); then, we colorize the grayscale images to new, colorized versions of the dataset using the colorization models (Section 2.2); and lastly, we use the colorized images in our gender classifica- tion models (Section 2.3). In Section 3, we provide our results. Finally, in Section 4, we discuss our conclusions.

2 Methods

2.1 Colorization

To perform colorization, we used a colorization CNN [39] implemented in the deep learning frame- work Caffe [14]. This CNN was created with the goal of predicting plausible color versions of grayscale images given as input. The CNN em- ploys the CIE Lab color space: it predicts the a and b channels of an image, given its L channel.

The L channel contains information about the im- age’s lightness or brightness — it thus provides a grayscale representation of the image.

The colorization CNN is unique due to its tai- lored loss function. Image color prediction is a mul- timodal problem [3]: objects in an image often have multiple plausible colors. By using a loss function like the Euclidean loss, the optimal solution would be the mean of all plausible color values, resulting in grayish colors [39]. To prevent this, the coloriza- tion CNN predicts a probability distribution over the possible colors. For this, the in-gamut ab out- put space is divided into 313 bins, over which the

probabilities are calculated. In addition, the loss function applies class rebalancing. This is done to compensate for the fact that some color values in the ab output space occur less often in photos.

Class rebalancing reweighs the loss of pixels based on the rarity of the color during training, ensuring that the model can still predict less common color values.

The Lab colorization model was trained on the training set of the ImageNet dataset [29], contain- ing 1.3M color images, in 450k iterations [39]. The ImageNet dataset consists of color images of many different types of objects collected from the In- ternet. We used the trained Lab models with1 and without class rebalancing2. More information about the loss function and training parameters of the Lab colorization model can be found in [39].

Moreover, we adopted the colorization CNN to create our own HSV colorization model. In this CNN, we predict the H and S channels of an im- age given its V channel. Similar to the L channel in the Lab color space, the V channel in the HSV color space provides a grayscale representation of an image. We applied binning over the HS output space using 18 equally sized bins in both directions, yielding 324 bins in total. We did not apply class re- balancing. We trained our HSV colorization model on the training set of the ImageNet dataset for 300k iterations using the unmodified training parameters of the Lab colorization CNN. This took ∼ 14 days on a single core of an NVIDIA Tesla K40 GPU.

2.2 Dataset

We employed a selection of 1242 images of the grayscale face image dataset FERET [26]. This dataset consists of photos of subjects from several angles with different facial expressions. Our selec- tion consisted of frontal images and images in which subjects had their face slightly turned. Our images were cropped to squares with the faces aligned in the middle and contained one color channel. From our selection of 1242 images, we put 255 aside as test set. We randomly split the remaining images into a training set of 789 images (80%) and a vali- dation set of 198 images (20%). This random split- ting procedure was repeated to create ten train- ing/validation set pairs for MCCV.

1colorization release v2.caffemodel

2colorization release v2 norebal.caffemodel

(3)

Figure 1: Example images from the FERET face dataset [26] used in our research. From top to bottom:

original grayscale, Lab colorized with class rebalanc- ing, Lab colorized without class rebalancing, and HSV colorized without class rebalancing.

We then colorized the images. To colorize an im- age using the Lab colorization models, we loaded it using the Caffe [14] function io.load image(), automatically converting it to three-channel RGB data in grayscale. Then we converted it to Lab us- ing the scikit-image [35] function color.rgb2lab() and provided the L channel as input to the coloriza- tion model. The output of the model, being the predicted ab color values, was combined with the calculated L channel. The resulting Lab image was then converted to RGB using color.lab2rgb() and stored. We repeated this for all images. Sim- ilarly, each image was colorized using our HSV colorization model: we converted it to HSV us- ing color.rgb2hsv(); provided the V channel to the model; combined the V channel with the pre- dicted HS channels; converted it to RGB using color.hsv2rgb(); and stored it.

We thus had four variants of the dataset: origi- nal grayscale, Lab colorized with class rebalancing, Lab colorized without class rebalancing, and HSV colorized without class rebalancing. Some example images are shown in Fig. 1.

2.3 Gender Recognition

Network

We performed gender classification using a reduced version of the GoogLeNet CNN [32] implemented in Caffe [14]. GoogLeNet contains Inception mod- ules, each consisting of six convolution layers and one pooling layer, outputting the concatenation of its contained layers. The original CNN is 27 layers deep and contains nine Inception modules.

It yielded an excellent performance on the Im- ageNet Large-Scale Visual Recognition Challenge 2014 [32].

We noticed that the extensive GoogLeNet struc- ture would not be required for our gender classifi- cation task. Therefore we reduced it, removing six of the nine Inception layers. The resulting reduced GoogLeNet still provided good performance for our task, while saving ∼ 40% computation time during training. The structure of our reduced GoogLeNet is shown in Figure 2.

Loss Function

Loss functions indicate the difference between pre- dicted values and target values in supervised learn- ing. During training, the loss is minimized by the CNN to get the predictions closer to the target val- ues, improving the model’s performance. The orig- inal GoogLeNet contains a softmax classification layer which employs the cross entropy loss func- tion (SoftmaxWithLossLayer in Caffe). We also applied this in our reduced CNN. Additionally, we created a variant of our CNN using the hinge loss function (HingeLossLayer). More information can be found in [1].

Gradient Solver

During training, the model’s parameters are up- dated to minimize the loss. The gradient solver specifies how the CNN updates its parameters.

We employed two different gradient optimization solvers: NAGD [23] and SGD. More information can be found in [2].

Initialization

The Caffe implementation of GoogLeNet employs Xavier initialization [9]. This algorithm initial- izes the model’s weights from a distribution around

(4)

Input

Conv 1 MaxPool 1 Conv 2

(depth 2) MaxPool 2

Inception 3a Inception 3b MaxPool 3 Inception 4a

AvgPool 1 Conv 3 FC 1

(70% dropout) FC 2

Softmax

(cross entropy loss) Hinge loss Figure 2: Structure of our reduced GoogLeNet CNN for gender classification

zero, with a variance based on the number of in- put/output neurons in the CNN. We employed this technique when training our models from scratch.

In addition to training from scratch, we also trained models by fine-tuning from pre-trained weights. In this case, the weights in our models were initialized using weights from the pre-trained Caffe GoogLeNet model3, which was trained on the ImageNet dataset similar to [32].

Training Parameters

When training our models, we set the base learn- ing rate to 0.001 (base lr parameter in Caffe’s solver.prototxt file). After each step of 10k training iterations (stepsize; lr policy=step) we lowered the learning rate by multiplying it with 0.96 (gamma). Furthermore we used a momentum of 0.9 (momentum) and a weight decay factor of 0.0002 (weight decay).

We trained gender classification models on the ten training sets for each color variant of our dataset. We trained each scratch model for 30k iterations, which took ∼ 1.5 hour on a single core of an NVIDIA Tesla K40 GPU. We trained each fine-tuned model for 20k iterations, which took ∼ 1 hour on the same GPU.

3bvlc googlenet.caffemodel

3 Results

The gender classification learning curves of our models after various numbers of training iterations are shown in Figure 3. The scratch models, which used Xavier weight initialization, show significant improvements after the first and second training epoch; after that, the accuracies slightly improve further. The same applies to some of the fine-tuned models, which were initialized using pre-trained weights. Other fine-tuned models only show slight improvements from the first training epoch on. Be- cause the models keep improving until the final training epoch, the models at the final training epoch will be used for further reporting.

The gender classification test accuracies of our various models are reported in Table 1. The results show that the maximum classification accuracy is obtained by the fine-tuned model using hinge loss and SGD, applied on the Lab colorized version of the dataset without class rebalancing, yielding 99.78%.

When we compare the accuracies of our models on the Lab and HSV colorized images without class balancing, they show a significant increase over the models with class rebalancing. This implies that colorization without class rebalancing works better for this study. Furthermore, the models without class rebalancing only show a slight improvement over some of the original grayscale image models.

(5)

Scratch, cross entropy loss, NAGD Fine-tuning, cross entropy loss, NAGD Scratch, cross entropy loss, SGD Fine-tuning, cross entropy loss, SGD

Scratch, hinge loss, NAGD Fine-tuning, hinge loss, NAGD Scratch, hinge loss, SGD Fine-tuning, hinge loss, SGD

1 2 3 4 5 6

97 98 99 100

Number of epochs

Accuracy(%)

(a) Original grayscale

1 2 3 4 5 6

94 95 96 97 98 99 100

Number of epochs

Accuracy(%)

(b) Lab colorized w/ class rebalancing

1 2 3 4 5 6

97 98 99 100

Number of epochs

Accuracy(%)

(c) Lab colorized w/o class rebalancing

1 2 3 4 5 6

97 98 99 100

Number of epochs

Accuracy(%)

(d) HSV colorized w/o class rebalancing Figure 3: Average gender classification accuracy of ten-fold MCCV on the test set after training for a number of epochs (1 epoch ≡ 5k iterations). The scratch CNN models were trained for 30k iterations, the fine-tuned CNN models for 20k iterations.

Table 1: Average gender classification accuracy in % (± standard deviation) of ten-fold MCCV on the test set after training for 30k (scratch) or 20k (fine-tuning) iterations.

Training Loss Function

Gradient

Solver Grayscale Lab Color’ed w/ class rebal.

Lab Color’ed w/o class rebal.

HSV Color’ed w/o class rebal.

From scratch (Xavier init.)

Cross entropy loss

NAGD 98.27 ± 0.50 95.56 ± 0.66 98.71 ± 0.46 98.58 ± 0.55 SGD 98.36 ± 0.40 95.02 ± 0.48 98.58 ± 0.62 98.49 ± 0.80 Hinge loss NAGD 98.58 ± 0.56 95.24 ± 0.63 98.62 ± 0.58 98.62 ± 0.42 SGD 98.49 ± 0.41 95.20 ± 0.91 98.58 ± 0.62 98.67 ± 0.40 Fine-

tuning (pre- trained weights)

Cross entropy loss

NAGD 99.51 ± 0.24 98.00 ± 0.46 99.51 ± 0.31 99.29 ± 0.22 SGD 99.51 ± 0.42 97.82 ± 0.37 99.60 ± 0.24 99.20 ± 0.33 Hinge loss NAGD 99.64 ± 0.27 97.96 ± 0.80 99.51 ± 0.42 99.16 ± 0.42 SGD 99.51 ± 0.31 97.91 ± 0.56 99.78 ± 0.30 99.47 ± 0.27

(6)

In general, our fine-tuned models showed more promising performance than their scratch counter- parts on all versions of the dataset we examined.

The results of the models using the different loss functions or gradient solvers show only very small differences — no conclusions can be drawn from this.

4 Conclusions

This study comprised a comprehensive evaluation using different variants of a reduced GoogLeNet CNN to determine the gender classification per- formance on original and colorized versions of the FERET grayscale face dataset. The results show that class rebalancing in colorization does not im- prove the gender classification performance in this study since all models without class rebalancing yielded better results than the models with class rebalancing. Additionally, fine-tuning from pre- trained weights seems to be the best approach in this study: all fine-tuned models yielded better performance than their scratch counterparts, while also requiring less training time.

In general, the gender classification accuracies from our models were all very high and similar.

While we observed some improvements in models using colorization, they were in the order of tenths of a percentage point. To attribute significant im- provement to colorization prior to gender classifi- cation, there should be a greater difference in accu- racy between the grayscale and colorized models.

A reason for the high classification accuracies could be that the model used some individual char- acteristics of subjects, rather than gender charac- teristics only. Because subjects were photographed from multiple angles for the FERET dataset, differ- ent images of the same face could have ended up in different sets, like our training and test set. While such images were not identical, they may have al- lowed the model to use face-specific characteristics.

Research has shown that color information im- proves the performance of face recognition systems [33]. Moreover, psychological research suggests that humans use color information in gender recog- nition if other information is unavailable [24, 38].

Our dataset consisted of high quality frontal whole face photos. Our approach of colorization prior to gender classification might be better applicable to

more complex datasets. In that case, the grayscale baseline results would be worse, and colorization could bring actual improvement.

Future work should examine multi-orientation data augmentation using colorized images, which may improve the performance further.

Acknowledgements

I would like to thank the Center for Information Technology of the University of Groningen for their support and for providing access to the Peregrine high performance computing cluster.

References

[1] Berkeley Vision and Learning Cen- ter. Caffe documentation. http:

//caffe.berkeleyvision.org/doxygen/, 2017. Accessed: 2017-03-17.

[2] Berkeley Vision and Learning Cen- ter. Caffe tutorial: Solver. http:

//caffe.berkeleyvision.org/tutorial/

solver.html, 2017. Accessed: 2017-03-17.

[3] Guillaume Charpiat, Matthias Hofmann, and Bernhard Sch¨olkopf. Automatic image col- orization via multimodal predictions. In David Forsyth, Philip Torr, and Andrew Zisserman, editors, European Conference on Computer Vision 2008, volume 5304 of Lecture Notes in Computer Science, pages 126–139, Berlin, Hei- delberg, 2008. Springer.

[4] Rama Chellappa, Charles L. Wilson, and Saad Sirohey. Human and machine recognition of faces: a survey. In Proceedings of the IEEE, volume 83, pages 705–741. IEEE, 1995.

[5] Tongbo Chen, Yan Wang, Volker Schillings, and Christoph Meinel. Grayscale image mat- ting and colorization. In 6th Asian Conference on Computer Vision, pages 1164–1169, 2004.

[6] Zezhou Cheng, Qingxiong Yang, and Bin Sheng. Deep colorization. In 2015 IEEE In- ternational Conference on Computer Vision, pages 415–423. IEEE, 2015.

(7)

[7] Ryan Dahl. Automatic colorization. http://

tinyclouds.org/colorize/, 2016. Accessed:

2017-02-14.

[8] Aditya Deshpande, Jason Rock, and David Forsyth. Learning large-scale automatic im- age colorization. In 2015 IEEE International Conference on Computer Vision, pages 567–

575. IEEE, 2015.

[9] Xavier Glorot and Yoshua Bengio. Under- standing the difficulty of training deep feed- forward neural networks. In Yee Whye Teh and Mike Titterington, editors, 13th Inter- national Conference on Artificial Intelligence and Statistics, volume 9 of JMLR Workshop and Conference Proceedings, pages 249–256.

PMLR, 2010.

[10] Raj Kumar Gupta, Alex Yong-Sang Chia, Deepu Rajan, Ee Sin Ng, and Huang Zhiyong.

Image colorization using similar images. In 20th ACM International Conference on Multi- media, pages 369–378, New York, 2012. ACM.

[11] Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa. Let there be color!: Joint end-to-end learning of global and local image priors for automatic image colorization with simultane- ous classification. In ACM SIGGRAPH 2016, volume 35 of ACM Transactions on Graphics, New York, 2016. ACM.

[12] Revital Irony, Daniel Cohen-Or, and Dani Lischinski. Colorization by example. In Kavita Bala and Philip Dutr´e, editors, 16th Euro- graphics Symposium on Rendering. The Euro- graphics Association, 2005.

[13] Taskeed Jabid, Md. Hasanul Kabir, and Ok- sam Chae. Gender classification using lo- cal directional pattern (LDP). In 20th Inter- national Conference on Pattern Recognition, pages 2162–2165. IEEE, 2010.

[14] Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Gir- shick, Sergio Guadarrama, and Trevor Dar- rell. Caffe: Convolutional architecture for fast feature embedding. In 22nd ACM Interna- tional Conference on Multimedia, pages 675–

678, New York, 2014. ACM.

[15] Gustav Larsson, Michael Maire, and Gregory Shakhnarovich. Learning representations for automatic colorization. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, European Conference on Computer Vision 2016, volume 9908 of Lecture Notes in Computer Science, pages 577–593, Cham, 2016. Springer.

[16] Gil Levi and Tal Hassncer. Age and gender classification using convolutional neural net- works. In 2015 IEEE Conference on Com- puter Vision and Pattern Recognition Work- shops, pages 34–42. IEEE, 2015.

[17] Anat Levin, Dani Lischinski, and Yair Weiss.

Colorization using optimization. In John C.

Hart, editor, ACM SIGGRAPH 2004, vol- ume 23 of ACM Transactions on Graphics, pages 689–694, New York, 2004. ACM.

[18] Hui-Cheng Lian and Bao-Liang Lu. Multi- view gender classification using local binary patterns and support vector machines. In Jun Wang, Zhang Yi, Jacek M. Zurada, Bao-Liang Lu, and Hujun Yin, editors, 3rd International Symposium on Neural Networks, volume 3972 of Lecture Notes in Computer Science, pages 202–209, Berlin, Heidelberg, 2006. Springer.

[19] Qing Luan, Fang Wen, Daniel Cohen-Or, Lin Liang, Ying-Qing Xu, and Heung-Yeung Shum. Natural image colorization. In Jan Kautz and Sumanta Pattanaik, editors, 18th Eurographics Symposium on Rendering, pages 309–320. The Eurographics Association, 2007.

[20] Erno M¨akinen and Roope Raisamo. An ex- perimental comparison of gender classifica- tion methods. Pattern Recognition Letters, 29(10):1544–1556, 2008.

[21] Jordi Mansanet, Alberto Albiol, and Roberto Paredes. Local deep neural networks for gen- der recognition. Pattern Recognition Letters, 70:80–86, 2016.

[22] Baback Moghaddam and Ming-Hsuan Yang.

Gender classification with support vector ma- chines. In IEEE International Conference on Automatic Face and Gesture Recognition, pages 306–311. IEEE, 2000.

(8)

[23] Yurii Nesterov. A method of solving a con- vex programming problem with convergence rate O(1/k2). Soviet Mathematics Doklady, 27(2):372–376, 1983.

[24] Adrian Nestor and Michael J. Tarr. Gender recognition of human faces using color. Psy- chological Science, 19(12):1242–1246, 2008.

[25] Choon Boon Ng, Yong Haur Tay, and Bok-Min Goi. Recognizing human gender in computer vision: A survey. In Patricia Anthony, Mit- suru Ishizuka, and Dickson Lukose, editors, 12th Pacific Rim International Conference on Artificial Intelligence, volume 7458 of Lecture Notes in Computer Science, pages 335–346, Berlin, Heidelberg, 2012. Springer.

[26] P. Jonathon Phillips, Harry Wechsler, Jeffery Huang, and Patrick J. Rauss. The FERET database and evaluation procedure for face- recognition algorithms. Image and Vision Computing, 16(5):295–306, 1998.

[27] Yingge Qu, Tien-Tsin Wong, and Pheng-Ann Heng. Manga colorization. In ACM SIG- GRAPH 2006, volume 25 of ACM Trans- actions on Graphics, pages 1214–1220, New York, 2006. ACM.

[28] Erik Reinhard, Michael Adhikhmin, Bruce Gooch, and Peter Shirley. Color transfer be- tween images. IEEE Computer Graphics and Applications, 21(4):34–41, 2001.

[29] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. ImageNet large scale visual recognition challenge. International Journal of Computer Vision, 115(3):211–252, 2015.

[30] Guillermo Sapiro. Inpainting the colors. In IEEE International Conference on Image Pro- cessing 2005, pages 698–701. IEEE, 2005.

[31] Caifeng Shan. Learning local binary patterns for gender classification on real-world face im- ages. Pattern Recognition Letters, 33(4):431–

437, 2012.

[32] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Van- houcke, and Andrew Rabinovich. Going deeper with convolutions. In 2015 IEEE Conference on Computer Vision and Pattern Recognition, pages 1–9. IEEE, 2015.

[33] Luis Torres, Jean-Yves Reutter, and Luis Lorente. The importance of the color infor- mation in face recognition. In 1999 Interna- tional Conference on Image Processing, vol- ume 3, pages 627–631. IEEE, 1999.

[34] Jos van de Wolfshaar, Mahir F. Karaaba, and Marco A. Wiering. Deep convolutional neural networks and support vector machines for gen- der recognition. In 2015 IEEE Symposium Se- ries on Computational Intelligence, pages 188–

195. IEEE, 2015.

[35] St´efan van der Walt, Johannes L. Sch¨onberger, Juan Nunez-Iglesias, Fran¸cois Boulogne, Joshua D. Warner, Neil Yager, Emmanuelle Gouillart, and Tony Yu. scikit-image: image processing in Python. PeerJ, 2, 2014.

[36] Tomihisa Welsh, Michael Ashikhmin, and Klaus Mueller. Transferring color to greyscale images. In ACM SIGGRAPH 2002, volume 21 of ACM Transactions on Graphics, pages 277–

280, New York, 2002. ACM.

[37] Liron Yatziv and Guillermo Sapiro. Fast im- age and video colorization using chrominance blending. IEEE Transactions on Image Pro- cessing, 15(5):1120–1129, 2006.

[38] Andrew W. Yip and Pawan Sinha. Contribu- tion of color to face recognition. Perception, 31(8):995–1003, 2002.

[39] Richard Zhang, Phillip Isola, and Alexei A.

Efros. Colorful image colorization. In Bas- tian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, European Conference on Computer Vision 2016, volume 9907 of Lec- ture Notes in Computer Science, pages 649–

666, Cham, 2016. Springer.

[40] Wenyi Zhao, Rama Chellappa, P. Jonathon Phillips, and Azriel Rosenfeld. Face recogni- tion: A literature survey. ACM Computing Surveys, 35(4):399–458, 2003.

Referenties

GERELATEERDE DOCUMENTEN

Naast de mogelijkheid die HI-gas biedt voor het goed in kaart brengen van sterrenstelsels in de meest verduisterde gebieden van de ZoA, biedt emissie van het HI-gas de mogelijkheid

If the universalization of a given maxim necessarily and immediately makes acting on this maxim impossible, then willing that the maxim be my indi- vidual principle and

Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers) Please check the document version of this publication:.. • A submitted manuscript is

Although equal employment opportunity policies ensured the involvement of more women in the labour force, they are still struggling with issues such as balancing

We assume that because of the long rental time of the bikes, an average of 8 hours, plus the fact that users have to bring back the bike at the same place, that the predicted

Fourth and final, I propose my idea of civil inequality that includes the idea that social structure might be embedded in social capital, and the circular mechanism of public

this paper, a smart city is therefore conceptualized as a city that makes use of Big Data Technology and the development of the Internet of Things to solve policy problems and

Nu de mate van vertrouwen en de mate van doelconsensus binnen het Warmtenet Hengelo duidelijk zijn, zal er worden gekeken naar de verschillende sturingsmethoden