CIT-GAN: Cyclic Image Translation Generative Adversarial Network With Application in Iris Presentation Attack Detection
Abstract
In this work, we propose a novel Cyclic Image Translation Generative Adversarial Network (CIT-GAN) for multi-domain style transfer. To facilitate this, we introduce a Styling Network that has the capability to learn style characteristics of each domain represented in the training dataset. The Styling Network helps the generator to drive the translation of images from a source domain to a reference domain and generate synthetic images with style characteristics of the reference domain. The learned style characteristics for each domain depend on both the style loss and domain classification loss. This induces variability in style characteristics within each domain. The proposed CIT-GAN is used in the context of iris presentation attack detection (PAD) to generate synthetic presentation attack (PA) samples for classes that are under-represented in the training set. Evaluation using current state-of-the-art iris PAD methods demonstrates the efficacy of using such synthetically generated PA samples for training PAD methods. Further, the quality of the synthetically generated samples is evaluated using Frechet Inception Distance (FID) score. Results show that the quality of synthetic images generated by the proposed method is superior to that of other competing methods, including StarGan.

1 Introduction
The unique texture of the iris has made iris-based recognition systems desirable for human recognition in a number of applications [21]. However, these systems are increasingly facing threats from presentation attacks (PAs) where an adversary attempts to obfuscate their own identity, impersonate someone’s identity, or create a virtual identity [28]. Some commonly known iris presentation attacks (as shown in Figure 1) are:


- •
-
•
Artificial eyes [10]: Prosthetic or doll eyes are typically hand-crafted by professionals to look as similar to the bonafide irides as possible. Attackers can utilize such artifacts to obfuscate their true identity.
-
•
Cosmetic contact lens [13, 15]: The term “cosmetic contact lens” refers to lenses that typically have texture over them and are tinted with some color. These patterns can obstruct the natural iris texture that is required to recognize a person. Therefore, they can be used to deceive recognition systems.
In addition, electronic displays [20], cadaver eyes [3] and holographic eye images [26] may be used to launch a presentation attack. Grasping the threat posed by these attacks, researchers have been working on devising methods for iris presentation attack detection (PAD) that aim to distinguish between bonafide and PAs. In [18, 27], researchers used textural descriptors like Local Binary Pattern (LBP) and multi-scale binarized statistical image features (BSIF) to detect print attacks. Kohli et al. [23] proposed a variation of LBP to obtain textual information from iris images that helps in detecting cosmetic contact lens. More recently, deep features from Convolutional Neural Networks (CNNs) have been used to detect multiple iris presentation attacks [6][20]. Yadav et. al. [35] utilized the Relativistic Discriminator from a RaSGAN as a one-class classifier for PA detection. These methods report high accuracy for iris PA detection, but their performance can be negatively impacted by the absence of a sufficient number of samples from different PA classes [11]. Therefore, we can conclude that current iris PAD methods need a copious amount of training data corresponding to different PA classes and scenarios. Unfortunately, in the real world, such a dataset is hard to acquire.


With recent advances in the field of deep learning, researchers have proposed different methods based on Convolutional Autoencoders [31, 33] and Generative Adversarial Networks (GANs) [17] for image-to-image style translation. Here, image-to-image translation refers to learning a mapping between different visual domain categories each of which has its own unique appearance and style. Gatys et al. [16] proposed a neural architecture that could separate image content from style and then combine the two in different combinations to generate new natural looking styles. Their paper mainly focused on learning styles from well known artworks to generate new high quality and natural looking artwork. Karras et al. [22] introduced StyleGAN that uses a non-linear mapping function to embed a style driven latent code to generate images with different styles. However, since the input to the generator in StyleGAN is a noise vector, non-trivial efforts are required to transform image from one domain to another. Some researchers overcame this issue by enforcing an overlay between generator’s input and output for diversity in generated images using either marginal matching [1] or diversity regularization [37]. Others approached style transfer with the guidance of some reference images [5, 7].
However, these methods are not scalable to more than two domains and often show instability in the presence of multiple domains [9]. Choi et al. [8, 9] proposed to solve this problem by using a unified GAN architecture called StarGAN for style transfer that can generate diverse images across multiple domains. StarGAN uses a single generator with a mapping and Styling Network to learn diverse mappings between all available domains. The Styling Network aims to learn style codes for all the domains, while the mapping network is used to produce random style codes from latent codes. On one hand, it enforces the generator to learn diverse mappings across domains, but it also introduces the risk of generating images that are far from the original domains. Therefore, when introducing the mapping network, it becomes important to regularize the generator with diversity sensitive loss [9] that helps ensure that the generated data are diverse in nature.

In this research, we propose a GAN architecture that uses a novel Styling Network to drive the translation of input image into multiple target domains. Here, the generator takes an image as input along with a domain label and then generates images with characteristics of the target domains. Apart from the domain label, the generative capability of the network is enhanced using a multi-task Styling Network that learns the style codes for multiple PA domains and helps the generator to synthesize images reflecting the style components of the target PA domains.111A PA domain refers to a specific PA category, e.g., printed image. The domain-specific style characteristics learned using Styling Network depend on both style loss and domain classification. This ensures variability in style characteristics within each domain. Since there are multiple domains, the discriminator has multiple output branches to decide if a given image is real or synthetic for each of the domains (see Figure 2).
The primary contributions of this paper are as follows:
-
•
We propose a unified GAN architecture, referred to as Cyclic Image Translation Generative Adversarial Network (CIT-GAN), with novel Styling Network to generate good quality synthetic images for multiple domains. The quality of generated samples is evaluated using Fréchet Inception Distance (FID) Score [19].
-
•
We demonstrate the usefulness of the synthetically generated data to train state-of-the-art iris PAD methods and improve their accuracy.
2 Background
Generative Adversarial Networks (GANs) have been used by researchers extensively for synthetic image generation [17], super-resolution [4], anomaly detection [38], etc. due to their effectiveness in learning intricate features and texture of a given data distribution. Details of these networks are presented below.
2.1 Standard Generative Adversarial Network (GAN)
A Generative Adversarial Network [17] consists of two components, a generator and a discriminator , that compete with each other. takes as input a noise vector and aims to generate good quality synthetic data that closely resembles the real data. On the other hand, aims to distinguish between synthetically generated data and real data. This is denoted by a min-max objective function ,
(1) |
Here, represents the real data distribution and represents the Gaussian noise distribution. outputs whether input belongs to the real distribution or not. The generator takes an input to generate a synthetic image.
2.2 Frechet Inception Distance (FID) Score
Due to rapid advances in the field of DeepFakes, researchers have been studying different methods to evaluate the quality of synthetically generated data. Salimans et al. [29] used a pre-trained inception-V3 to compare the marginal and conditional label distribution of synthetically generated data to compute the inception score. Higher the inception score, better the quality of the generated dataset. However, the inception score does not include the statistics of the real data distribution when computing the score. In [19], Heusel et al. exploited the statistics of real data and compared it with the statistics of the synthetically generated dataset to compute the Frechet Inception Distance (FID) score:
(2) |
Here, are the statistics of the synthetic () and real () distributions. Since FID computes distance between the two distributions, the lower the FID score, better is the quality of generated dataset.
3 Cyclic Image Translation Generative Adversarial Network (CIT-GAN)
Let be an input image and be an arbitrary domain from domain space . The proposed method aims to translate image to synthetic image with style characteristics of domain . This is achieved using a Styling Network that is trained to learn domain specific style codes, and then train to generate synthetic images with the given target style codes (see Figure 3).
3.1 Generative Adversarial Network
Unlike standard GAN, the generative adversarial network in the proposed method has been updated to include domain level information. These changes are reflected in each component of the proposed architecture (as shown in 4) :
-
•
Generator: For image-to-image translation between multiple domains, takes an input image and translates it to an image with the desired style code . The style code is facilitated by Styling Network and injected into G.
-
•
Discriminator: Discriminator in the proposed architecture has multiple branches, where each branch decides whether the input image is a real image in domain or a synthetic image.
With the new objective for the generative adversarial network, the adversarial loss can be updated as,
(3) |
Here, outputs a decision on image for domain branch . The Styling Network takes an image from target domain and outputs a style code . generates image with style characteristics of target domain .

3.2 Styling Network
Given an input image belonging to domain , the Styling Network encodes the image into a style code . Similar to , the Styling Network is a multi-task network that learns the style code for an input image and injects the style code into to generate images with the given style codes. This is achieved using [9],
(4) |
Here, is the style code of reference image belonging to target domain . This ensures that generates images with the specified style code. However, poor quality synthetic data in the initial training iterations can affect the quality of the domain specific style codes learned by . To avoid this, we introduce a domain classification loss at the shared layer of from soft-max layer (as shown in Figure 4) to ensure that the learnt style code aligns with the correct domain. Further, this helps the Styling Network to learn style vectors (or feature characteristics) of varying samples from same domain.
(5) |
Here, is the true domain of input .

3.3 Cycle Consistency
While translating images from the source domain to the domains depicted by the reference images, it is important to preserve some characteristics of the input images (such as geometry, pose and eye lashes in case of irides). This is achieved using the cycle consistency loss [8],
(6) |
Here, represents the style code of input image with domain , and is the style code of reference image in target domain . This ensures that image with style can be reconstructed using synthetic image .
Hence, the overall loss function for the proposed Cyclic Image Translation Generative Adversarial Network can be defined as:
(7) |
Here, and represent the hyperparameters for each loss term.
4 Dataset Used
In this research, we utilized five different iris PA datasets, viz., Casia-iris-fake [32], Berc-iris-fake [25], NDCLD15 [14], LivDet2017 [36] and MSU-IrisPA-01 [34] for training and testing different iris presentation attack detection (PAD) algorithms. These iris datasets contain bonafide images and images from different PA classes such as cosmetic contact lenses, printed eyes, artificial eyes and kindle-display attack (as shown in Figure 1). The images in these datasets are pre-processed and cropped to a size of 256x256 around the iris using the coordinates from a software called VeriEye.222www.neurotechnology.com/verieye.html The images that were not properly processed by VeriEye were discarded from the datasets as this paper focuses primarily on image synthesis. This give us a total of 24,409 bonafide irides, 6,824 cosmetic contact lenses, 680 artificial eyes and 13,293 printed eyes.
5 Image Quality
The proposed architecture is trained using 6,450 bonafide images, 2,104 cosmetic contact lenses, 4,482 printed eyes and 276 artificial eyes randomly selected from the aforementioned datasets. The trained network is then utilized to generate synthetic PA samples. To achieve this, 6,000 bonafide images were utilized as source images. The source images are then translated to different PA classes using 2,000 printed eyes, 2,000 cosmetic contact lens and 276 artificial eyes as reference images. Using this approach, we generated 8,000 samples for each PA class. The generated samples from CIT-GAN obtained an average FID score of 32.79.
For comparison purposes, we used the same train and evaluation setup to generate synthetic samples using Star-GAN [8], Star-GAN v2 [9] and Style-GAN [22]. As mentioned before, Style-GAN and Star-GAN are not well-equipped to handle multi-domain image translation. Therefore, they obtained a high average FID score of 86.69 and 44.76, respectively. On the other hand, Star-GAN v2 is equipped to handle multi-domains using a styling and mapping network. A trained Star-GAN v2 utilizes the mapping network to generate diverse style codes to diversify images. The synthetic iris PAs generated using this method were diverse in nature, but failed to capture the true characteristics of PAs like artificial eyes. Hence, the average FID score of the generated image using Star-GAN v2 was 38.81 - much lower than that of Style-GAN and Star-GAN, but still a bit higher than CIT-GAN. This can also be seen in the FID score distribution in Figure 10 that compares the synthetically generated data using Star-GAN v2 with that of CIT-GAN.


6 Experimental Setups
In this section, we describe different experimental setups that are used to evaluate the quality and usefulness of synthetic PA samples generated using CIT-GAN. We evaluated the performance of different iris PAD methods viz., VGG-16 [16], BSIF [14], DESIST [23], D-NetPAD [30] and AlexNet [24] under these different experimental setups for analysis purposes. Note that D-NetPAD is one of the best performing PAD algorithms in the iris liveness detection competition (LivDet-20 edition) [12].
6.1 Experiment-1
This is the baseline experiment that demonstrates the performance of the current iris PAD methods on the previously mentioned datasets with imbalanced samples across different PA classes. The PAD methods are trained using 14,970 bonafide samples and 10,306 PA samples consisting of 276 artificial eyes, 4,014 cosmetic contact lenses and 6,016 printed eyes. The test set consists of 9,439 bonafide samples and 9,896 PA samples corresponding to 404 artificial eyes, 2,720 cosmetic contact lenses and 6,772 printed eyes.
6.2 Experiment-2
In this experiment, our aim is to evaluate the equivalence between real PAs and synthetically generated PAs. Therefore, the iris PAD methods are trained using 14,970 bonafide samples and PAs consisting of both real PA images and synthetic PA images. The real PA dataset has 138 artificial eyes, 2,007 cosmetic contact lenses and 3,008 printed eyes. The synthetic PA dataset is generated using the remainder of the real PA dataset as reference images (i.e., 138 artificial eyes, 2,007 cosmetic contact lenses and 3,008 printed eyes) in order to capture their style characteristics in the generated dataset. As before, the test set consists of 9,439 bonafide and 9,896 PA samples corresponding to 404 artificial eyes, 2,720 cosmetic contact lenses and 6,772 printed eyes.
6.3 Experiment-3
This experiment aims to evaluate the efficacy of the proposed method, CIT-GAN, in generating synthetic PA samples that represent the real PA distribution across various PA domains. Here, the iris PAD methods are trained using 14,970 bonafide samples and synthetically generated 276 artificial eyes, 4,014 cosmetic contact lenses and 6,016 printed eyes. The test set consists of 9,439 bonafide and 9,896 PA samples corresponding to 404 artificial eyes, 2,720 cosmetic contact lenses and 6,772 printed eyes.
6.4 Experiment-4
As mentioned in Experiment-1, current iris PAD methods are trained and tested on imbalanced samples from PA classes thereby affecting their accuracy. To overcome this, we train the iris PAD methods using 14,970 bonafide samples and a balanced set of 15,000 PA samples corresponding to 276 artificial eyes, 4,014 cosmetic contact lenses and 5000 printed eyes that are real; and 4,724 artificial eyes and 986 cosmetic contact lenses that are synthetic. This balances the number of samples across PA classes. The testing was done on 9,439 bonafide samples and 9,896 PA samples consisting of 404 artificial eyes, 2,720 cosmetic contact lenses and 6,772 printed eyes.
7 Results and Analysis
In this section, we discuss the results obtained for the four different experiments described in the previous section. Experiment-1 is the baseline experiment that evaluates the performance of various iris PAD methods. The training set for this experiment contains 14,970 bonafide and 10,306 PA samples consisting of 276 artificial eyes, 4,014 cosmetic contact lenses and 6,016 printed eyes. Due to imbalance in the number of samples across various PA domains, the performance of the PAD methods is affected. This becomes apparent when comparing the results of Experiment-1 with that of Experiment-4 where PAD methods are trained using 9,439 bonafide samples and a balanced number of PA samples (i.e., 5,000 samples from each PA domain) containing both real and synthetic PAs. As seen from the results in Table 1 and Table 4, performance for each PAD method improves in Experiment-4. For example, in the case of D-NetPAD, the TDR at a 1% FDR improved from 92.54% in Experiment-1 to 97.89% in Experiment-4 (as shown in Figure 8). A huge increase in performance was also noticed for BSIF+SVM where TDR improved from 28.11% in Experiment-1 to 51.11% in Experiment-4, at a FDR of 1%.
In addition, the equivalence of synthetically generated PA samples and real PA samples was established using Experiment-2 and Experiment-3. In Experiment-2, some of the real PA samples in the training set were replaced with synthetically generated PAs. Comparing the performance in Table 1 and Table 2, a very slight difference in PAD performance is observed (see Figure 7). Similarly, in Experiment-3, where all the real PAs are replaced with synthetically generated PA samples, only a slight decrease in performance was seen for the PAD methods (as shown in Figure 9) signifying underlying similarities between real and synthetically generated data.
8 Conclusion and Future Work
In this research work, we designed an image-to-image translation method (CIT-GAN) that can synthetically generate images across multiple iris presentation attack domains, i.e., multiple types of iris PAs. The results obtained in the previous sections show the equivalence of synthetically generated PA samples and real PA samples. Furthermore, the results in Table 1 and Table 4 demonstrate that the performance of the iris PAD methods can be improved by adding synthetically generated data to different PA classes for balanced training.
We would like to extend this work to generate PA samples with different variations and multi-domain style characteristics such as ”printed” cosmetic contact lenses, etc. This will allow current iris PAD methods to generalize over different styles of PAs for enhanced PA detection.
9 Acknowledgment
This research is based upon work supported in part by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA), via IARPA R&D Contract No. 2017 - 17020200004. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, either expressed or implied, of ODNI, IARPA, or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation therein.
References
- [1] Amjad Almahairi, Sai Rajeswar, Alessandro Sordoni, Philip Bachman, and Aaron Courville. Augmented cycleGAN: Learning many-to-many mappings from unpaired data. arXiv preprint arXiv:1802.10151, 2018.
- [2] Rajesh Bodade and Sanjay Talbar. Fake iris detection: A holistic approach. International Journal of Computer Applications, 19(2):1–7, 2011.
- [3] Aidan Boyd, Shivangi Yadav, Thomas Swearingen, Andrey Kuehlkamp, Mateusz Trokielewicz, Eric Benjamin, Piotr Maciejewicz, Dennis Chute, Arun Ross, Patrick Flynn, et al. Post-Mortem Iris Recognition—A Survey and Assessment of the State of the Art. IEEE Access, 8:136570–136593, 2020.
- [4] Jiancheng Cai, Han Hu, Shiguang Shan, and Xilin Chen. Fcsr-GAN: End-to-end learning for joint face completion and super-resolution. In 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pages 1–8. IEEE, 2019.
- [5] Huiwen Chang, Jingwan Lu, Fisher Yu, and Adam Finkelstein. Paired cycleGAN: Asymmetric style transfer for applying and removing makeup. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 40–48, 2018.
- [6] Cunjian Chen and Arun Ross. A multi-task convolutional neural network for joint iris detection and presentation attack detection. In IEEE Winter Applications of Computer Vision Workshops (WACVW), pages 44–51, 2018.
- [7] Wonwoong Cho, Sungha Choi, David Keetae Park, Inkyu Shin, and Jaegul Choo. Image-to-image translation via group-wise deep whitening-and-coloring transformation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 10639–10647, 2019.
- [8] Yunjey Choi, Minje Choi, Munyoung Kim, Jung-Woo Ha, Sunghun Kim, and Jaegul Choo. StarGAN: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8789–8797, 2018.
- [9] Yunjey Choi, Youngjung Uh, Jaejun Yoo, and Jung-Woo Ha. StarGAN v2: Diverse image synthesis for multiple domains. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8188–8197, 2020.
- [10] Adam Czajka. Iris liveness detection by modeling dynamic pupil features. In Handbook of Iris Recognition, pages 439–467. Springer, 2016.
- [11] Adam Czajka and Kevin W Bowyer. Presentation attack detection for iris recognition: An assessment of the state-of-the-art. ACM Computing Surveys (CSUR), 51(4):1–35, 2018.
- [12] Priyanka Das, Joseph McGrath, Zhaoyuan Fang, Aidan Boyd, Ganghee Jang, Amir Mohammadi, Sandip Purnapatra, David Yambay, Sébastien Marcel, Mateusz Trokielewicz, et al. Iris liveness detection competition (LivDet-Iris)–the 2020 edition. In IEEE International Joint Conference on Biometrics (IJCB), 2020.
- [13] John Daugman. Demodulation by complex-valued wavelets for stochastic pattern recognition. International Journal of Wavelets, Multiresolution and Information Processing, 1(01):1–17, 2003.
- [14] James S Doyle and Kevin W Bowyer. Robust detection of textured contact lenses in iris recognition using BSIF. IEEE Access, 3:1672–1683, 2015.
- [15] James S Doyle, Patrick J Flynn, and Kevin W Bowyer. Automated classification of contact lens type in iris images. In International Conference on Biometrics (ICB), pages 1–6. IEEE, 2013.
- [16] Leon A Gatys, Alexander S Ecker, and Matthias Bethge. A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576, 2015.
- [17] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative Adversarial Nets. In Advances in Neural Information Processing Systems (NIPS), pages 2672–2680, 2014.
- [18] Priyanshu Gupta, Shipra Behera, Mayank Vatsa, and Richa Singh. On iris spoofing using print attack. In 22nd International Conference on Pattern Recognition, pages 1681–1686. IEEE, 2014.
- [19] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in neural information processing systems, pages 6626–6637, 2017.
- [20] Steven Hoffman, Renu Sharma, and Arun Ross. Iris + Ocular: Generalized Iris Presentation Attack Detection Using Multiple Convolutional Neural Networks. In IAPR International Conference on Biometrics (ICB), 2019.
- [21] Anil K Jain, Karthik Nandakumar, and Arun Ross. 50 years of biometric research: Accomplishments, challenges, and opportunities. Pattern Recognition Letters, 79:80–105, 2016.
- [22] Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4401–4410, 2019.
- [23] Naman Kohli, Daksha Yadav, Mayank Vatsa, Richa Singh, and Afzel Noore. Detecting medley of iris spoofing attacks using DESIST. In IEEE International Conference on Biometrics Theory, Applications and Systems (BTAS), pages 1–6, 2016.
- [24] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097–1105, 2012.
- [25] Sung Joo Lee, Kang Ryoung Park, Youn Joo Lee, Kwanghyuk Bae, and Jai Hie Kim. Multifeature-based fake iris detection method. Optical Engineering, 46(12):1–10, 2007.
- [26] Andrzej Pacut and Adam Czajka. Aliveness detection for iris biometrics. In Proceedings of 40th Annual International Carnahan Conference on Security Technology, pages 122–129. IEEE, 2006.
- [27] R. Raghavendra and C. Busch. Presentation attack detection algorithm for face and iris biometrics. In European Signal Processing Conference (EUSIPCO), pages 1387–1391, 2014.
- [28] Arun Ross, Sudipta Banerjee, Cunjian Chen, Anurag Chowdhury, Vahid Mirjalili, Renu Sharma, Thomas Swearingen, and Shivangi Yadav. Some Research Problems in Biometrics: The Future Beckons. In IAPR International Conference on Biometrics (ICB), 2019.
- [29] Tim Salimans, Han Zhang, Alec Radford, and Dimitris Metaxas. Improving GANs using optimal transport. arXiv preprint arXiv:1803.05573, 2018.
- [30] Renu Sharma and Arun Ross. D-NetPAD: An Explainable and Interpretable Iris Presentation Attack Detector. In IEEE International Joint Conference on Biometrics (IJCB), 2020.
- [31] Ashish Shrivastava, Tomas Pfister, Oncel Tuzel, Joshua Susskind, Wenda Wang, and Russell Webb. Learning from simulated and unsupervised images through adversarial training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2107–2116, 2017.
- [32] Z. Sun, H. Zhang, T. Tan, and J. Wang. Iris image classification based on hierarchical visual codebook. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(6):1120–1133, 2014.
- [33] Aaron Van den Oord, Nal Kalchbrenner, Lasse Espeholt, Oriol Vinyals, Alex Graves, et al. Conditional image generation with CNN decoders. In Advances in Neural Information Processing Systems (NIPS), pages 4790–4798, 2016.
- [34] Shivangi Yadav, Cunjian Chen, and Arun Ross. Synthesizing Iris Images using RaSGAN with Application in Presentation Attack Detection. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2019.
- [35] Shivangi Yadav, Cunjian Chen, and Arun Ross. Relativistic Discriminator: A One-Class Classifier for Generalized Iris Presentation Attack Detection. In IEEE Winter Conference on Applications of Computer Vision, pages 2635–2644, 2020.
- [36] David Yambay, Benedict Becker, Naman Kohli, Daksha Yadav, Adam Czajka, Kevin W Bowyer, Stephanie Schuckers, Richa Singh, Mayank Vatsa, Afzel Noore, et al. LivDet Iris 2017 - Iris Liveness Detection Competition. In IEEE International Joint Conference on Biometrics (IJCB), pages 733–741, 2017.
- [37] Dingdong Yang, Seunghoon Hong, Yunseok Jang, Tianchen Zhao, and Honglak Lee. Diversity-sensitive conditional generative adversarial networks. arXiv preprint arXiv:1901.09024, 2019.
- [38] Houssam Zenati, Chuan Sheng Foo, Bruno Lecouat, Gaurav Manek, and Vijay Ramaseshan Chandrasekhar. Efficient GAN-based anomaly detection. arXiv preprint arXiv:1802.06222, 2018.