Pitfalls in studying “big data” from operational scenarios

Analyzing a larger dataset is sometimes assumed, in itself, to confer a greater degree of validity to the results of a study. In biometrics, analyzing an “operational” dataset is also sometimes assumed, in itself to confer a greater degree of validity. And so studying a large, operational biometric dataset may seem to guarantee valid results. However, a number of basic questions should be asked of any “found” big data, in order to avoid pitfalls of the data not being suitable for the desired analysis. We explore such issues using a large operational iris recognition dataset from the Canada Border Services Agency's NEXUS program, similar to the dataset analyzed in the NISTIREX VI report.

[1]  Patrick J. Flynn,et al.  Pupil dilation degrades iris biometric performance , 2009, Comput. Vis. Image Underst..

[2]  Kevin W. Bowyer,et al.  Exploratory analysis of an operational iris recognition dataset from a CBSA border-crossing application , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[3]  George W. Quinn,et al.  IREX VI : Temporal Stability of Iris Recognition Accuracy , 2013 .

[4]  Kevin W. Bowyer,et al.  Statistical evaluation of up-to-three-attempt iris recognition , 2015, 2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS).

[5]  Adam Czajka Influence of Iris Template Aging on Recognition Reliability , 2013, BIOSTEC.

[6]  Arun Ross,et al.  Guidelines for best practices in biometrics research , 2015, 2015 International Conference on Biometrics (ICB).

[7]  Kevin W. Bowyer,et al.  Template Aging Phenomenon in Iris Recognition , 2013, IEEE Access.

[8]  John Daugman,et al.  New Methods in Iris Recognition , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[9]  Arun Ross,et al.  A study on quality-adjusted impact of time lapse on iris recognition , 2012, Defense + Commercial Sensing.