• No results found

University of Groningen Computer vision techniques for calibration, localization and recognition Lopez Antequera, Manuel

N/A
N/A
Protected

Academic year: 2021

Share "University of Groningen Computer vision techniques for calibration, localization and recognition Lopez Antequera, Manuel"

Copied!
13
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Computer vision techniques for calibration, localization and recognition

Lopez Antequera, Manuel

DOI:

10.33612/diss.112968625

IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from

it. Please check the document version below.

Document Version

Publisher's PDF, also known as Version of record

Publication date:

2020

Link to publication in University of Groningen/UMCG research database

Citation for published version (APA):

Lopez Antequera, M. (2020). Computer vision techniques for calibration, localization and recognition.

University of Groningen. https://doi.org/10.33612/diss.112968625

Copyright

Other than for strictly personal use, it is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license (like Creative Commons).

Take-down policy

If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim.

Downloaded from the University of Groningen/UMCG research database (Pure): http://www.rug.nl/research/portal. For technical reasons the number of authors shown on this cover page is limited to 10 maximum.

(2)

Akhtar, N. and Mian, A.: 2018, Threat of adversarial attacks on deep learning in computer vision: A survey, IEEE Access .

Alitto, H. J. and Dan, Y.: 2010, Function of inhibition in visual cortical processing, Current Opinion in Neurobiology (3). Sensory systems.

Arandjelovic, R., Gronat, P., Torii, A., Pajdla, T. and Sivic, J.: 2016, NetVLAD: CNN archi-tecture for weakly supervised place recognition, Conference on Computer Vision and Pattern Recognition (CVPR).

Arroyo, R., Alcantarilla, P. F., Bergasa, L. M. and Romera, E.: 2015, Towards Life-Long Visual Localization using an Efficient Matching of Binary Sequences from Images, International Conference on Robotics and Automation (ICRA).

Azzopardi, G. and Petkov, N.: 2012, A CORF computational model of a simple cell that relies on LGN input outperforms the Gabor function model, Biological Cybernetics .

Azzopardi, G. and Petkov, N.: 2013, Trainable COSFIRE filters for keypoint detection and pattern recognition, IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) . Azzopardi, G. and Petkov, N.: 2014, Ventral-stream-like shape representation: from pixel

intensity values to trainable object-selective cosfire models, Frontiers in Computational Neu-roscience .

Azzopardi, G., Rodr´ıguez-S´anchez, A., Piater, J. and Petkov, N.: 2014, A push-pull corf model of a simple cell with antiphase inhibition improves snr and contour detection, PLoS ONE . Azzopardi, G., Strisciuglio, N., Vento, M. and Petkov, N.: 2015, Trainable COSFIRE filters for

vessel delineation with application to retinal images, Medical Image Analysis .

Badrinarayanan, V., Kendall, A. and Cipolla, R.: 2015, Segnet: A deep convolutional encoder-decoder architecture for image segmentation., CoRR .

(3)

Balntas, V., Johns, E., Tang, L. and Mikolajczyk, K.: 2016, PN-Net: Conjoined Triple Deep Network for Learning Local Image Descriptors, CoRR .

Bay, H., Tuytelaars, T. and Van Gool, L.: 2006, SURF: Speeded Up Robust Features, European Conference on Computer Vision (ECCV).

Blanco, J.-L., Moreno, F.-A. and Gonz´alez-Jim´enez, J.: 2014, The M´alaga Urban Dataset: High-rate Stereo and Lidars in a realistic urban scenario, The International Journal of Robotics Re-search (IJRR) .

Bolz, J. and Gilbert, C. D.: 1986, Generation of end-inhibition in the visual cortex via inter-laminar connections., Nature .

Brown, D. C.: 1971, Close-Range Camera Calibration, Photogrammetric Engineering And Remote Sensing .

Calonder, M., Lepetit, V., Strecha, C. and Fua, P.: 2010, BRIEF: Binary Robust Independent Elementary Features, European Conference on Computer Vision (ECCV).

Caprile, B. and Torre, V.: 1990, Using vanishing points for camera calibration, International Journal of Computer Vision (IJCV) .

Carlini, N. and Wagner, D. A.: 2016, Towards evaluating the robustness of neural networks, CoRR .

Caruana, R.: 1997, Multitask learning, Machine Learning (1).

Chen, Z., Badrinarayanan, V., Lee, C.-Y. and Rabinovich, A.: 2017, GradNorm: Gradient normalization for adaptive loss balancing in deep multitask networks, CoRR .

Chen, Z., Jacobson, A., Sunderhauf, N., Upcroft, B., Liu, L., Shen, C., Reid, I. and Milford, M.: 2017, Deep Learning Features at Scale for Visual Place Recognition, CoRR .

Chen, Z., Lam, O., Jacobson, A. and Milford, M.: 2014, Convolutional Neural Network-based Place Recognition, CoRR .

Cohen, T. S. and Welling, M.: 2016, Steerable cnns, CoRR .

Cummins, M. and Newman, P.: 2008, FAB-MAP: Probabilistic Localization and Mapping in the Space of Appearance, The International Journal of Robotics Research (IJRR) .

Deretey, E., Ahmed, M. T., Marshall, J. A. and Greenspan, M.: 2015, Visual indoor positioning with a single camera using PnP, International Conference on Indoor Positioning and Indoor Navigation (IPIN).

Deutscher, J., Isard, M. and MacCormick, J.: 2002, Automatic Camera Calibration from a Sin-gle Manhattan Image, in A. Heyden, G. Sparr, M. Nielsen and P. Johansen (eds), European Conference on Computer Vision (ECCV), Springer Berlin Heidelberg, Berlin, Heidelberg.

(4)

Devernay, F. and Faugeras, O.: 2001, Straight lines have to be straight, Machine Vision and Applications .

Dodge, S. and Karam, L.: 2017, A study and comparison of human and deep learning recog-nition performance under visual distortions, 2017 26th International Conference on Computer Communication and Networks (ICCCN).

Ferris, B., Haehnel, D. and Fox, D.: 2006, Gaussian processes for signal strength-based loca-tion estimaloca-tion, Robotics: Science and Systems (RSS).

Fitzgibbon, A. W.: 2001, Simultaneous linear estimation of multiple view geometry and lens distortion, Conference on Computer Vision and Pattern Recognition (CVPR), IEEE.

Fox, D.: 2001, KLD-sampling: Adaptive particle filters, Neural Information Processing Systems (NIPS).

Freeman, T. C., Durand, S., Kiper, D. C. and Carandini, M.: 2002, Suppression without inhibi-tion in visual cortex, Neuron .

Fukushima, K.: 1980, Neocognitron: A self-organizing neural network model for a mecha-nism of pattern recognition unaffected by shift in position, Biological Cybernetics .

Galvez-Lopez, D. and Tardos, J. D.: 2012, Bags of binary words for fast place recognition in image sequences, IEEE Transactions on Robotics (TRO) .

Gecer, B., Azzopardi, G. and Petkov, N.: 2017, Color-blob-based COSFIRE filters for object recognition, Image and Vision Computing .

Geiger, A., Lenz, P. and Urtasun, R.: 2012, Are we ready for autonomous driving? The KITTI vision benchmark suite, Conference on Computer Vision and Pattern Recognition (CVPR), IEEE. Geirhos, R., Temme, C. R. M., Rauber, J., Sch ¨utt, H. H., Bethge, M. and Wichmann, F. A.: 2018, Generalisation in humans and deep neural networks, in S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi and R. Garnett (eds), Advances in Neural In-formation Processing Systems 31.

Gomez-Ojeda, R. and Gonz´alez-Jim´enez, J.: 2016, Robust Stereo Visual Odometry through a Probabilistic Combination of Points and Line Segments, International Conference on Robotics and Automation (ICRA).

Gomez-Ojeda, R., Lopez-Antequera, M., Petkov, N. and Jim´enez, J. G.: 2015, Training a con-volutional neural network for appearance-invariant place recognition, CoRR .

Gonzalez-Aguilera, D., Gomez-Lahoz, J., Rodriguez-Gonzalvez, P., Gonz´alez-Aguilera, D., G ´omez-Lahoz, J. and Rodr´ıguez-Gonz´alvez, P.: 2011, An automatic approach for radial lens distortion correction from a single image, IEEE Sensors Journal .

Goodfellow, I., Shlens, J. and Szegedy, C.: 2015, Explaining and harnessing adversarial exam-ples, International Conference on Learning Representations.

(5)

Hartley, R. and Zisserman, A.: 2003, Multiple view geometry in computer vision, Cambridge university press.

He, K., Zhang, X., Ren, S. and Sun, J.: 2015, Deep residual learning for image recognition, CoRR .

Hendrycks, D. and Dietterich, T.: 2019, Benchmarking neural network robustness to com-mon corruptions and perturbations, Proceedings of the International Conference on Learning Representations .

Hold-Geoffroy, Y., Sunkavalli, K., Eisenmann, J., Fisher, M., Gambaretto, E., Hadap, S. and Lalonde, J.-F.: 2018, A Perceptual Measure for Deep Single Image Camera Calibration, Conference on Computer Vision and Pattern Recognition (CVPR).

Honegger, D., Sattler, T. and Pollefeys, M.: 2017, Embedded real-time multi-baseline stereo, IEEE ICRA.

hua Liu, B., tang Li, Y., pei Ma, W., jie Pan, C., Zhang, L. I. and Tao, H. W.: 2011, Broad inhibition sharpens orientation selectivity by expanding input dynamic range in mouse simple cells, Neuron .

Huang, G., Liu, Z., Van Der Maaten, L. and Weinberger, K. Q.: 2017, Densely connected convolutional networks., Conference on Computer Vision and Pattern Recognition (CVPR). Hubel, D. and Wiesel, T.: 1962, Receptive fields, binocular interaction and functional

archi-tecture in the cat’s visual cortex, Journal of Physiology-London .

Hui, T.-W., Tang, X. and Loy, C. C.: 2018, Liteflownet: A lightweight convolutional neural network for optical flow estimation, Conference on Computer Vision and Pattern Recognition (CVPR).

Huitl, R., Schroth, G., Hilsenbeck, S., Schweiger, F. and Steinbach, E.: 2012, TUMindoor: An extensive image and point cloud dataset for visual indoor localization and mapping, IEEE International Conference on Image Processing (ICIP), IEEE.

Ishii, C., Sudo, Y. and Hashimoto, H.: 2003, An image conversion algorithm from fish eye image to perspective image for human eyes, International Conference on Advanced Intelligent Mechatronics, IEEE.

Jayaraman, D. and Grauman, K.: 2015, Learning image representations tied to egomotion, IEEE International Conference on Computer Vision (ICCV).

Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., Guadarrama, S. and Darrell, T.: 2014, Caffe: Convolutional architecture for fast feature embedding, Proceedings of the ACM International Conference on Multimedia, ACM.

Kendall, A., Gal, Y. and Cipolla, R.: 2018, Multi-task learning using uncertainty to weigh losses for scene geometry and semantics, Conference on Computer Vision and Pattern Recog-nition (CVPR).

(6)

Ko, J. and Fox, D.: 2008, GP-BayesFilters: Bayesian filtering using Gaussian process prediction and observation models, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

Kremkow, J., Perrinet, L. U., Monier, C., Alonso, J.-M., Aertsen, A., Fr´egnac, Y. and Masson, G. S.: 2016, Push-pull receptive field organization and synaptic depression: Mechanisms for reliably encoding naturalistic stimuli in v1, Frontiers in Neural Circuits .

Krizhevsky, A., Sutskever, I. and Hinton, G. E.: 2012, Imagenet classification with deep con-volutional neural networks, in F. Pereira, C. J. C. Burges, L. Bottou and K. Q. Weinberger (eds), Neural Information Processing Systems (NIPS), Curran Associates, Inc.

Kumar, B. G. V., Carneiro, G. and Reid, I.: 2015, Learning Local Image Descriptors with Deep Siamese and Triplet Convolutional Networks by Minimising Global Loss Functions, CoRR .

Kurakin, A., Goodfellow, I. J. and Bengio, S.: 2016, Adversarial examples in the physical world., CoRR .

Larlus, D. and Jurie, F.: 2009, Latent mixture vocabularies for object categorization and seg-mentation, Image and Vision Computing .

Lazebnik, S., Schmid, C. and Ponce, J.: 2004, Semi-local Affine Parts for Object Recognition, in A. Hoppe, S. Barman and T. Ellis (eds), British Machine Vision Conference (BMVC), The British Machine Vision Association (BMVA), Kingston, United Kingdom.

LeCun, Y., Bengio, Y. and Hinton, G.: 2015, Deep learning, Nature .

Lecun, Y., Bottou, L., Bengio, Y. and Haffner, P.: 1998, Gradient-based learning applied to document recognition, Proceedings of the IEEE.

LeCun, Y., Haffner, P., Bottou, L. and Bengio, Y.: 1999, Object recognition with gradient-based learning, in D. Forsyth (ed.), Shape, Contour and Grouping in Computer Vision, Springer. Lee, C.-Y., Xie, S., Gallagher, P., Zhang, Z. and Tu, Z.: 2015, Deeply-Supervised Nets,

ings of the Eighteenth International Conference on Artificial Intelligence and Statistics, Proceed-ings of Machine Learning Research, PMLR.

Leyva-Vallina, M., Strisciuglio, N., L ´opez Antequera, M., Tylecek, R., Blaich, M. and Petkov, N.: 2019, Tb-places: A data set for visual place recognition in garden environments, IEEE Access .

Li, Y.-t., Ma, W.-p., Li, L.-y., Ibrahim, L. A., Wang, S.-z. and Tao, H. W.: 2012, Broadening of inhibitory tuning underlies contrast-dependent sharpening of orientation selectivity in mouse visual cortex, Journal of Neuroscience .

Lopez-Antequera, M., Gomez-Ojeda, R., Petkov, N. and Gonzalez-Jimenez, J.: 2017, Appearance-invariant place recognition by discriminatively training a convolutional neu-ral network, Pattern Recognition Letters .

(7)

Lopez-Antequera, M., Petkov, N. and Gonzalez-Jimenez, J.: 2016, Image-based localization using Gaussian processes, International Conference on Indoor Positioning and Indoor Naviga-tion (IPIN).

Lowe, D. G.: 2004, Distinctive Image Features from Scale-Invariant Keypoints, International Journal of Computer Vision (IJCV) .

Lowry, S., S ¨underhauf, N., Newman, P., Leonard, J. J., Cox, D., Corke, P. and Milford, M. J.: 2016, Visual place recognition: A survey, IEEE Transactions on Robotics (TRO) .

Lu, J., Sibai, H., Fabry, E. and Forsyth, D. A.: 2017, Standard detectors aren’t (currently) fooled by physical adversarial stop signs, CoRR .

Maddern, W., Milford, M. and Wyeth, G.: 2011, Continuous appearance-based trajectory SLAM, International Conference on Robotics and Automation (ICRA).

Maddern, W., Milford, M. and Wyeth, G.: 2012a, CAT-SLAM: probabilistic localisation and mapping using a continuous appearance-based trajectory, The International Journal of Robotics Research (IJRR) .

Maddern, W., Milford, M. and Wyeth, G.: 2012b, Towards persistent indoor appearance-based localization, mapping and navigation using CAT-Graph, IEEE/RSJ International Con-ference on Intelligent Robots and Systems (IROS).

Madry, A., Makelov, A., Schmidt, L., Tsipras, D. and Vladu, A.: 2018, Towards deep learning models resistant to adversarial attacks, CoRR .

Maneewongvatana, S. and Mount, D. M.: 1999, It’s okay to be skinny, if your friends are fat, Center for Geometric Computing 4th Annual Workshop on Computational Geometry.

Marˆcelja, S.: 1980, Mathematical description of the responses of simple cortical cells, Journal of the Optical Society of America .

Metzen, J. H., Genewein, T., Fischer, V. and Bischoff, B.: 2017, On detecting adversarial per-turbations, Proceedings of 5th International Conference on Learning Representations (ICLR). Milford, M.: 2013, Vision-based place recognition: how low can you go?, The International

Journal of Robotics Research (IJRR) .

Milford, M. J. and Wyeth, G. F.: 2012, SeqSLAM: Visual route-based navigation for sunny summer days and stormy winter nights, International Conference on Robotics and Automation (ICRA).

Moosavi-Dezfooli, S., Fawzi, A., Fawzi, O. and Frossard, P.: 2017, Universal adversarial per-turbations, CVPR, IEEE Computer Society.

Moosavi-Dezfooli, S., Fawzi, A. and Frossard, P.: 2016, Deepfool: A simple and accurate method to fool deep neural networks, CVPR, IEEE Computer Society.

(8)

Moreno, F. A., Blanco, J.-L. and Gonzalez, J.: 2009, Stereo vision specific models for particle filter-based SLAM, Robotics and Autonomous Systems (RAS) .

Moreno, F. A., Blanco, J. L. and Gonzalez-Jimenez, J.: 2016, A constant-time SLAM back-end in the continuum between global mapping and submapping: application to visual stereo SLAM, The International Journal of Robotics Research (IJRR) .

Mur-Artal, R., Montiel, J. M. M. and Tardos, J. D.: 2015, ORB-SLAM: A Versatile and Accurate Monocular SLAM System, IEEE Transactions on Robotics (TRO) .

Mur-Artal, R. and Tard ´os, J. D.: 2014, Fast Relocalisation and Loop Closing in Keyframe-Based SLAM, International Conference on Robotics and Automation (ICRA).

Neubert, P. and Protzel, P.: 2015, Local region detector+ cnn based landmarks for practical place recognition in changing environments, European Conference on Mobile Robots (ECMR), IEEE.

Neubert, P. and Protzel, P.: 2016, Beyond holistic descriptors, keypoints, and fixed patches: Multiscale superpixel grids for place recognition in changing environments, IEEE Robotics and Automation Letters (RAL) .

Nist´er, D. and Stew´enius, H.: 2006, Scalable recognition with a vocabulary tree, Conference on Computer Vision and Pattern Recognition (CVPR).

Papernot, N., McDaniel, P. D., Wu, X., Jha, S. and Swami, A.: 2015, Distillation as a defense to adversarial perturbations against deep neural networks, CoRR .

Pasupathy, A. and Connor, C.: 2002, Population coding of shape in area v4, Nature Neuro-science .

Pasupathy, A. and Connor, C. E.: 1999, Responses to contour features in macaque area v4, Journal of Neurophysiology .

Pepperell, E., Corke, P. and Milford, M.: 2014, All-environment visual place recognition with SMART, International Conference on Robotics and Automation (ICRA).

Pepperell, E., Corke, P. and Milford, M.: 2016, Routed roads: Probabilistic vision-based place recognition for changing conditions, split streets and varied viewpoints, The International Journal of Robotics Research (IJRR) .

Ramalingam, S., Lodha, S. K. and Sturm, P.: 2006, A generic structure-from-motion frame-work, Computer Vision and Image Understanding .

Ranzato, M., Poultney, C., Chopra, S. and Cun, Y. L.: 2007, Efficient Learning of Sparse Rep-resentations with an Energy-Based Model, in B. Sch ¨olkopf, J. C. Platt and T. Hoffman (eds), Neural Information Processing Systems (NIPS), MIT Press.

Rasmussen, C. E. and Williams, C. K. I.: 2005, Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning series), The MIT Press.

(9)

Razavian, A. S., Azizpour, H., Sullivan, J. and Carlsson, S.: 2014, CNN Features off-the-shelf: an Astounding Baseline for Recognition, Computer Vision and Pattern Recognition Workshops (CVPRW).

Razavian, A. S., Sullivan, J., Maki, A. and Carlsson, S.: 2014, A Baseline for Visual Instance Retrieval with Deep Convolutional Networks, CoRR .

Rong, J., Huang, S., Shang, Z. and Ying, X.: 2017, Radial Lens Distortion Correction Using Convolutional Neural Networks Trained with Synthesized Images, in S.-H. Lai, V. Lepetit, K. Nishino and Y. Sato (eds), Asian Conference on Computer Vision (ACCV), Springer Inter-national Publishing, Cham.

Rublee, E., Rabaud, V., Konolige, K. and Bradski, G.: 2011, ORB: An efficient alternative to SIFT or SURF, IEEE International Conference on Computer Vision (ICCV), Ieee.

Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., Berg, A. C. and Fei-Fei, L.: 2015, ImageNet Large Scale Visual Recognition Challenge, International Journal of Computer Vision (IJCV) .

Santana-Cedr´es, D., Gomez, L., Alem´an-Flores, M., Salgado, A., Esclar´ın, J., Mazorra, L. and Alvarez, L.: 2016, An iterative optimization algorithm for lens distortion correction using two-parameter models, Image Processing On Line .

Scalzo, F. and Piater, J. H.: 2007, Adaptive Patch Features for Object Class Recognition with Learned Hierarchical Models, Conference on Computer Vision and Pattern Recognition (CVPR). Scaramuzza, D. and Fraundorfer, F.: 2011, Visual odometry [tutorial], IEEE Robotics &

Au-tomation Magazine (RAM) .

Schairer, T., Huhle, B., Vorst, P., Schilling, A. and Straßer, W.: 2011, Visual mapping with un-certainty for correspondence-free localization using Gaussian process regression, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

Sch ¨onberger, J. L. and Frahm, J.-M.: 2016, Structure-from-Motion Revisited, Conference on Computer Vision and Pattern Recognition (CVPR).

Schroff, F., Kalenichenko, D. and Philbin, J.: 2015, FaceNet: A Unified Embedding for Face Recognition and Clustering, Conference on Computer Vision and Pattern Recognition (CVPR). Schussel, M. and Pregizer, F.: 2015, Coverage gaps in fingerprinting based indoor

position-ing: The use of hybrid Gaussian Processes, International Conference on Indoor Positioning and Indoor Navigation (IPIN), IEEE.

Sermanet, P., Eigen, D., Zhang, X., Mathieu, M., Fergus, R. and LeCun, Y.: 2013, OverFeat : Integrated Recognition , Localization and Detection using Convolutional Networks, CoRR .

Simonyan, K. and Zisserman, A.: 2014, Very deep convolutional networks for large-scale image recognition, CoRR .

(10)

Sivic, J. and Zisserman, A.: 2003, Video Google: a text retrieval approach to object matching in videos, IEEE International Conference on Computer Vision (ICCV), IEEE.

Song, J. and Park, Y.: 2015, Fingerprint-Based User Positioning Method Using Image Data of Single Camera, International Conference on Indoor Positioning and Indoor Navigation (IPIN). Song, X., Zhao, X., Hu, H. and Fang, L.: 2018, Edgestereo: A context integrated residual

pyramid network for stereo matching, CoRR .

Sparck Jones, K.: 1972, A statistical interpretation of term specificity and its application in retrieval, Journal of documentation .

Strisciuglio, N., Azzopardi, G. and Petkov, N.: 2019, Brain-inspired robust delineation opera-tor, Computer Vision – ECCV 2018 Workshops.

Strisciuglio, N., Azzopardi, G., Vento, M. and Petkov, N.: 2016, Supervised vessel delineation in retinal fundus images with the automatic selection of B-COSFIRE filters, Machine Vision and Applications .

Strisciuglio, N. and Petkov, N.: 2017, Delineation of line patterns in images using b-cosfire filters, IWOBI.

Strisciuglio, N., Tylecek, R., Petkov, N., Bieber, P., Hemming, J., van Henten, E., Sattler, T., Pollefeys, M., Gevers, T., Brox, T. and Fisher, R. B.: 2018, Trimbot2020: an outdoor robot for automatic gardening, International Symposium on Robotics, VDE Verlag GmbH Berlin -Offenbach.

Sturm, P. and Ramalingam, S.: 2004, A Generic Concept for Camera Calibration, in J. Matas and T. Pajdla (eds), European Conference on Computer Vision (ECCV), Lecture Notes in Com-puter Science (LNCS), Springer-Verlag, Prague, Czech Republic.

S ¨underhauf, N., Neubert, P. and Protzel, P.: 2013, Are we there yet? challenging seqslam on a 3000 km journey across all four seasons, Workshop on Long-Term Autonomy, International Conference on Robotics and Automation (ICRA) .

S ¨underhauf, N., Shirazi, S., Dayoub, F., Upcroft, B., Milford, M., Sunderhauf, N., Dayoub, F., Sareh, S., Ben, U. and Michael, M.: 2015, On the Performance of ConvNet Features for Place Recognition, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). Sunderhauf, N., Shirazi, S., Jacobson, A., Dayoub, F., Pepperell, E., Upcroft, B. and Milford,

M.: 2015, Place recognition with convnet landmarks: Viewpoint-robust, condition-robust, training-free, Robotics: Science and Systems (RSS).

Szegedy, C., Inc, G., Zaremba, W., Sutskever, I., Inc, G., Bruna, J., Erhan, D., Inc, G., Goodfel-low, I. and Fergus, R.: 2014, Intriguing properties of neural networks, In ICLR.

Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V. and Rabinovich, A.: 2015, Going deeper with convolutions, Conference on Computer Vision and Pattern Recognition (CVPR).

(11)

Szeliski, R.: 1999, Prediction error as a quality metric for motion and stereo, IEEE International Conference on Computer Vision (ICCV).

Taneja, A., Ballan, L. and Pollefeys, M.: 2015, Never Get Lost Again: Vision Based Navigation Using StreetView Images, Asian Conference on Computer Vision (ACCV), Springer.

Taylor, M. M., Sedigh-Sarvestani, M., Vigeland, L., Palmer, L. A. and Contreras, D.: 2018, Inhibition in simple cell receptive fields is broad and off-subregion biased, Journal of Neu-roscience .

Temel, D., Kwon*, G., M.Prabhuhankar* and AlRegib, G.: 2017, CURE-TSR: Challenging un-real and un-real environments for traffic sign recognition, Advances in Neural Information Pro-cessing Systems (NIPS) Machine Learning for Intelligent Transportations Systems Workshop . Temel, D., Lee, J. and AlRegib, G.: 2018, CURE-OR: challenging unreal and real environments

for object recognition, CoRR .

Thrun, S., Burgard, W. and Fox, D.: 2005, Probabilistic robotics, MIT press.

Tomasi, M. and Anedda, P.: 2013, Using smartphone sensor to improve image-based scene recognition, International Conference on Indoor Positioning and Indoor Navigation (IPIN). Triggs, B., McLauchlan, P. F., Hartley, R. I. and Fitzgibbon, A. W.: 2000, Bundle adjustment

— a modern synthesis, in B. Triggs, A. Zisserman and R. Szeliski (eds), Vision Algorithms: Theory and Practice, Springer Berlin Heidelberg, Berlin, Heidelberg.

Vaca-Castano, G., Zamir, A. R. and Shah, M.: 2012, City scale geo-spatial trajectory estimation of a moving camera, Conference on Computer Vision and Pattern Recognition (CVPR). Vasiljevic, I., Chakrabarti, A. and Shakhnarovich, G.: 2016, Examining the impact of blur on

recognition by convolutional networks, CoRR .

Wang, J., Song, Y., Leung, T., Rosenberg, C., Wang, J., Philbin, J., Chen, B. and Wu, Y.: 2014, Learning Fine-Grained Image Similarity with Deep Ranking, Conference on Computer Vision and Pattern Recognition (CVPR).

Weiler, M., Hamprecht, F. A. and Storath, M.: 2017, Learning steerable filters for rotation equivariant cnns, CoRR .

Werner, M., Kessel, M. and Marouane, C.: 2011, Indoor positioning using smartphone camera, International Conference on Indoor Positioning and Indoor Navigation (IPIN).

Wohlhart, P. and Lepetit, V.: 2015, Learning Descriptors for Object Recognition and 3D Pose Estimation, Conference on Computer Vision and Pattern Recognition (CVPR).

Workman, S., Greenwell, C., Zhai, M., Baltenberger, R. and Jacobs, N.: 2015, DEEPFOCAL: A method for direct focal length estimation, IEEE International Conference on Image Processing (ICIP).

(12)

Worrall, D. E., Garbin, S. J., Turmukhambetov, D. and Brostow, G. J.: 2016, Harmonic net-works: Deep translation and rotation equivariance, CoRR .

Xiao, J., Ehinger, K. A., Oliva, A. and Torralba, A.: 2012, Recognizing scene viewpoint us-ing panoramic place representation, Conference on Computer Vision and Pattern Recognition (CVPR), IEEE.

Yin, X., Wang, X., Yu, J., Zhang, M., Fua, P. and Tao, D.: 2018, Fisheyerecnet: A multi-context collaborative deep network for fisheye image rectification, CoRR .

Zagoruyko, S. and Komodakis, N.: 2016, Wide residual networks, British Machine Vision Con-ference (BMVC).

Zhai, M., Workman, S. and Jacobs, N.: 2016, Detecting Vanishing Points Using Global Image Context in a Non-Manhattan World, Conference on Computer Vision and Pattern Recognition (CVPR).

Zheng, S., Song, Y., Leung, T. and Goodfellow, I.: 2016, Improving the robustness of deep neural networks via stability training, Conference on Computer Vision and Pattern Recognition (CVPR).

Zoumpourlis, G., Doumanoglou, A., Vretos, N. and Daras, P.: 2017, Non-linear convolution filters for cnn-based learning, IEEE International Conference on Computer Vision (ICCV).

(13)

Referenties

GERELATEERDE DOCUMENTEN

The results that we achieved using LeNet on the MNIST data set, and ResNet and DenseNet models on corrupted versions of the CIFAR data set, namely the CIFAR-C bench- mark data

The work described in chapter 2 involves training a convolutional neural network using a fully supervised scheme where panoramas are cropped to simulate images taken with cameras

• Best Paper Award: Manuel L ´opez-Antequera, Nicolai Petkov, and Javier Gonz´alez- Jim´enez, “Image-based localization using Gaussian processes,” International Conference on

Manuel Lopez-Antequera, Roger Mar´ı Molas, Pau Gargallo, Yubin Kuang, Javier Gonzalez-Jimenez, Gloria Haro, “ Deep Single Image Camera Calibration with Radial Distortion” The

Contemporary machine learning techniques might achieve surprisingly good re- sults when used as black boxes, however, proper use of domain knowledge will make the difference

We configure an inhibition-augmented COSFIRE filter by using two different types of prototype patterns, namely one positive pattern and one or more negative pattern(s), in order

In the second approach, we first localize the optic disc by using two types of trainable COSFIRE (Combination of Shifted Filter Responses) filters (Azzopardi and Petkov, 2013b):

We evaluate the proposed approach on eight public data sets and compare the obtained results with the manual annotation provided by a glau- coma expert from the University