Login Paper Search My Schedule Paper Index Help

My ICIP 2021 Schedule

Note: Your custom schedule will not be saved unless you create a new account or login to an existing account.
  1. Create a login based on your email (takes less than one minute)
  2. Perform 'Paper Search'
  3. Select papers that you desire to save in your personalized schedule
  4. Click on 'My Schedule' to see the current list of selected papers
  5. Click on 'Printable Version' to create a separate window suitable for printing (the header and menu will appear, but will not actually print)

Paper Detail

Paper IDMLR-APPL-IP-8.6
Paper Title CRAQUELURENET: MATCHING THE CRACK STRUCTURE IN HISTORICAL PAINTINGS FOR MULTI-MODAL IMAGE REGISTRATION
Authors Aline Sindel, Andreas Maier, Vincent Christlein, Friedrich-Alexander University Erlangen-Nürnberg (FAU), Germany
SessionMLR-APPL-IP-8: Machine learning for image processing 8
LocationArea E
Session Time:Wednesday, 22 September, 14:30 - 16:00
Presentation Time:Wednesday, 22 September, 14:30 - 16:00
Presentation Poster
Topic Applications of Machine Learning: Machine learning for image processing
IEEE Xplore Open Preview  Click here to view in IEEE Xplore
Abstract Visual light photography, infrared reflectography, ultraviolet fluorescence photography and x-radiography reveal even hidden compositional layers in paintings. To investigate the connections between these images, a multi-modal registration is essential. Due to varying image resolutions, modality dependent image content and depiction styles, registration poses a challenge. Historical paintings usually show crack structures called craquelure in the paint. Since craquelure is visible by all modalities, we extract craquelure features for our multi-modal registration method using a convolutional neural network. We jointly train our keypoint detector and descriptor using multi-task learning. We created a multi-modal dataset of historical paintings with keypoint pair annotations and class labels for craquelure detection and matching. Our method demonstrates the best registration performance on the multi-modal dataset in comparison to competing methods.