Please use this identifier to cite or link to this item:
http://hdl.handle.net/1893/35587
Appears in Collections: | Computing Science and Mathematics Conference Papers and Proceedings |
Author(s): | Goncalves, Diogo Nunes Junior, Jose Marcato Zamboni, Pedro Pistori, Hemerson Li, Jonathan Nogueira, Keiller Goncalves, Wesley |
Contact Email: | keiller.nogueira@stir.ac.uk |
Title: | MTLSegFormer: Multi-Task Learning With Transformers for Semantic Segmentation in Precision Agriculture |
Citation: | Goncalves DN, Junior JM, Zamboni P, Pistori H, Li J, Nogueira K & Goncalves W (2023) MTLSegFormer: Multi-Task Learning With Transformers for Semantic Segmentation in Precision Agriculture. In: <i>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops</i>. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 18.06.2023-22.06.2023. Piscataway, NJ, USA: IEEE, pp. 6290-6298. https://doi.org/10.1109/CVPRW59228.2023.00669 |
Issue Date: | 2023 |
Date Deposited: | 27-Nov-2023 |
Conference Name: | 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) |
Conference Dates: | 2023-06-18 - 2023-06-22 |
Conference Location: | Vancouver, BC, Canada |
Abstract: | Multi-task learning has proven to be effective in improving the performance of correlated tasks. Most of the existing methods use a backbone to extract initial features with independent branches for each task, and the exchange of information between the branches usually occurs through the concatenation or sum of the feature maps of the branches. However, this type of information exchange does not directly consider the local characteristics of the image nor the level of importance or correlation between the tasks. In this paper, we propose a semantic segmentation method, MTLSegFormer, which combines multi-task learning and attention mechanisms. After the backbone feature extraction, two feature maps are learned for each task. The first map is proposed to learn features related to its task, while the second map is obtained by applying learned visual attention to locally re-weigh the feature maps of the other tasks. In this way, weights are assigned to local regions of the image of other tasks that have greater importance for the specific task. Finally, the two maps are combined and used to solve a task. We tested the performance in two challenging problems with correlated tasks and observed a significant improvement in accuracy, mainly in tasks with high dependence on the others. |
Status: | AM - Accepted Manuscript |
Rights: | © 2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
NunesGoncalves-etal-CVPR-2023.pdf | Fulltext - Accepted Version | 10.01 MB | Adobe PDF | View/Open |
This item is protected by original copyright |
Items in the Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
The metadata of the records in the Repository are available under the CC0 public domain dedication: No Rights Reserved https://creativecommons.org/publicdomain/zero/1.0/
If you believe that any material held in STORRE infringes copyright, please contact library@stir.ac.uk providing details and we will remove the Work from public display in STORRE and investigate your claim.