Task-Agnostic Vision Transformer for Distributed Learning of Image Processing

Research output: Contribution to journalArticlepeer-review

16 Scopus citations

Abstract

Recently, distributed learning approaches have been studied for using data from multiple sources without sharing them, but they are not usually suitable in applications where each client carries out different tasks. Meanwhile, Transformer has been widely explored in computer vision area due to its capability to learn the common representation through global attention. By leveraging the advantages of Transformer, here we present a new distributed learning framework for multiple image processing tasks, allowing clients to learn distinct tasks with their local data. This arises from a disentangled representation of local and non-local features using a task-specific head/tail and a task-agnostic Vision Transformer. Each client learns a translation from its own task to a common representation using the task-specific networks, while the Transformer body on the server learns global attention between the features embedded in the representation. To enable decomposition between the task-specific and common representations, we propose an alternating training strategy between clients and server. Experimental results on distributed learning for various tasks show that our method synergistically improves the performance of each client with its own data.

Original languageEnglish
Article number3226892
Pages (from-to)203-218
Number of pages16
JournalIEEE Transactions on Image Processing
Volume32
DOIs
StatePublished - 2023
Externally publishedYes

Keywords

  • Distributed learning
  • image processing
  • task-agnostic learning
  • transformer

Fingerprint

Dive into the research topics of 'Task-Agnostic Vision Transformer for Distributed Learning of Image Processing'. Together they form a unique fingerprint.

Cite this