2

within Delving into Local Features for Open-Set Domain Adaptation in Fundus Image Analysis paper. I got trouble in understanding their cluster-aware contrastive adaption $\mathcal{L}_\text{cca}$.

cluster-aware contrastive adaption loss

I don't get it how they defined "$i$ and $j$ are aligned", and how is consider "mis-aligned".

This is what they say when they highlighted their contributions: (CCA) method is proposed to enforce the local features of aligned clusters across domains to remain close-by, while pushing those of the misaligned clusters far apart.

CCA illustration

(Blue is source domain, Orange is target domain).This is open-set domain adaptation. I believe each prototype is the mean (center) of each cluster. and they say $i$ (a cluster) from source, $j$ (a cluster) from target.

After scanning the paper for a week, I still don't understand what is "aligned" here. I sent them email twice (politely) but no reply.

Can any one enlighten me a little bit?

nbro
  • 42,615
  • 12
  • 119
  • 217

1 Answers1

1

To provide some context for this answer, the referenced paper is dealing with the problem of unsupervised domain adaptation (UDA). In UDA, there are two (or more) datasets, each drawn from a different, but related domain. One of these datasets (the source) is labelled, and we want to use that data to train a machine learning model to predict the labels for the unlabelled (or target) datasets. The complicating factor is that the distribution of the data from each domain is different, so a model trained on the source data in the usual way does not generalise to the target domain. In open-set domain adaptation the distributions of the source and target labels are different and there are one or more target labels that are not represented in the source dataset.

The paper addresses this problem by training a model jointly on the source and target data. The idea behind the joint training is that the deep learning model is trained to extract domain-invariant features from the input, which are then used in a classifier. As the classifier is trained on these domain-invariant features, it should be able to classify samples from the target domain, as well as from the source domain. For the source data, the usual cross-entropy loss is used. For the target data they use two loss functions, one of which is the above-mentioned cluster-aware contrastive adaptive loss.

To calculate the cluster-aware contrastive adaptive loss, the learned feature maps for the source and target data are clustered separately, and then the clusters are aligned. Minimizing this loss function trains the model to extract features that bring aligned clusters closer together and pushes unaligned clusters further apart.

The authors don't go into a lot of detail about how the clusters are aligned. But they do tell us they use the "domain consensus score" and cite the paper Domain consensus clustering for universal domain adaptation by Guangrui Li et al. According to this paper, they find the centre of each cluster, then for each cluster in one domain they simply find the cluster in the other domain with the nearest cluster centre. If the two clusters (one from each domain) are each other's nearest centre, then they describe this as reaching consensus. They also compute a domain consensus score, which is simply the mean of the proportion of cluster members that are located in the intersection of the two clusters.

Lynn
  • 141
  • 1
  • 4