Domain Adaptation with Randomized Multilinear Adversarial Networks Domain Adaptation with Randomized Multilinear Adversarial Networks
Paper summary _Objective:_ Perform domain-adaptation by adapting several layers using a randomized representation and not just the final layer thus performing alignment of the joint distribution and not just the marginals. _Dataset:_ [Office]( and [ImageCLEF-DA1]( ## Inner-workings: Basically an improvement on [RevGrad]( where instead of using the last embedding layer for the discriminator, a bunch of them is used. To avoid dimension explosion when using the tensor product of all layers they instead use a randomized multi-linear representation: [![screen shot 2017-06-01 at 5 35 46 pm](]( Where: * d is the dimension of the embedding (they use 1024) * R is random matrix for which each element as a null average and variance of 1 (Bernoulli, Gaussian and Uniform are tried) * z^l is the l-th layer * ⊙ represents the Hadamard product In practice they don't use all layers but just the 3-4 last layers for ResNet and AlexNet. ## Architecture: [![screen shot 2017-06-01 at 5 34 44 pm](]( They use the usual losses for domain adaptation with: - F minimizing the cross-entropy loss for classification and trying to reduce the gap between the distributions (indicated by D). - D maximizing the gap between the distributions. [![screen shot 2017-06-01 at 5 40 53 pm](]( ## Results: Improvement on state-of-the-art results for most tasks in the dataset, very easy to implement with any pre-trained network out of the box.
Domain Adaptation with Randomized Multilinear Adversarial Networks
Long, Mingsheng and Cao, Zhangjie and Wang, Jianmin and Jordan, Michael I.
arXiv e-Print archive - 2017 via Local Bibsonomy
Keywords: dblp

Summary by Léo Paillier 3 years ago
Your comment: allows researchers to publish paper summaries that are voted on and ranked!

Sponsored by: and