1: Initialize the encoder parameters Θ E r , weights and biases of n adaption layers: A Θ A i = { w A i , b A i } ( i = 1 , ⋯ , n ) . Freeze all parameters of D Θ D 1 , ⋯ , D Θ D n .
2: for each epoch do
3: Sample v i ∈ d i ( i = 1 , ⋯ , n )
4: Traverse θ ∈ { Θ E r , Θ A 1 , ⋯ , Θ A i , ⋯ , Θ A n }
5: Apply Backpropagation to update θ for all adaption layers and encoder.
θ ← θ − η ∇ θ { ∑ i = 1 n E v i L ( v i , D Θ D i [ A Θ A i ( E Θ E r ( v i ) ) ] ) } (8)
where v i ∈ V { v 1 , ⋯ , v i , ⋯ , v n } ( i = 1 , ⋯ , n )