From 073b821aa2a1ea9dfc56c66d7cfc17e25d755608 Mon Sep 17 00:00:00 2001 From: Junnan Li Date: Tue, 15 Feb 2022 11:20:04 +0800 Subject: [PATCH] Update blip_pretrain.py --- models/blip_pretrain.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/models/blip_pretrain.py b/models/blip_pretrain.py index 0684202..e42ce5f 100644 --- a/models/blip_pretrain.py +++ b/models/blip_pretrain.py @@ -91,7 +91,7 @@ class BLIP_Pretrain(nn.Module): decoder_config.encoder_width = vision_width self.text_decoder = BertLMHeadModel.from_pretrained('bert-base-uncased',config=decoder_config) self.text_decoder.resize_token_embeddings(len(self.tokenizer)) - tie_encoder_decoder_weights(self.text_decoder.bert,self.text_encoder,'','/attention') + tie_encoder_decoder_weights(self.text_encoder,self.text_decoder.bert,'','/attention') def forward(self, image, caption, alpha):