| from modules.layers.transformers import (TransformerDecoderLayer, |
| TransformerEncoderLayer, |
| TransformerSpatialDecoderLayer) |
|
|
| class UnifiedSpatialCrossEncoderV2(nn.Module): |
| """ |
| spatial_dim: spatial feature dim, used to modify attention |
| dim_loc: |
| """ |
|
|
| def __init__(self, cfg, hidden_size=768, dim_feedforward=2048, num_attention_heads=12, num_layers=4, dim_loc=6): |
| super().__init__() |
|
|
| |
| unified_encoder_layer = TransformerEncoderLayer(hidden_size, num_attention_heads, dim_feedforward=dim_feedforward) |
| self.unified_encoder = layer_repeat(unified_encoder_layer, num_layers) |
|
|
| |
| loc_layer = nn.Sequential( |
| nn.Linear(dim_loc, hidden_size), |
| nn.LayerNorm(hidden_size), |
| ) |
| self.loc_layers = layer_repeat(loc_layer, 1) |
|
|
| |
| self.token_type_embeddings = nn.Embedding(2, hidden_size) |
|
|
| self.apply(_init_weights_bert) |
|
|
| def forward( |
| self, txt_embeds, txt_masks, obj_embeds, obj_locs, obj_masks, |
| output_attentions=False, output_hidden_states=False, **kwargs |
| ): |
| txt_len = txt_embeds.shape[1] |
| obj_len = obj_embeds.shape[1] |
|
|
| for i, unified_layer in enumerate(self.unified_encoder): |
| |
| query_pos = self.loc_layers[0](obj_locs) |
| pc_token_type_ids = torch.ones((obj_embeds.shape[0:2])).long().cuda() |
| pc_type_embeds = self.token_type_embeddings(pc_token_type_ids) |
| obj_embeds = obj_embeds + query_pos + pc_type_embeds |
|
|
| |
| lang_token_type_ids = torch.zeros((txt_embeds.shape[0:2])).long().cuda() |
| lang_type_embeds = self.token_type_embeddings(lang_token_type_ids) |
| txt_embeds = txt_embeds + lang_type_embeds |
|
|
| |
| joint_embeds = torch.cat((txt_embeds, obj_embeds), dim=1) |
| joint_masks = torch.cat((txt_masks, obj_masks), dim=1) |
|
|
| |
| joint_embeds, self_attn_matrices = unified_layer(joint_embeds, |
| tgt_key_padding_mask=joint_masks.logical_not()) |
|
|
| |
| txt_embeds, obj_embeds = torch.split(joint_embeds, [txt_len, obj_len], dim=1) |
|
|
| return txt_embeds, obj_embeds |