Skip to content

HAN的attention里为什么加reduce_sum和reduce_max? #139

@fengdoudou1895

Description

@fengdoudou1895

在HAN的attention里面看到:
attetion_logits = tf.reduce_sum(hidden_state_context_similarity,axis = 2) attention_logits_max = tf.reduce_max(attention_logits, axis = 1,keep_dims = True) p_attention = tf.nn.softmax(attetion_logits-attention_logits_max)
原论文里没看到这个操作,请问这是为什么呢?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions