Skip to content

Why using relu before softmax? #1

@wang975679801

Description

@wang975679801

In the line 196,
gate_network = tf.contrib.layers.fully_connected( inputs=input, num_outputs=subexpert_nums, activation_fn=tf.nn.relu, \ weights_regularizer=l2_reg)
and then gate_network = tf.nn.softmax(gate_network, axis=1)
It is confusing.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions