-
Notifications
You must be signed in to change notification settings - Fork 185
Open
Description
g_loss = tf.reduce_mean(tf.nn.sigmoid_cross_entropy_with_logits(logits = Dg, labels = tf.ones_like(Dg)))
trainerG = tf.train.AdamOptimizer().minimize(g_loss, var_list=g_vars)
These codes seem to minimize g_loss for Generator to generlize nearly true pictures, but when I training, gLoss returned by _,gLoss = sess.run([trainerG,g_loss],feed_dict={z_placeholder:z_batch}) is increasing, while dLoss is decreasing as designed, why did this happen?
ps. the results seem that generator did learned something
Metadata
Metadata
Assignees
Labels
No labels