Skip to content

Sigmoid activation in mixed activation for linear outputs #45

@tsrobinson

Description

@tsrobinson

As discovered by @antndlcrx, we can get better performance by using sigmoid rather than identity activation functions for continuous inputs, since we scale these to 0-1 in the data preprocessing.

@tsrobinson to update codebase to make this default behavior.

Metadata

Metadata

Assignees

Labels

No labels
No labels

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions