You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The paper say: "standard deviations of the outputs of the last hidden layer on the training data as surrogate for Rademacher complexities",
And customizing_adanet.ipynb example use tf.constant(1), is the value choose is robust to the conclusion ?
Or should set the value by standard deviations ?
The text was updated successfully, but these errors were encountered:
@svjack: We may open source some of our complexity measures including Standard Deviation. Currently the one we have in adanet_objective.ipynb is sqrt(d) where d is the number of hidden layers. For the customizing_adanet.ipynb, we use a constant complexity since there is only one candidate at each iteration and they all have the same hyperparameters and therefore complexity.
The paper say: "standard deviations of the outputs of the last hidden layer on the training data as surrogate for Rademacher complexities",
And customizing_adanet.ipynb example use tf.constant(1), is the value choose is robust to the conclusion ?
Or should set the value by standard deviations ?
The text was updated successfully, but these errors were encountered: