-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
negative weight for "right" feature #7
Comments
Thanks for asking this interesting question. I'm not affiliated with the paper but here's a plot that I think helps me understand the problem: it shows the Even assuming that This would suggest that the bulk of the regression's predictive power comes from the bias term, which the authors allude to in Table 2's last row, which gives the mean-absolute-error in per-session score for a dumb predictor that just always returned the average score of 0.859 ( If this understanding is correct, it could imply that optimizing for weights that drive error in
both to zero is problematic. If I were to hazard a guess, you'd really want to be Bayesian about this: after potentially hundreds (thousands?) of times seeing a word, you have to arrive at some prior about your future performance: if you happen to fail after that, that shouldn't automatically become a major outlier for the optimization to deal with, but rather should be more gracefully be folded into a posterior. (Ebisu tries to do something like this.) You might argue that with sufficient data, that should automatically happen. Given that the bias term is so much larger than the two features' weights in HLR-lex, it would appear that's not happening. And furthermore, relying on more data to get the "right" answer (in our case, a positive and maybe sizable weight for |
Thanks for commenting with so many details! I will be looking into Ebisu :) |
@sueyhan previously opened a ticket for this, but then closed it without getting a response.
These are the model weights I get for training without lexical features (
python experiment.py -l settles.acl16.learning_traces.13m.csv.gz
):wrong -0.2245
right -0.0125
bias 7.5365
I do not see how it can be correct that the
right
feature has a negative weight. This will cause the half life to get shorter as a user gets more correct answers, and therefore the model will predict a lower and lower probability of the user getting correct answers.How can this be correct?
The text was updated successfully, but these errors were encountered: