-
Notifications
You must be signed in to change notification settings - Fork 12
Adding "SignBLEU: Automatic Evaluation of Multi-channel Sign Language Translation" and some eval metrics discussion #77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
All right, I think we're about ready for review: https://chatgpt.com/share/2a913f6c-e344-4a00-86ab-357bbae6ec8d is the conversation with ChatGPT and revisions of the summary. Took some, discarded others Questions:
|
Added some more detail on pose output metrics, and some citations for APE |
src/index.md
Outdated
Naively, works in this domain have used metrics such as mean squared error (MSE) or Average Position Error (APE) for pose outputs [ahuja2019Language2PoseNaturalLanguage;ghosh2021SynthesisCompositionalAnimations;petrovich2022TEMOSGeneratingDiverse]. | ||
However, these metrics have significant limitations for Sign Language Production. | ||
|
||
For example, MSE and APE do not account for variations in sequence length, for cases where where the same sign might take different amounts of time to produce. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
where where
"different amounts of time" sounds strange to me
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
How about:
For example, MSE and APE do not account for variations in sequence length.
The same sign may not always take exactly the same amount of time to produce.
Or various options here: https://chatgpt.com/share/e4e5cde3-fccf-4e40-ab2f-5b210db6cf6f
cleong110#21 details and checklist