-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rating methods comparison #33
Comments
@amirrr @markwhiting Here's a comparison between humans and LLaMA-3.1 in labeling 6 dimensions for statements in our previous corpus. Note that |
Thanks, do these look right to you? I note that we did have reversed classes for some features that we reverse in the start of the R script from the prior experiment, and just want to ensure that the ratings you're comparing with reflect the correct state. |
I think they should be correct. In the original data file (and R script), only |
OK, thanks. Qualitatively do you have any qualms with this? Also, how much will this change the design points we can cover with our current corpus? |
The numbers are lower than I expected, although upon manual inspection some human-given labels may not be reasonable. For example:
I'm not sure how much the design points will have to change, but I think we should do a manual evaluation of LLaMA-3.1's labels in the new dataset. Perhaps we can sample 200 statements, label them ourselves independently (at least by me and Amir), and then check whether we agree with each other and with the LLM. |
Thanks — yeah a manual pass seems worth it. At worst we will then have some ground truth to work with.
If we want to scale that a bit more we can also put a more sophisticated version of the task in front of turkers.
|
TODO:
|
I have prompted LLaMA-3-8B to classify all N = 10,110 statements here. The same features previously obtained by Amir are in this file. Note that they are only available for the first N = 8,814 statements. For comparison, I have put them together. For feature |
Here are 200 statements randomly sampled from those for which |
Will start working on them now. |
This is for the 200-statement dataset that Amir and I annotated independently. I used LLaMA-3.1-8B to make predictions for the 6 dimensions. The script can be found here. Essentially I asked LLaMA to output either of the two tokens: The figure attached here shows the ROC curves for these scores. There are two settings: when using my labels as the ground truth (black), and when using Amir's labels (blue). |
I performed the same analysis for the original 4,407-statement dataset. Here I treat the mode of human annotations as the ground truth, similar to what we've done so far. I also use the mean of these annotations. For example, if a statement gets 5 annotations, 3 of which say "physical" and 2 say "social", then the human score for "physical" is 3 / 5 = 60%. The figure below shows the correlation between LLaMA scores and human scores. I use the Spearman correlation here, and all of them are statistically significant. |
Interesting. Thanks for this. It feels like we still don't have something that is strongly aligned, but I wonder what the best next step is for getting to a place where we are satisfied. |
Would you mind doing the average analysis on the 200 that you two rated? |
I would say the following categories are pretty robust: The hardest for me to annotate were:
- |
I think even the resolution of three possible values 0, .5 or 1, I'm hoping that we see improved alignment over the binary case, because I think that would let us argue that the continuous categories from the model are sufficiently more informative that it's worth treating those as our standard. So I'd still be interested to look at that if you can easily run it. Another option would be to take these 200 statements and a few rounds of human raters on them to see if we get more consistent responses that way too. Lastly, I really don't mind if we want to try to improve our definitions of these concepts as I think that will benefit us as well as the models. |
Thank you. Interesting that the model sometimes captures disagreement nicely and sometimes does not. Also interesting that LLaMA's probability for some is skewed so high. I wonder if we should post tune those to have better coverage and if that would help at all.
Is this after a revisit to the ratings (I think @amirrr was planning to look at |
The text was updated successfully, but these errors were encountered: