-
Notifications
You must be signed in to change notification settings - Fork 478
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🙅 Inaccurate model coref predictions master thread #215
Comments
updated to include article url, doh * For this article, the model struggles with This is the mentions array:
Webb is broken out as if it is a last name when it is the part of the telescope's name. In general the model struggles to tell the difference between the two telescopes mentioned in the article. I'm wondering if Bert Span-based model might be an option for the next release? I tried the above text on it and it is slightly better (though still imperfect). https://github.com/mandarjoshi90/coref |
Hey @svlandeg ,
|
+1. I've noticed discrepancies between performance locally and in the dev environment. @svlandeg @thomwolf, can you comment on what exactly is currently served by the demo environment? |
I wasn't involved with this project when the demo environment was created. However, note that it's not just the version that was trained that makes a difference, but also the specific hyperparameters used for making the predictions. So that is definitely something you can "play" with too. |
Thanks, |
Same here. With exactly the same sentence provided from the sample site, I tried all kinds of options for hyperparameters, but I am still unable to reproduce the result. I installed Spacy and Neuralcoref from source on a brand new Docker, so it should not be a problem of environment. In the sample page, the score is usually like 3 to 15, where on my environment the result is always like -2 to 2. I am wondering how exactly to reproduce the result on the sample page. Thank you very much! resolved by compiling spacy 2.1.0 and neuralcoref from source code. |
hello, |
Not exactly. I am just saying that this would increase the accuracy on my side, from unusable to usable. |
It would be very helpful if someone could shed some light on what model and combination of package versions are used in the demo environment. Like others here I'm not able to reproduce what I see on the demo in my local setup, even when rolling spacy back to 2.1.3 and building neuralcoref from source. It feels like the model served in the demo environment is a different model or it was trained with different word embeddings. Are pretrained neuralcoref models tied to any specific spacy language model tag? I've played with the parameters like others here with some improvement but I'm still seeing systematic differences with the live demo. |
Totally agree!!! Online demo is an ideal starting point, but only if it is reproducible :( |
I'm not able to share much in the way of text for confidentiality reasons, but I'm noticing that the pre-trained model seems to be gravitating toward resolving "us" to "We." It might be useful to be able to blacklist certain words (e.g. "We") as never being satisfactory coreferents.
|
I agree- to have a more customizable blacklist (including they, it, these,
who) would be wonderful. this is great tech as it is but just a suggestion
…On Mon, Jul 27, 2020 at 2:57 PM Keating950 ***@***.***> wrote:
I'm not able to share much in the way of text for confidentiality reasons,
but I'm noticing that the pre-trained model seems to be gravitating toward
resolving "us" to "We." It might be useful to be able to blacklist certain
words (e.g. "We") as never been satisfactory coreferents.
\< It is not up to us to rectify things
\---
\> It is not up to We to rectify things
\< It is absolutely an issue, but not only to us
\---
\> It is absolutely an issue, but not only to We
- neuralcoref 4.0
- spacy 2.3.2
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#215 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALGMOQXWSF373JA2LBJJMJDR5XEXZANCNFSM4JBIG4EQ>
.
|
If you're interested in this feature, I've added to in my fork of this project. I'm still making sure it works, so I'm all ears to any feedback and review. |
Thanks alot Keating950!
I see that in your fork NO_COREF_LIST = ["i", "me", "my", "you", "your"] no
longer exists in *train/document.py*, nor in *neuralcoref.pyx*. So to me it
seems we do not place our custom blacklist directly in the source code.
Does that mean each time neuralcoref is instantiated I just pass the custom
blacklist? for example:
*coref = neuralcoref.NeuralCoref(nlp.vocab, greedyness=0.75 ,
blacklist = ["i", "me", "my", "you", "your", "they", "their", "it"] )*
Just verify I implement your fork as intended. Thanks
…On Sun, Aug 9, 2020 at 3:21 PM Keating950 ***@***.***> wrote:
I agree- to have a more customizable blacklist (including they, it, these,
who) would be wonderful. this is great tech as it is but just a suggestion
If you're interested in this feature, I've added it in my fork
<https://github.com/Keating950/neuralcoref> of this project. I'm still
making sure it works, so I'm all ears to any feedback and review.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#215 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALGMOQQCPSE3FI3ZWR7F6U3R73ZL5ANCNFSM4JBIG4EQ>
.
|
Yup, that's exactly right. I've updated the README. Feel free to open an issue on that repo if you have any other questions. |
Thanks alot man.
…On Sun, Aug 9, 2020 at 10:31 PM Keating950 ***@***.***> wrote:
Yup, that's exactly right. I've updated the README. Feel free to open an
issue on that repo if you have any other questions.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#215 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALGMOQTPEWADAXYJ75QCNWLR75LX5ANCNFSM4JBIG4EQ>
.
|
Thanks for your work @Keating950! Very helpful! |
@lauwauw Thanks! I've merged in the latest changes from this repo in light of the renewed interest. |
Master thread for collecting incorrect and/or problematic coreference predictions with the pretrained models. These can be interesting test cases when training the next version of the model.
The text was updated successfully, but these errors were encountered: