-
Notifications
You must be signed in to change notification settings - Fork 614
Guidelines for new contributions #2194
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
We partially talked about this on the upstreaming treshold at tensorflow/community#239 (comment) |
Yep, this is a good idea. A "feature submission" suggestion under the contribution section in |
We need to continue here cause that one was closed. We don't handle the upstreaming process anymore on our side. |
Just to collect another resource: |
Yeah that's very interesting. Thanks @bhack cc: @seanpmorgan @WindQAQ what are your thoughts? |
Great suggestion. We've hinted at it, but never formalized anything as has been mentioned.
Very cool but when I tried the live version it wasn't very performant. I think I'd prefer a simple way to determine this, though the tool is super cool / useful
This would be a nice way to offload the decision, a bit concerned about highly cited things being turned down if not there though.
Overall I'd lean to a simple citation number as metric, but could be convinced easily if there is any consensus. Also interested in ways to identify that number of citations? We could pull some data on our historical acceptances. |
Yes. We need to do this. We can use |
This Is really about define a treshold policy. With a citation only metric we will probably exclude proposal on tops like: http://www.arxiv-sanity.com/top and http://www.arxiv-sanity.com/toptwtr |
@bhack yes. But we will make another proposal for these things. Sometimes a method that is very promising takes some time before getting into the limelight or enough number of citations. In that case, we will go through the relevant paper and a consensus can be taken whether we want to include the new functionality or not. Though this will take much longer to review |
In that case having an official reference implementation it will have its weight in the evaluation. |
Why not to put a simple voting system? In the end of the day, it's all about implementation in TFA. Something cited doesn't necessary imply eagering to utilize it [in TF/TFA/whatever]. Although the same issue with threshold. 😄 Maybe some % of active users? |
@failure-to-thrive this could be a feature we could include, but I don't think relying on the opinion of members is an effective evaluation metric. Some users may not have an opinion on a certain implementation or they might bring their personal bias into the voting system. This system might become chaotic for the user's who want to suggest a new feature since they won't be able to determine the requirements for their proposal. I think the number of
|
Thanks @failure-to-thrive for your feedback.
A voting system here would be abused. For example, let's say there is a new optimizer that came out yesterday which claims better results than any other existing optimizer. Any PR related to it would gather a huge number of votes because everyone wants to try out new things. The problem is that papers don't cover enough use cases to prove that whatever is presented in a paper is actually useful. You would only know if something works properly if it has a good number of citations. |
Having observed and/or contributed in scientific publication process in several disciplines for more than a decade, In my opinion it is generally not worth it to implement anything until it has been widely recognized by the community. Especially so on hot topics like ML.
The most critical single requirement for any article to get trough peer review is that it shows improvement in something, so that really should not be taken as an indicator of a particularly useful method. In my experience, review works recapping the advances in some area are particularly good sources for new contributions. A work or a series of works cited in such an article are typically useful enough to be implemented. The problem here is that such articles seem to be rare, and might sometimes lead to excessive delays.
As a practical example, I recently tested a bunch of brand new optimizers like Radam, LAMB and Yogi with and without Lookahead against the good old Adam in my classification tasks. Found out that most combinations decreased the mean accuracy and/or increased result variance as compared to Adam. In this case I was positively surprised that plain Radam actually significantly improved the mean accuracy and decreased result variance over Adam.
26.10.2020 15.44 Aakash Kumar Nain <[email protected]> kirjoitti:
Thanks @failure-to-thrive<https://github.com/failure-to-thrive> for your feedback.
Why not to put a simple voting system?
A voting system here would be abused. For example, let's say there is a new optimizer that came out yesterday which claims better results than any other existing optimizer. Any PR related to it would gather a huge number of votes because everyone wants to try out new things. The problem is that papers don't cover enough use cases to prove that whatever is presented in a paper is actually useful. You would only know if something works properly if it has a good number of citations.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub<#2194 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/ADNNQ2OGOBJOTNFE2EHWETLSMV4MVANCNFSM4SHDECZQ>.
|
This has been mentioned in a lot of PRs. Let's take a decision and close it in the next meeting |
I believe in the previous meeting we decided to keep 50 citations as a suggested requirement with the ability for any maintainer to override the requirement. I'm currently working on updating the I still think there's a lot of room for more suggestions to narrow down the quality of the requests. |
I will take a look at that. Thank you @Harsh188 |
I really like this package. @seanpmorgan @bhack @WindQAQ thoughts? |
Nice it could be useful for an interactive Github bot |
Yup. |
New contributions are always welcome. However, ML is a rapidly changing field. There are hundreds of papers being published on
arXiv
everyday. With so many papers coming out, we expect new things that aren't readily available in thecore
and people would be wiling to contribute toaddons
.From the perspective of an end user, it makes sense that
addons
incorporate new contributions as much as possible. However from the perspective of a maintainer, things can become overwhelming at some point in time because all contributors won't be active all the time which can lead to stalled issues and problems with new updates/releases (if any).To this end, we need to decide how many
citations
should be there for a new functionality to be added in the ecosystem. For thecore
, it's around 100 citations IIRC. For addons, this bar would be much lower but we still need to decide a number. We should update this in theREADME
as well.There are certain ways to track the number of citations but one of the coolest tools is this one: https://github.com/dennybritz/papergraph
The text was updated successfully, but these errors were encountered: