Skip to content

Reviewer rating #52

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
ThomasKranitsas opened this issue Jan 20, 2017 · 20 comments
Open

Reviewer rating #52

ThomasKranitsas opened this issue Jan 20, 2017 · 20 comments

Comments

@ThomasKranitsas
Copy link

There were many discussions on Slack about bad reviews that lead to the win of an average or even worse a bad submission and in most cases re-appeals.

Many members asked for a rating system for reviewers.

Having a rating system for reviewers might not directly affect the choice of the reviewers for a challenge but it's a good reason for reviewers to perform better reviews. Potentially Topcoder could give some bonus $$ to reliable reviewers.

The rating of a reviewer could be a factor on the reviewer's payment formula thus a good rating will increase the payment but a bad rating will decrease the final payment.

What could be a reason to give a negative rate to a reviewer:

  • Inconsistency in reviews.
  • Missing major/critical issues.
  • Late deliverables.
  • Successful re-appeals.
  • etc..
@birdofpreyru
Copy link

birdofpreyru commented Jan 20, 2017

With a good rating system, we may:

  1. Exclude reviewers with poor rating from the review boards.
  2. Even better, to make a periodical rotations of the review boards, i.e. each month we can exclude from the boards some % of reviewers with lowest ratings (e.g. the bottom 10% of the reviewers leaderboard), and replace them by new people. I believe, it will be interesting for fresh members (now it is not that easy to make a way to the review board, e.g. I am not a reviewer yet :-D also quite motivating for reviewers to be the good ones :)

@lstkz
Copy link

lstkz commented Jan 22, 2017

We had a reviewer rating, and TC staff decided to remove it because it didn't work well.
Copilots and PMs don't have time to review reviewers.

@ThomasKranitsas
Copy link
Author

@lsentkiewicz My idea was to provide feedback (rating) only when there is a reason to give negative rating. Something like giving penalties to reviewers and not require PMs/copilots to review reviewers.

@birdofpreyru
Copy link

@lsentkiewicz
Yeap, in situations where some competitors reach PM for re-appeals anyway, it won't be a big extra work to officially add some "penalty points" to reviewers for their mistakes (i.e. the re-appeals found to be valid).

@lstkz
Copy link

lstkz commented Jan 23, 2017

If the reviewer gets "penalty points" he will try to appeal it, and it will create another discussion copilot vs. reviewer.

TC staff tried to implement a system that payments are related to the review quality. It was probably 6 or 8 years ago. They dropped it and decided to create a simple system.

btw
This is not a new problem. Try to search for old topics in the forums. Use "review", "review quality" as a search query.
Example:
https://apps.topcoder.com/forums/?module=Thread&threadID=842927&start=0&mc=26#1970612

@birdofpreyru
Copy link

@lsentkiewicz

If the reviewer gets "penalty points" he will try to appeal it, and it will create another discussion copilot vs. reviewer.

What about a system, that such problematic cases are collected in some queue, and a selected jury of most experienced copilots/managers/members/reviewers reviews these cases with some periodicity? Their decisions won't have an effect on the outcomes of the contests affected by the cases (it would create a real mess, as the money have been paid already, and, probably, the winning code already have passed further into dev/production), but they will have impact on the reviewer rating, contributing to preventing similar issues in further. This will remove the overhead from specific contests (the copilot will have just file the case into the system), but will (i) range reviewers by the reliability; (ii) create a reference for reviewers; (iii) reduce the appellation hell, as it will be a group decision of reputed guys, so one have to accept it.

What do you think?

@lstkz
Copy link

lstkz commented Jan 23, 2017

a selected jury of most experienced copilots/managers/members/reviewers

They won't work for free, and TC admin don't want to pay anything extra.
They reduced 3 -> 2 reviewers to limit review costs.

@birdofpreyru
Copy link

birdofpreyru commented Jan 23, 2017

They won't work for free, and TC admin don't want to pay anything extra.

Well, you get CAB, and hot discussions in Slack working for free, why won't it work for such jury? :)
Also, budget matters, and whether it is worth spendings is a secondary question, what do you think about the idea itself, irrelevant of the costs?

@lstkz
Copy link

lstkz commented Jan 23, 2017

Reviewing is a time-consuming task :)

@hokienick
Copy link
Contributor

Speaking hypothetically, what about something similar to the lyft/uber system?

The way they operate, riders and drivers rate each other 1-5 stars to ensure good quality. If a driver gets less than 3.5 stars or something they are then banned from being a driver on their app for a period of time. If I rate a driver 5 stars, it is optional for me to write any further comments. If I rate lower than 5 stars I am required to give a reason why. This would make it easier for the Topcoder Team as we wouldn't have to monitor every situation and every reviewer, rather the focus would be on the reviewer(s) that have fallen below the mark and investigate as to why that is and if it is fair.

@ThomasKranitsas
Copy link
Author

@hokienick that's exactly what I had in mind 😄

@lstkz
Copy link

lstkz commented Jan 24, 2017

@hokienick We had almost the same system.
The copilot could rate a reviewer with Bad/Average/Good.
The base rating was 100. Good reviewers had rating >110, but bad reviewers < 90.
Reviewers were assigned based on some algorithm. Good reviewers were assigned more often than bad reviewers, but also there was a random factor.

Copilots didn't have time to analyze the scorecards, and in 90% of all reviews, they assigned 'Good' rate. 10% was re-appeals or very bad review.

@lijulat
Copy link

lijulat commented Jan 24, 2017

I think it should be the submitter who should be given a chance to give feedback . And the co pilot can look at feedbacks that are below a certain threshold.

Ultimately if the competitors think that if the review is fair and good then only people will be motivated to take part in the challenges .

So if it's time vs attracting quality competitors in would choose the latter.

@ThomasKranitsas
Copy link
Author

@lijulat then every submitter who won't be placed 1st/2nd (get paid) will give negative feedback :)

@birdofpreyru
Copy link

@ThomasKranitsas
What you tell, is an universal objection about any feedback-rating system :-D
And it is not quite true, if the process organised in the correct way, like:

  1. The competitor (usually the most interested party when it comes to review errors) raises the complain, if he truly believe, the review was not fair.
  2. Everybody involved in the contest (copilot / reviewers / other competitors) should give their opinion, as they are familiar with the matter.
  3. A jury (kind of conflict commission, selected similarly to CAB) decides who is right in this situation, and scores up/down the reviewer rating according to some system (on the other hand, all successful reviews by reviewer, which did not led to complains, are counted to rise his rating).
  4. If some participant goes to complain each time he does not get a place, and each time it is found the review was right, it will be clearly visible after several such occasions, thus he may be banned from the right to raise the complains.

@hokienick
Copy link
Contributor

Putting @dmessing comment from Slack here:

The idea of the old system was. Good but it did not produce worthwhile results. We need to update the algorithm and who can rate the reviewers

@iversonLv
Copy link

@hokienick and @dmessing In fact I wonder what is 'did not produce worthwhile' results.

@hokienick
Copy link
Contributor

CAB Meeting

  • In the feedback email, there could be a "Do you think the reviewer was fair?"
  • Create some tool to gather rankings of reviewers
  • Reviewers should know what their feedback is

@hokienick
Copy link
Contributor

Cardillo leads Reviewers
Next quarter, they will set up goals for this issue.
Will keep CAB in the loop

@iversonLv
Copy link

Any update or schedule for new rating system is introduced?
1 year past.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

6 participants