[Gluster-devel] Reducing merge conflicts

Joe Julian joe at julianfamily.org
Thu Jul 14 17:59:01 UTC 2016


On 07/07/2016 08:58 PM, Pranith Kumar Karampuri wrote:
>
>
> On Fri, Jul 8, 2016 at 8:40 AM, Jeff Darcy <jdarcy at redhat.com 
> <mailto:jdarcy at redhat.com>> wrote:
>
>     > What gets measured gets managed.
>
>     Exactly.  Reviewing is part of everyone's job, but reviews aren't
>     tracked
>     in any way that matters.  Contrast that with the *enormous*
>     pressure most
>     of us are under to get our own patches in, and it's pretty predictable
>     what will happen.  We need to change that calculation.
>
>
>     > What I have seen at least is that it is easy to find
>     > people who sent patches, how many patches someone sent in a
>     month etc. There
>     > is no easy way to get these numbers for reviews. 'Reviewed-by'
>     tag in commit
>     > only includes the people who did +1/+2 on the final revision of
>     the patch,
>     > which is bad.
>
>     That's a very good point.  I think people people who comment also get
>     Reviewed-by: lines, but it doesn't matter because there's still a
>     whole
>     world of things completely outside of Gerrit.  Reviews done by
>     email won't
>     get counted, nor will consultations in the hallway or on IRC.  I
>     have some
>     ideas who's most active in those ways.  Some (such as yourself)
>     show up in
>     the Reviewed-by: statistics.  Others do not.  In terms of making sure
>     people get all the credit they deserve, those things need to be
>     counted
>     too.  However, in terms of *getting the review queue unstuck* I'm
>     not so
>     sure.  What matters for that is the reviews that Gerrit uses to
>     determine
>     merge eligibility, so I think encouraging that specific kind of review
>     still moves us in a positive direction.
>
>
> In my experience at least it was only adding 'reviewied-by' for the 
> people who gave +1/+2 on the final version of the patch
>
> I agree about encouraging specific kind of review. At the same time we 
> need to make reviewing, helping users in the community as important as 
> sending patches in the eyes of everyone. It is very important to know 
> these statistics to move in the right direction. My main problem with 
> this is, everyone knows that reviews are important, then why are they 
> not happening? Is it really laziness? Are we sure if there are people 
> in the team who are not sharing the burden because of which it is 
> becoming too much for 1 or 2 people to handle the total load? All 
> these things become very easy to reason about if we have this data. 
> Then I am sure we can easily find how best to solve this issue. Same 
> goes for spurious failures. These are not problems that are not faced 
> by others in the world either. I remember watching a video where 
> someone shared (I think it was in google) that they started putting 
> giant TVs in the hall-way in all the offices and the people who don't 
> attend to spurious-build-failure problems would show up on the screen 
> for everyone in the world to see. Apparently the guy with the biggest 
> picture(the one who was not attending to any build failures at all I 
> guess) came to these folks and asked how should he get his picture 
> removed from the screen, and it was solved in a day or two. We don't 
> have to go to those lengths, but we do need data to nudge people in 
> the right direction.
>
>

Perhaps it's imposter syndrome. I know that even when I do leave 
comments on a patch, I don't add a +-1 because I don't think that my 
vote counts. I know I'm not part of the core developers so maybe I'm 
right, I don't know. Maybe some sort of published guidelines or 
mentorship could help?
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-devel/attachments/20160714/b9a95437/attachment.html>


More information about the Gluster-devel mailing list