Last updated: | Permalink
Discussion: Open Source
These notes roughly capture the (mostly unedited) comments from the class:
- Review course requirements:
- Individual Reflection Paper (proposal with reading list: Feb 7, ~2500 word final document: Mar 14, Mar 14 & 16 presentations)
- Group Project (Feb 23 preliminary proposal, Mar 2 final proposal, Mar 28 status update, Apr 18 report, Apr 18 & 20 presentations)
- In-class participation (attendance, completing polls, participating).
- Review course grading:
- Each of the above 3 aspects gets an overall grade: Check-, Check, Check+
- Paper and project each have extremely detailed rubrics
- Participation: Can miss up to two weeks with no penalty
- Earn at least check on all 3: A
- Earn at least check on 2: B
- Earn at least check on 1: C
- +/- are discretionary
- All grades manually computed at end of semester, do not rely on Canvas averages
- Poll and high-level discussion: OSS
- Poll will ask for evaluation on two dimensions: https://pollev.com/jbell
- Significance: Do you think that there is a practical impact of this study? (We will discuss those impacts)
- Soundness: Do you think that the methods are appropriate, and that the results support the conclusions that are drawn?
- Forking paper:
- Significance
- Interviews on collaboration interesting and can have relevance
- Avoiding hard forks is interesting, but not the main focus of the article
- Good contribution for learning about forking, history, implications, etc.
- There has been a significant increase in forking (good), increase in OSS, collaboration (good), increase in hard forks (?)
- Soundness
- How to interpret response rate? What conclusions to draw?
- Heuristics-based classifiers for detecting social forks vs hard forks, hard to tell exactly what the heuristics optimize for, whether it is even possible to make a clear classification
- Significance
- Open source donations
- Significance
- Selection of platforms (can’t make claims beyond what they studied)
- Interesting conclusion that there is not a significant correlation between level of activity and donations
- What is the right question to ask here?
- Over what time period to consider activity?
- Soundness
- Are there strong confounds: projects only ask for money once they are already in a relatively mature phase?
- Using only public artifacts vs interviews/survey
- RQ5: How was the money spent?
- Compares across domains
- Volume of donations, value of donations, amount activity really varies by project AND by domain
- Significance
- Poll will ask for evaluation on two dimensions: https://pollev.com/jbell
- Broader OSS discussion/community issues
- What are the expectations of community members
- Contributors
- Contributions to be accepted?
- Users
- Bug-free?
- Contributors
- What are the expectations of community members
- Forking
- There is a spectrum between hard forks and social forks
- “We started as a social fork, then it became a hard fork”
- “Significant increase in forks” - but short-lived
- How to make forks sustainable?
- What are our own experiences about making a fork?
- Merging back upstream?
- Intentional not to merge upstream: a feature just for us
- “Successful PR” - make the fork, then merge upstream
- “I just had to make a fork of the class project”
- What are the motivations for project maintainers to NOT merging in a PR?
- Supply chain attacks
- Maintenance burden
- Vanity [or maybe something is intentionally snapshotted]
- It is not a feature that we want
- Licensing conflicts
- What are the motivations for project maintainers TO merge them in?
- Diversity the volunteer pool, bring in new talent
- Free bug fixes
- What about a “disruptive innovation?”
- Creativity, community contribution
- Build on top of existing projects to make something new/bigger
- Forks can increase the utilization of the overall project
- That fork might have its own pool of users, contributors, etc. WebRTC
- Methodology:
- Is there a more objective way to determine hard forks, OR: is it OK, OR: should we change the question to be easier to answer?
- Maybe this should be a non-discrete (Not hard/social)?
- Based on a metric based on the flow of changes in both directions
- Maybe the study is interesting without the “hard fork” part?
- (“We found that the stigma around hard forks is mostly gone”)
- Maybe this should be a non-discrete (Not hard/social)?
- Is there a more objective way to determine hard forks, OR: is it OK, OR: should we change the question to be easier to answer?
- Results:
- “Hard forks are not likely to be avoidable”
- Community fragmentation is problematic… could we measure that directly?
- Forks, new projects, projects that just don’t happen
- What about community/governance models?
- It is organic.
- What is GitHub’s role in governance? (Nudging)
- BDFL vs hierarchy
- Donations:
- Did the decision of the authors of the donations to not interview developers cause them to miss out on significant information?
- This is a difficult topic to discuss and get answers on
- This is a possible topic to get valid/good answers on, but it is hard and time consuming to do right
- Other soundness complaints:
- Bitcoin
- Projects supported by a foundation?
- Other support mechanisms (paid version vs free version)
- Other significance complaints:
- We really want to compare self-run donations vs other funding sources
- “How are they spending the money” - this is a philanthropy-rooted question, maybe turn to that research
- What is the level of funding that is needed to sustain a project?
- Bringing funding in to a project may have unexpected outcomes (again, look to philanthropy research)
- Methodology
- Is this “look at open source projects” and then follow-up with interviews a good methodology overall?
- Good, but hard to draw general conclusions
- Can be very limiting based on the selection of data sources.
- OK with just using GitHub?
- OK with just using OpenCollective?
- There are non-open-source collaboratively funded projects that are excluded from such a study (e.g. games)
- Is this “look at open source projects” and then follow-up with interviews a good methodology overall?
- Did the decision of the authors of the donations to not interview developers cause them to miss out on significant information?