Project Evaluation Activity V1

From Foss2Serve
Revision as of 11:39, 8 September 2018 by Clif.kussmaul (Talk | contribs)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search


Title

Project Evaluation Activity V1

Overview

Learners will gain an understanding of the breadth of available FOSS projects. Learners will also gain an understanding of the identifying characteristics of FOSS projects including pattern of contributions, patterns of commits, programming languages used, and more.

Prerequisites

Completion of Browsing a Forge Activity or understanding of SourceForge and Ohloh; Understanding of course in which students will be participating in an HFOSS project.

Learning
Objectives
After successfully completing this activity, the learner should be able to:
  • Utilize the rubric to identify likely HFOSS projects.
Process Skills
Practiced


Background

This activity is intended to give you an overview of what to consider when evaluating a FOSS project for student participation and for you to gain experience using the rubric.

Directions

Part 1-Learn about the rubric

  1. Watch the video describing mission critical criteria
  2. Watch the video describing secondary criteria

Part 2-Walk through of an evaluation of the Mifos project

Mission Critical criteria-Viability
  1. Size/Scale/Complexity - An ideal project should be neither overly simple nor overly complex. One heuristic to use is the number of contributors as an indicator of project complexity.
    1. Go to Ohloh.net, type Mifos into the Search Projects box.
    2. On the results page click on Mifos (above the Mifos logo) to see the Project Summary page.
    3. Scroll down to the Community area and calculate the average number of contributors in the last 12 months -- note that the graph is interactive. The average was 9 so it passed the minimum average number of contributors metric of 6.
    4. Go to the Mifos web page (http://www.mifos.org) and choose Tech Overview from the Contributors tab. From examination of the technology stack, the architecture looks modular and further search shows it is documented elsewhere on the site.
    5. Result: Based on the modular design and meeting the minimum average number of contributors metric, the project is scored a 2 for size/scale/complexity.
  2. Activity - To support student participation a project should be reasonably active. Number of committers can be used as an indicator of activity.
    1. Return to the Mifos project summary page in Ohloh. Scroll to the Activity area on the page.
    2. Compute the 12-month average of commits. The 12-month average was about 108, much higher than the recommended range of 10 - 30 commits per month.
    3. Result: Because the number of commits for this project exceeds the favorable level of activity, it may be a little large/complex. However, it still appears manageable, the project is scored a 2 for activity.
  3. Community - A suitable project has an active user community. While it is difficult to quantitatively evaluate the activity of a user community, some indicators include a regular history of project downloads and documentation updates over time, current activity on user mailing lists, and testimonials on the project web site.
    1. Examine download activity
      1. Go to Sourceforge.net and enter Mifos into the search box.
      2. Choose Mifos-Microfinance Open Source from the search results.
      3. Click on the number of downloads that is listed on the project page.
      4. Change the date range to give a graph of downloads over the last year.
    2. Examine user mailing list activity
    3. Examine the IRC logs
    4. Result: Downloads appear steady so the project has a community of users. Developers are responsive on mailing list and have a presence on IRC. Project is scored a 3.


Mission Critical criteria-Approachability
Here you are evaluating a project's on-ramp to contribution, scoring as follows:
1-Insufficient-Few or no pointers on how to become involved.
2-Sufficient-Suggestions about how to get involved other than contributing money with accompanying high-level instructions.
3-Ideal-Obvious link to get started, list of suggestions for things to do and detailed instructions.
  1. Link to get started-There is a Get Started page with links to what Mifos is, how to contribute, community processes, and tools used.
  2. List of suggestions for things to do - The Volunteer Project page provides a list of ways to contribute including testing, translation, development and documentation. There is also a volunteer bug queue listed as a good way for developers to get started.
  3. Detailed instructions- On the web site instructions and information are provided in many areas including process, architecture, licensing, product functionality, and developer documentation.
  4. Result-Was scored a 3.


Mission Critical criteria-Suitability
  1. Appropriate Artifacts -Since evaluation is dependent on class objectives, in this example we'll assume an objective is to learn the process of working in authentic development project through contributing bug fixes.
    1. Opportunities to contribute bug fixes - Examined the volunteer bite-sized bug queue. There were 10 open bugs for new contributors. There were many more listed for more experienced contributors.
    2. Documentation on how to contribute bug fixes - From the Tech Overview page there are links to details on the code submission process.
    3. Result - May score a 1, 2 or 3 depending on the number of bugs suitable for students to tackle and class size.
  2. Contributor Support-Does the project have a high volume of guidance to help students as they learn?
    1. Are communication tools documented?-Communication tools are documented under the Collaboration and Communication section of the Development Tools page. Instructions on how to access the mailing lists with tips on how to participate are available from the Communications page.
    2. Do developers have a web presence?-Examination of IRC logs shows scattered activity over the last week.
    3. Are operating processed documented?-Links to information about coding standards, code submission process, and commit privileges process can be found on the Tech Overview page. The process for making feature requests and for prioritizing feature request is available on the Roadmap page.
    4. Do questions posed have timely and supportive answers?-Responses to user mailing list and developer mailing list over the last month have timely and supportive responses.
    5. Result - Not a lot of activity on IRC, but mailing lists show lots of timely feedback and communication methods and operating procedures are well documented, score a 3.


Overall evaluation for Mission Critical criteria - Since no mission-critical criteria were scored lower than a 2 the project is then evaluated on secondary criteria. Otherwise, the project would have been considered not suitable for student participation.


Secondary criteria-Viability
  1. Domain
    1. Does this project require domain knowledge that may be difficult for students to learn? - As a domain microfinance students should be able to grasp it well enough to contribute a bug fix, which is the learning objective assumed in this example.
    2. Result - Score a 2 since the domain isn’t as simple to grasp as say a desktop application for word processing or compressing files.
  2. Maturity
    1. To have the organization to support student learning, the project should have at least one stable production release - The roadmap page lists releases.
    2. Result - The Download Mifos page says 2.6.0 is the 4th major community-supported release. Scored a 3.
  3. User Support
    1. The project should have clear instructions for downloading, installing, and using the project - There is a demo server, video, and slide presentation that explains system functionality. This information can be found looking at pages listed under the Product tab that can be used to learn about the system. There is also a user manual available. On the Download Mifos page, there are detailed instructions related to installation, configuration, system requirements, and troubleshooting.
    2. Result - Given the wealth of detailed documentation, score a 3.
  4. Roadmap
    1. Student learning is best supported by projects that have a roadmap that includes new feature development, a method for users to submit new feature requests and a process for identifying how new features are prioritized - The process for making feature requests and for prioritizing feature request is available on the Roadmap page. The roadmap has features that were implemented in the last release, but no listing for the next release.
    2. Result - Scored a 2 because there is no information listed for feature planning in the next release.


Secondary criteria-Approachability
  1. Contribution Types
    1. Does the project contain opportunities for multiple types of contribution and of the type that fits the class? - There are multiple projects for testers, tech writers, and developers. These can be seen on the Volunteer Projects page.
    2. Result - May be a 1, 2 or 3 depending on whether the number of bugs is suitable for students is enough given the class size.
  2. Openness to Contributions
    1. Acceptance of a student contribution to a project provides valuable affirmation to student learning. Determine whether the project accepts student patches. - The process for contribution is documented on the Tech Overview page.
    2. Result - Score a 3 because the contribution process is documented.
  3. Student Friendliness
    1. Do community members moderate the tone of communication? Review mailing lists and IRC to gauge tone - Review of user mailing list and developer mailing list and IRC logs during evaluation of contributor support showed a positive tone during communication.
    2. Result - Score a 3, no inappropriate or demeaning messages.


Secondary criteria-Suitability
  1. Project Description
    1. Students must be able to understand the purpose of the project. Does the project clearly describe the product? Can students understand the intended uses of the product? - The About tab on the web page has links to the vision for the product and how it is used by microfinance institutions.
    2. Result - Score a 3, how the product is used and the vision for it is well documented and should be understandable by students.
  2. Platform
    1. What software and hardware platform does the FOSS project run on? - Development environment can be built on Windows, Ubuntu or Mac desktop completely with FOSS software. (Project development information found here)
    2. Are there resources to support these platforms? - In this example, yes.
    3. Are students familiar with the platforms? - In this example, yes.
    4. Result - Score a 2, assumption in this example is students all have newer personal computers and given the ability to set up a development environment on different operating systems that makes the availability of student resources greater. However, there is some risk because machine requirements for setting up developer environment are not provided and some documentation may be out of date.
  3. Development Features - Is the class dependent on specific development features? (Project development information found here)
    1. Programming language - Is primarily Java.
    2. Development environment - Can be built on Windows, Ubuntu or Mac completely with FOSS software.
    3. Supporting technologies - Suggested IDE is Eclipse, requires Maven, Jetty, and mySQL.
    4. Result - Need to gauge this on knowledge of students and requirements of class. Assumption here is students know Java and are familiar with mySQL. While students are not familiar with Maven and Jetty this may not be necessary for intro bug fix plus the community is very supportive so assistance can be found there. Given there is some risk, score a 2.


Overall evaluation for secondary criteria - Total score for criteria is over 20, the project passes. However, criteria scoring below 1 and criteria for which there was some risk noted should be reexamined to see if steps can be taken to mitigate risk.

Part 3-Evaluate a project

  1. Choose a project to evaluate from the project list project list
  2. Use the blank evaluation template to help record your results.
  3. Upload your evaluation.
  4. Add a link on your user page to the evaluation you uploaded.

Deliverables

Wiki posting of evaluation of a project from the list of HFOSS projects.


Assessment:

  • How will the activity be graded?
  • How will learning will be measured?
  • Include sample assessment questions/rubrics.
Criteria Level 1 (fail) Level 2 (pass) Level 3 (good) Level 4 (exceptional)
The purpose of the project
Why the project is open source

Comments

  • What should the instructor know before using this activity?
  • What are some likely difficulties that an instructor may encounter using this activity?


Additional Information:

ACM BoK
Area & Unit(s)
ACM BoK
Topic(s)
Difficulty
Estimated Time
to Complete

60-90 minutes

Environment /
Materials

Access to Internet/Web and web browser, SIGCSE paper on evaluating FOSS projects, Blank evaluation template

Author(s)
Source
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License

CC license.png


Suggestions for Open Source Community:

  • Suggestions for an open source community member who is working in conjunction with the instructor.
Personal tools
Namespaces
Variants
Actions
Events
Learning Resources
HFOSS Projects
Evaluation
Navigation
Toolbox