I agree to Idea Store results in the cloud with workflow available for reuse
Voting Disabled

37 votes

I disagree to Idea Store results in the cloud with workflow available for reuse

Rank1

Idea#13

This idea is active.
Analysis Priorities »

Store results in the cloud with workflow available for reuse

Store results in the cloud (e.g. variant file) with methodologies documented and workflow available in a workflow management system so the analyses can be reproduced with other data

Submitted by 1 year ago

Comments (4)

  1. One of the clear benefits of the cloud data hosting model is that all data sets of a similar type can be analyzed consistently, something which individuals have to do themselves now.

    So the example of having variant files generated with a documented methodology is a good one, as long one can get variant files for all samples produced with that same methodology.

    This speaks to a broad issue of how much top-down control there will be on the cloud to ensure consistency and standards on the data. We'll have to strike the right balance here to both encourage participation and enable comprehensive analysis.

    1 year ago
    0 Agreed
    0 Disagreed
  2. I agree that cloud arrangements with data and algorithms are an important advance. I think it is also important to recognize the value of metadata. Including systems to coordinate metadata with the primary data, and also the algorithms is essential.

    Others have pointed out that technologies like iPython notebooks that show code and RESULTS are valuable because it lets you have the complete package.

    1 year ago
    0 Agreed
    0 Disagreed
  3. Storing results in the cloud will be most useful of the results and their interpretations are represented using standard semantic terminologies. These will also need to be defined and systems developed to facilitate their use.

    1 year ago
    0 Agreed
    0 Disagreed
  4. In conjunction we might also cultivate online dashboards for reproducibility, where results might be evaluated according to well-defined metrics to derive some form of plausibility score. This may not be straightforward for wet results, but seems well within reach for computational/algorithmic results. Imagine the time & effort saved if investigators could visit something like Angie's List for science, allowing one to ascertain at a glance which results show the greatest promise for replication. More details are in a draft proposal available from the Broad Institute at https://docs.google.com/document/d/1Ts1LAMyv9j3sb8F3UX87Gd1tfaBTQht08QQNtNgsjVQ/edit?usp=sharing

    1 year ago
    0 Agreed
    0 Disagreed

Vote Activity Show

(latest 20 votes)

Events

  1. The idea was posted
    1 year ago