|
|||
|
This is version 11.
It is not the current version, and thus it cannot be edited. Terrestrial usecasesThe following are the Usecases the terrestrial working group discussed during the July 2007 REAP Requirements Workshop. The group also conducted some informal ranking of these. First, usescases were given a 1-3 importance ranking, with 1 being most important. Second, of those cases ranked 1, group members voted for a few they felt were most important or should be implemented first.
2)Importance: 2 Use Case Name: Smart Reruns Version: Preconditions: User has designed a workflow that specifies:
Users want the ability to rerun certain workflows in a "smart" way, that is, without regenerating intermediary derived data products if the source data (and hence the derived data for non-stochastic models) have not changed. Some intermediate derived data products take a very long time to produce, hence it is beneficial to skip their generation when possible.
User executes a workflow that uses Smart Rerun features and all Smart Rerun criteria have been met.
Workflow developer designs a workflow that makes use of Smart Rerun components. Upon execution of this workflow, Smart Rerun components check to see if their input data sources have changed (possibly the user can also specify conditions which must also be met). If these data sources have changed (or are being used for the first time), components (re)produce derived data products and save those that are marked as Smart Rerun derived data products. If data sources have not changed, and previously saved derived data products exist, a Smart Rerun can occur reusing these existing products. Special considerations for stochastic systems must be taken into account in the design of this usecase (do we allow Smart Reruns for certain stochastic models?).
Intermediary derived data products produced by Smart Rerun actors are saved.
3)Importance: 2 Use Case Name: Optional archiving of intermediate derived datasets.
Users want to be able to save intermediate derived datasets.
4)Importance: 1, votes: 5 Use Case Name: Archive a complete end to end run ("publication ready archive") Version: Summary: Users would like the ability to create a "publication ready archive"--an archive that contains the Kepler workflow, all required inputs (data sources like database tables are included), and all produced output. For example, at the completion of a study, a user can produce this archive and submit it (or a link to it) alongside their paper to peers and journals.
User exports workflow
User creates a workflow. User turns on "publication ready archive" mode. User runs workflow, and exports to an archive at completion.
User uses features from "Smart Rerun", "Manage a group of model runs", and "save intermediate derived data products" usecase (ie a special "publication ready archive" mode might not be required).
A specific "publication ready archive" option would not be required if requirements for usecases mentioned in Alt Paths are implemented.
5)Importance: 2 Use Case Name: Database actor option: Snapshot Version: Summary: Users would like the ability to use a database (or more likely simply some data from a table in a database) in two ways.
Workflow containing database actor is run.
If Snapshot feature is used, data snapshot is saved.
6)Importance: 2 Use Case Name: Actor specific annotations (gui linked) Version: Summary: Users would like to be able to produce annotations that are visually closely coupled with actors. When an actor is dragged around the campus, if it has an associated annotation, this annotation moves alongside it. This does not require a graphical link (eg line or box) that groups the two, but this could be done.
A component has been created with which to associate an annotation.
User creates a component and chooses to create an associated annotation for it.
Another way to approach this would be mouseover "alt text" popups.
7)Importance: 1, votes: 2 Use Case Name: Actor for Data-Merge for specific product Version: Summary: Users want the ability to easily merge many different data sources into one. For example, in the REAP terrestrial case study users might want to merge a) data in differently formatted excel and text files b) data from near-realtime streams and c) data pulled from an archive. Users want this process to convert units if needs be.
An instance of data-merge component has been created with inputs connected to desired data sources.
User executes workflow.
User creates an instance of data-merge component and connects its input ports to desired data sources. User specifies certain properties in the data-merge actor regarding how to merge data (eg how to convert units, if interpolation or decimation should occur, which differently names fields correspond, etc). Depending on how complex the data-merge actor becomes, it may have to require user interaction.
User merges data themselves :)
Merged derived source data product is produced--and potentially saved.
8)Importance: 2 Use Case Name: Integrate database actors Version: Summary: make database actors more consistent. DatabaseQuery actor has different output ports than metacat database actors -- users would like these actors to be more consistent. Users would also like to be able to view a database and its tables with this integrated actor.
9)Importance: 1 Use Case Name: Event Notification Monitoring system Version: Summary: Users would like to be able to design workflows that send out alerts when certain criteria are met. Users would typically like these kinds of workflows to run automatically and periodically on a cron or cron-like system that easy to setup. Examples:
12)Importance: 1, votes: 1 Use Case Name: Manage a group of model runs Version: Summary: Users would like to be able to design a workflow that allows for the management a group of model runs. This means:
This might be accomplished with a parameter sweep actor.
User develops workflow that includes a Parameter Sweep component. The input port(s) of this component are connected to parameters or parameter source(s) like a text file.
User designs a workflow that uses and changes (existing) Parameter components.
13) Importance: 1, votes: 1 Use Case Name: Workflows need to accommodate sub-workflows that operate at different frequencies
14)Importance: 1, votes: 4 Use Case Name: Tutorial or Wizard for wrapping external programs Version: Summary: Users would like to easily wrap an external program in Kepler. By "wrap" we mean create a simple workflow that consists of being able to run an external program that has already been developed, eg a C or Fortran program. Users would like either or both of the following:
Working external program has already been written.
User runs wizard or user opens tutorial.
User executes wizard inside Kepler, points it at their external program, and configures any required settings via the wizard.
User reads tutorial that explains how to "wrap" external programs and builds the workflow.
15)Importance: 1 Use Case Name: Easily create and share workflows and actors Version: Summary: Users want an easy way to create and share workflows with others. This means making sure exporting, emailing, and importing workflows is painless, or the submission of workflows/actors to a repository that others can import from is painless.
17)Importance: 3 Use Case Name: facilitate peer review process
18)Importance: 3 Use Case Name: Use Kepler as a teaching tool Version: Summary: Users would like to use Kepler as a teaching tool. This means adding features such as:
19)Importance: 3 Use Case Name: Ecological Actors Version: Summary: Users would like a set of components to do common ecological functions (ie mean or sum...) over temporal and spatial scales
If ecologists want to create some of these: improvements to workflow and component submission to repository process have been made.
User makes use of new ecological components in library as they would any other.
20)Importance: 2 Use Case Name: FakeData Generator Actor(s) Version: Summary: Users would like a set of FakeData Generator Actors appropriate to the REAP Case Studies. A FakeData actor should produce data in the same way a real data source would. For example, a "FakeData_CR800_Temperature" Actor might produce a datastream of temperature data at 1 "sample" per minute. The actual data in the datastream should be indecipherable from a real temperature stream coming out of a Campbell CR800 datalogger deployed in the field. Somewhere in the metadata however, a FakeData Generator Actor should clearly identify itself (as being fake).
Depending on how exactly we wish to simulate certain datalogger datastreams, these specs need to be available during the development of this/these actor(s).
21)Importance: 3 Use Case Name: 3d plots Version: Summary: Users would like to be able to produce 3d plots with Kepler. (ggobi, gnuplot, java3d ?)
3d visualization software is installed on user's machine.
Use Case Name: Iterator wrapper actor Version: Summary: Users would like an Iterator Actor that iterates a workflow, or parts of a workflow in a smart way. See Paramater Sweep Usecase.
22)Importance: 3 Use Case Name: Markov Chain Monte Carlo (mcmc) Version: Summary: Allow this different technique for running models. Can not use if model is stochastic.
An understanding of mcmc. :)
23)Importance: 2 Use Case Name: Breakpoints Version: Summary: Users would like to be able to add breakpoints in a workflow. When a breakpoint is reached, workflow execution is paused or stopped. Users would also like breakpoints to be conditional, eg a breakpoint might only pause or stop execution once a certain condition (like equilibrium) has been reached.
24)Importance: 2 Use Case Name: Debugging Version: Summary: Users would like an easy way to debug workflows. Some ideas on this: a special debug output port on all actors; or a specific debug actor (most workflow developers make extensive use of the display actor to verify output after each step as they build workflows--new users might not think of this, an actor named or tagged Debug might be of use)
|
This material is based upon work supported by the National Science Foundation under award 0619060. Any opinions, findings and conclusions or recomendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation (NSF). Copyright 2007 |