This is version 18.
It is not the current version, and thus it cannot be edited.
[Back to current version]
[Restore this version]
- Refactor PAR wf into two wfs (data acquisition, data analysis)
- Complete PAR data acquisition wf: reorganize par-data-clean step, replace with generic actors (Elementwise: min, max, is.na; UnivariateDistribution: min, max, range, interquartile range, mean, variance, median, stddev, historgram, densityplot, autocorrelation, partial autocorrelation; bivariate: correlation; statistical model fit: lm, glm, gam; filter actors) (Derik: top priority)
- Complete PAR data analysis workflow (Derik, Eric, Parviez)
- Parviez' workflow model for wrapping a simulation using CommandLineActor -- review annotations and resolve issues (Derik, Daniel)
- Terr wf -- need to select data based on spatial and temporal bounds
- Parameter sweep example workflow (define reqs of param sweep actor, implement actor after checking existing actors and directors) (Ilkay, Chad, Lucas, Parviez)
- Review user manual section on Directors and Command Line Exec actor (section 5.5) (Parviez, Daniel)
- Example wf pulling data from SensorBase at CENS using WebService actor or DBQuery actor (Eric G, Derik)
- Take loggernet out of the loop in the hardware demo (buy logger for testing) (Derik, Eric G)
- New workflow to create derived data for Met data (e.g., wind chill, dewpoint, etc) (Eric G, Derik)
- Determine how to archive data (e.g., EarthGrid w/ metadata) (Derik)
- Generate list of Level 3 SST data URLs (some with with global coverage) to be used in prototyping workflow (Ilkay, Nathan, Peter)
- Generate details of computational codes to be used to process the data (flowcharts+?)
- Layman version (without technical and scientific buzwords) of the description of the goals for each step and workflow(couple of paragraphs) (April 4th; 1 week) (Ilkay + Mark)
- Peter will write up his realization of why this is useful on top of just matlab with a few bullets(April 4th; 1 week) (provenance, chaining what is available, ...) (Peter)
- Conceptual workflow with input/output details and placeholders for different computational components (April 9th; 2 weeks) (Ilkay and Dan)
- Design the schema ad build a relational database for the match up datasets (April 11th; 2 weeks) (Dan and Ilkay)
- Recollection of requirements and comments (April 17th; 3 weeks) (Ilkay and everyone on SST)
- Working version0 (May 15th; 1.5 months - 4 weeks for implementation) (Ilkay and Dan)
- Definition of additional analysis steps (June 1st; 2 months) (Ilkay with input from Peter)
- GHRSST Meeting (June 9-13, 2008; France) presentation and demonstration (June 9th; 2 months) (Peter)
- Develop a mechanism for choosing a time and space bounded region (Coverage object) and subsetting data using that coverage specification; may be a sampling actor and/or search (when: April 30 who: )
- Evaluate map display clients for inclusion in search GUI and GIS display
- Develop a GUI for specifying space/time coverage constraints that returns a Coverage object that can be passed to servers for querying. The GUI would include widgets for selecting (possibly discontiguous) space and time windows, possibly providing real-time feedback about number of results
- Document OPeNDAP actor, add to tree, remove ODC feature, assign icon, test for release (Nathan (with Kirsten)), (Next week)) COMPLETED
- Matlab actor -- prepare for release, get library loading working, remove MatlabExpression version of actor (cretae bug to merge this as a fallback to JNI), make actor use TextArea for script, check documentation, complete test workflows (Daniel, with help from Kirsten, by mid April) (See http://bugzilla.ecoinformatics.org/show_bug.cgi?id=3208)
- Change ProvenanceRecorder to use local hsql db, turned on by default?, possibly output to text format on demand (Daniel, by mid April for Kepler release)
- Combine EML and DAP DataSource actors
- Create data location actor (finds data and spits out URL)
- Add server registration features to OPeNDAP server
- Add crawler/metadata indexing to OPeNDAP server
- Pass data by-reference in addition to by-value
- DataTurbine actor: PortParam bug, change name, gap filling feature (probably separate actor if doing interpolation), only output data on connected ports, assign icon
- CSV file reader (check name and documentation (tab, comma, csv should all return actor), ignore headers and footers (check the existing R Read.Table actor)
- Add DateTime token type
- Add DataFrame token type
- Remove External Execution actor from actor list, docs, and maybe code base, and/or maybe consolidate CommandLineExec with External Execution (Daniel) (See http://bugzilla.ecoinformatics.org/show_bug.cgi?id=3207)
- Modify Command Line Exec actor to trap errors so that workflow can continue after error or exception or core dump (Daniel)
- Check for bugs in CommandLineExec actor with respect to port names display and copy/paste and cloning operations (Daniel)
- Change exit code behavior of Command Line Exec actor (return code from unerlying app, add port for boolean success value (Daniel)
- Develop common access interface for DataTurbine, SensorBase, and Antelope (Derik, Eric G/CENS,Matt,Ilkay)
- Initiate discussion between CENS/REAP/DataTurbine about commonalities in goals, approach, and interface (Eric G)
|