AQ 2007 10 31 Discussion

From Earth Science Information Partners (ESIP)

back to 2007-10-31 Workshop page


What is the particular niche for this group?

Comparison to science meeting

  • Was this more like a science meeting?
Can we segment topics so all parties participate in all discussions?
Is there a way to capture all the discussion that happened around presentations?

Special opportunities to build "knowledge base"

  • Journal articles may not capture all available discussion
Could we test tools that would allow us to elaborate and apportion the significance of topics? Maybe we could use a wiki for this.
The comment was made that our measurements are "unstable", which elicited the comment, "Then what are we doing here?" How do we avoid defaming all our data with details? Certainly these satellite differences are significant. How can we state that while questioning the details of our understanding?

Opportunity to reach across communities

  • Clearly an advantage of the ESIP Federation is the opportunity to put technology people together with data people, scientists, educators, and applied science people. How can we take best advantage of this opportunity?
It might be that we could use discussion tools to help create an effective knowledge base that communicates appropriately to different audiences. Policy makers need to see that there is a human effect, not that there is still some question which of the effects are the most significant.

Data/tool Decision Tree

wonder if we could set up table comparing products
might do by datasource, visualization tool, processing tool, etc.

Usage considerations

Legal

  • EPA always has to defend its judgments in court
  • Court needs "preponderance of evidence" or "beyond a reasonable doubt"
This is quite different from 0.95% certainty.

Science

  • Needs detailed information about sources and models used in "correcting" data

Education

  • Special considerations for "real-time" data
  • Special considerations for hiding and introducing complexity

Data lineage tracking

  • Do we need to coordinate conventions for tracking provenance, even in readme files,?
  • How do we track sources and magnitude of variance within and across datasets throughout the processing chain?
    • Take as an example the case x and y variance are not the same.
    • Which processing tools compound error? How do we account for it?

Data quality considerations

  • Can we list spurious sources of variance that must be taken into consideration as we visualize or composite datasets?
    • Cloud cover
    • Registration issues
      • sensitivity to neighboring measurements
    • Instrument issues
      • resolution
        • types of interpixel contamination
      • interference (NOx, chemical, electrical)
      • "viewing" angle (each pixel, terrain issues)
    • Timing issues
      • Time of day
      • Day of week
      • Lunar cycle
      • Special (car exhaust, sporting event, volcanic activity)
    • Magnetic issues (no omi for Rio)
  • What is the best way to determine "best available evidence"?
    • How do we know if a remote sensing product has been verified with ground data and in situations comparable to use?
  • How and why might we tag a dataset "bad data"?

Use Case

How much NO2 is man made?

  • How would we provide data to justify the statement, "Man-made emissions of nitrogen oxides dominate total emissions"?
See http://www.apis.ac.uk/overview/pollutants/overview_NOx.htm

Thessolonika dataset

  • Perhaps useful to discuss formally Ernest Hilsenrath's Thessolonika slide

Solutions

Technology problem

  • Statement was made that this may not be technology problem
  • But we don't have technology for capturing inference or causes
    • we are barely handling ontology when we need epistemology, grammar, rhetoric, and dialectics
  • We don't capture emergent properties
    • Rudy's frames provide slots for lots of the necessary ancillary data
    • Frame-based reasoning is brittle and doesn't handle ambiguity

Social problem

  • How do we get people to use the technology
    • Dedicate a person to capturing ephemera
    • But there is an inestimable volume of such, it would be too expensive
  • Question may be, "How do we elevate those few really significant details?"
    • What tools would facilitate this? Easy way to raise "level 3" objections or warnings
  • Different communities don't know enough to ask each other for the "right" things
    • Need low level interaction to raise common ground

Systemic problem

  • Are there appropriate rewards for investment of time in metadata?
  • Is a non-rhetorical assumption valid?
    • There are always reasons to "sell" aspects of what was done
    • Everyone has a bias; courts require id for advocacy; should science?
  • Are the right people at the table?
    • ESIP Federation was established to assure that the right mix of experts were involved in data decisions.
    • Can we expedite reviews using "federalism" (balanced interests)?