Difference between revisions of "Output from Visualization Summit"

From Earth Science Information Partners (ESIP)
(Created page with "* RESPONSES TO BIG QUESTIONS (edited) ===================================== still in progress Overall questions Question 1: Imagine the perfect earth...")
 
Line 1: Line 1:
* RESPONSES TO BIG QUESTIONS (edited)
+
<h1>* RESPONSES TO BIG QUESTIONS (edited)</h1>
=====================================
 
  
still in progress
+
<div class="writeboardbody">
 
+
  <p><em>still in progress</em></p>
                          Overall questions
+
<h1>Overall questions</h1>
 
+
<h2>Question 1: Imagine the perfect earth data remote sensing visualization tool/system. What are the main components of this system?</h2>
Question 1: Imagine the perfect earth data remote sensing visualization
+
<ul>
tool/system. What are the main components of this system?
+
<li>Needs to closely track needs and abilities of a variety of audiences</li>
 
+
<li>Support multiple GUIs (meaning <span class="caps">API</span>-based back-end to support tools?)</li>
  * Needs to closely track needs and abilities of a variety of
+
<li>Open Source</li>
    audiences
+
<li>Solves data format/standards problem &#8212; new data type? derivative data types?</li>
  * Support multiple GUIs (meaning API-based back-end to support
+
<li>Social web, community-based: support collaborative workflow and data discovery</li>
    tools?)
+
<li>One-stop repository to support aggregate querying &#8212; e.g., What is the causality relationship between meningitis outbreaks and precipitation patterns in sub-Saharan Africa?</li>
  * Open Source
+
<li>Provide ability to store and track provenance</li>
  * Solves data format/standards problem -- new data type? derivative
+
</ul>
    data types?
+
<h2>Question 2: Visualization for understanding vs. conveying &#8212; It&#8217;s all about the user: How should tools and visualizations be tailored for distinct user groups?</h2>
  * Social web, community-based: support collaborative workflow and
+
<p>Two different issues:</p>
    data discovery
+
<ul>
  * One-stop repository to support aggregate querying -- e.g., What is
+
<li>Conveying &#8212; you know the answer, need to display it in a good way. &#8220;Static&#8221;</li>
    the causality relationship between meningitis outbreaks and
+
<li>Understanding &#8212; trying to find the answer via exploration. Needs more facilities. &#8220;Interactive&#8221;</li>
    precipitation patterns in sub-Saharan Africa?
+
</ul>
  * Provide ability to store and track provenance
+
<p>Hard to build a general-purpose tool, to satisfy power users vs. non-power users, cross-discipline. Same tool needs to scale with abilities.</p>
 
+
<h2>Question 3: Are there user groups that you know are under-served by the current data visualization technology? What needs to change to serve these groups?</h2>
Question 2: Visualization for understanding vs. conveying -- It's all about
+
<ul>
the user: How should tools and visualizations be tailored for distinct user
+
<li>Everyone is underserved, but not equally.</li>
groups?
+
<li>PIs/those closest to data are best served by way of their familiarity with the data.</li>
 
+
<li>Solving this problem is an argument in favor of plug-in based development.</li>
Two different issues:
+
<li>Tools need to be developed to support each audience &#8212; different audiences have wildly different needs in terms of capabilities and end result.</li>
 
+
<li>Of course, serving user groups only comes after they have located the data.</li>
  * Conveying -- you know the answer, need to display it in a good
+
</ul>
    way. &quot;Static&quot;
+
<h2>Question 4: Delivering data vs. images of data. What are the sweet spots for each? What are the areas where we need to focus or change?</h2>
  * Understanding -- trying to find the answer via exploration. Needs
+
<ul>
    more facilities. &quot;Interactive&quot;
+
<li>It depends on audience: images <strong>and</strong> data can serve both &#8220;public&#8221; and scientific audiences<br />
 
+
- In general images are great for the general public or less interested specialists (maybe 80%), data needs to be available for the interested 20%.<br />
Hard to build a general-purpose tool, to satisfy power users vs.
+
- The <a href="http://www.ifp.illinois.edu/nabhcs/abstracts/shneiderman.html">Ben Shneiderman UI design mantra</a> &#8220;overview, zoom &amp; filter, details on demand&#8221; applies in this case.</li>
non-power users, cross-discipline. Same tool needs to scale with
+
</ul>
abilities.
+
<p>Images sweet spots</p>
 
+
<ul>
Question 3: Are there user groups that you know are under-served by the
+
<li>As a means of discovery and monitoring production</li>
current data visualization technology? What needs to change to serve these
+
<li>Able to meet needs of wider audience [formats and general ease of use]</li>
groups?
+
</ul>
 
+
<p>Data sweet spots</p>
  * Everyone is underserved, but not equally.
+
<ul>
  * PIs/those closest to data are best served by way of their
+
<li>Scientist users who create their own visualizations</li>
    familiarity with the data.
+
<li>maintaining metadata (provenance)</li>
  * Solving this problem is an argument in favor of plug-in based
+
<li>can be used to derive multiple visualizations</li>
    development.
+
</ul>
  * Tools need to be developed to support each audience -- different
+
<h2>Question 5: How can visualizers bridge the science-outreach divide? How to teach the public about science and the scientists about the public?</h2>
    audiences have wildly different needs in terms of capabilities and
+
<p>Visualization is a compelling medium that science communicators can use to make complex scientific ideas approachable to a broad audience. Carl Sagan’s Cosmos series is the prototypical example, weaving visuals with narrative to explain astrophysics. It is crucial to define an audience: there is no such thing as a uniform general public. Concrete visualizations, such as planet walks and painted lines representing sea level rise, can be particularly effective. Effective visualization requires focus: emphasize important elements of a dataset, and de-emphasize or eliminate less important data. Ideally, tools would be designed by visualizers, not computer scientists. Try for verisimilitude: make things appear how the audience expects it to appear (for example Google Earth’s discontinuous boundaries between scenes are very distracting).</p>
    end result.
+
<p>(T1)<br />
  * Of course, serving user groups only comes after they have located
+
Perhaps current outreach programs are too top-down, we need a tighter, iterative relationship between viz developers, scientists and the outreach audience. We didn’t like the notion that the public can’t understand science, and consider the question: aren’t scientist part of the public? It’s important to bring scientists and their work into the community, can effective visualization facilitate the needed two-way motivation needed between scientists the community they serve?</p>
    the data.
+
<p>Features for tools, easy to read, embedded description with the displays. Simple initial presentation, but allows a progressive disclosure of information and concepts as far as the user desires.</p>
 
+
<p>(T2)</p>
Question 4: Delivering data vs. images of data. What are the sweet spots for
+
<ul>
each? What are the areas where we need to focus or change?
+
<li>Tune the visualization to the audience &#8212; but how?<br />
 
+
      &#8211; use of toolkits, plug-ins, component frameworks<br />
  * It depends on audience: images and data can serve both &quot;public&quot;
+
      &#8211; providing different presentations of the same data</li>
    and scientific audiences
+
</ul>
    - In general images are great for the general public or less
+
<ul>
    interested specialists (maybe 80%), data needs to be available for
+
<li>Provide <span class="caps">HTML</span>/Flash/<span class="caps">CMS</span> for data<br />
    the interested 20%.
+
      &#8211; easier mechanism for scientists/educators to convey info<br />
    - The Ben Shneiderman UI design mantra &quot;overview, zoom &amp; filter,
+
      &#8211; equivalent of wiki or “build your own web page” for general public</li>
    details on demand&quot; applies in this case.
+
</ul>
 
+
<p>Part B, How to teach the public about science and the scientists about the<br />
Images sweet spots
+
public?</p>
 
+
<ul>
  * As a means of discovery and monitoring production
+
<li>Force scientists through <span class="caps">NSF</span>, etc, new requirements on publishing</li>
  * Able to meet needs of wider audience [formats and general ease of
+
<li>Use emerging technologies like social media</li>
    use]
+
<li>Enable new ways of publishing<br />
 
+
      – e.g. Data <span class="caps">CMS</span>, like <span class="caps">RAMADDA</span></li>
Data sweet spots
+
</ul>
 
+
<p>General comments:</p>
  * Scientist users who create their own visualizations
+
<ul>
  * maintaining metadata (provenance)
+
<li>Visualization is conveying information</li>
  * can be used to derive multiple visualizations
+
<li>Teacher knows the answer, needs to find way to convey</li>
 
+
<li>Scientist does not know the answer, needs exploratory analysis</li>
Question 5: How can visualizers bridge the science-outreach divide? How to
+
<li>There is a declining importance of traditional journal publications</li>
teach the public about science and the scientists about the public?
+
</ul>
 
+
<h1>Current Issues</h1>
Visualization is a compelling medium that science communicators can
+
<h2>Question 1C: You tackle visualization tasks every day. What is the one thing that you do every day that needs to be different in 5 years for your life to improve?</h2>
use to make complex scientific ideas approachable to a broad audience.
+
<ul>
Carl Sagan’s Cosmos series is the prototypical example, weaving
+
<li>data search and retrieval</li>
visuals with narrative to explain astrophysics. It is crucial to
+
<li>one data format to rule them all &#8212; standards-based in all aspects (data structure, metadata)</li>
define an audience: there is no such thing as a uniform general
+
</ul>
public. Concrete visualizations, such as planet walks and painted
+
<h2>Question 2C: Name the three (or more) top obstacles today that limit the effective development/deployment of earth data visualizations.</h2>
lines representing sea level rise, can be particularly effective.
+
<h2>Question 3C: State of gridded data fusion capabilities: what are the main obstacles/opportunities?</h2>
Effective visualization requires focus: emphasize important elements
+
<p>(T1)<br />
of a dataset, and de-emphasize or eliminate less important data.
+
Error tracking and error propagation during data fusion has to be done and provided to the user. For example when resampling data for fusion we need to keep track of the errors and how it affects the grid fusion,</p>
Ideally, tools would be designed by visualizers, not computer
+
<p>Knowledge embedded in the tool to facilitate data fusion. Sort of an knowledge based system that recommends to the user how the data fusion has to be done.</p>
scientists. Try for verisimilitude: make things appear how the
+
<p>Provide raw data,processed re-gridded data and software to data fusion packaged together. Users have all the pieces needed for data fusion.</p>
audience expects it to appear (for example Google Earth’s
+
<p>Algorithm/Tools/Gridding are data dependent.What techniques are used are dependent on the data brought together.</p>
discontinuous boundaries between scenes are very distracting).
+
<p>Embed data units,scaling, offsets within the data to facilitate fusion. We have data in different units, for example we have temperature data in Fahrenheit and in Celsius. Scaling/ data unit conversion has to be embedded in the data or interpreted readily by the tool.</p>
 
+
<p>Strengthen interaction between standards for data fusion. OpenDAP to <span class="caps">OGC</span> to web services.</p>
(T1)
+
<p>(T2)<br />
Perhaps current outreach programs are too top-down, we need a tighter,
+
How do we get the grid itself – some domains end up creating gridded data from point observation data. How does the different algorithms and parametrizations affect the outcome? How do we show the provenance?</p>
iterative relationship between viz developers, scientists and the
+
<p>How to handle large scale grids – tiling, etc. Its not just a visualization issue – we need integrated data systems to deal with this issue – not just client applications.</p>
outreach audience. We didn’t like the notion that the public
+
<p>Services include – gridding (e.g., Barnes objective analysis), varying temporal and spatial resolutions, resampling, irregular and unstructured grids, pushing analysis onto the server due to the data set size. Need to stream them, etc.</p>
can’t understand science, and consider the question:
+
<p>For example, NCEP’s global 0.5 degree <span class="caps">GFS</span> model a single 3D field has dimensions of 720×361×26 and 61 time steps. This results in 412233120 points or 1.6 GB of data per field. Lots of data!</p>
aren’t scientist part of the public? It’s important to
+
<p>(T3)<br />
bring scientists and their work into the community, can effective
+
Grids are a sampling of an underlying continuous function that is reality. That sampling has aliasing artifacts. Those artifacts may vary from point to point within the data (imagine a satellite image at an angle across the Earth’s surface – the sample size varies from one side of the image to the next).</p>
visualization facilitate the needed two-way motivation needed between
+
<p>Grid to grid fusion often involves resampling those grids to merge them into a common grid. This introduces more aliasing artifacts. To reduce those artifacts we should use an interpolation function that models that underlying continuous function. However, we often do not. Perhaps the “correct” interpolation function is not known, a subject of debate, or not available in the software.</p>
scientists the community they serve?
+
<p>To get a handle on the artifacts introduced, and the interpolation function to use (or not use), we need to track the source of the data and the propagation of error. This becomes a file format issue because files often store the data result, but not the path to that result and the error function.</p>
 
+
<h2>Question 4C: Gridded data and <span class="caps">GIS</span>&#8230; getting these to talk to each other. How to make smart maps of <span class="caps">GIS</span> point and polygon data <span class="caps">AND</span> gridded satellite data.</h2>
Features for tools, easy to read, embedded description with the
+
<h1>Future Development Ideas</h1>
displays. Simple initial presentation, but allows a progressive
+
<h2>Question 1F: 2D, 3D, 4D: What&#8217;s the future of data display?</h2>
disclosure of information and concepts as far as the user desires.
+
<p>(T1)<br />
 
+
- Data, software, and hardware all can have different dimensionality.</p>
(T2)
+
<p>- How soon, if ever will 3 spatial dimension displays (stereo glasses, etc) become common.</p>
 
+
<p>- Issues of user interaction/control of displays, augmented reality (such iPhone location app) where user location in real world drives position in a 3 spatial dimension world shown on 2d hardware.</p>
  * Tune the visualization to the audience -- but how?
+
<p>- Single user vs. collaborative displays – for single user you could have level of detail optimizations, where eye look direction drives detail of display.</p>
    - use of toolkits, plug-ins, component frameworks
+
<p>(T2)<br />
    - providing different presentations of the same data
+
Google earth but with data.<br />
 
+
Immersive technology:<br />
  * Provide HTML/Flash/CMS for data
+
2D – image, contour<br />
    - easier mechanism for scientists/educators to convey info
+
3D – volume, surface<br />
    - equivalent of wiki or “build your own web page” for
+
4D – 3D + time, animated<br />
    general public
+
5D – multiple parameters</p>
 
+
<p>User interacts with data. Showing data in different ways, both geographically and analytically. Not just pretty pictures.<br />
Part B, How to teach the public about science and the scientists about
+
Probing<br />
the
+
Transects through data<br />
public?
+
Scatter plot<br />
 
+
Slice &amp; dice<br />
  * Force scientists through NSF, etc, new requirements on publishing
+
Time series analysis<br />
  * Use emerging technologies like social media
+
Multiple linked views into same data (“4-up”)<br />
  * Enable new ways of publishing
+
Geographic displays coupled with charts</p>
    – e.g. Data CMS, like RAMADDA
+
<p>For doing science, 3D has a problem with perspective</p>
 
+
<p>Exploration capability. Educators are adopting 3D, via Google Earth.</p>
General comments:
+
<p>Seems like we have the tools. But on a 2D display, you can control pixel color, transparency, and glyph, that’s it.</p>
 
+
<p>(T3)<br />
  * Visualization is conveying information
+
Our consensus is that 3D and beyond are not a good use of resources and that visualizations should concentrate on 2D and 2.5D visualizations. It is desirable but with very little return on the massive efforts.</p>
  * Teacher knows the answer, needs to find way to convey
+
<p>There are many reasons for this. Those being:</p>
  * Scientist does not know the answer, needs exploratory analysis
+
<p>Humans are really poor at perceiving depth. Studies have shown that humans do not perceive more than 6 depths at one time. Humans eyes are rarely of the same strength and contributes to depth perception problems.</p>
  * There is a declining importance of traditional journal
+
<p>Computer technology, at least at this point, does a very poor job or creating the illusion of depth.</p>
    publications
+
<p>Human perception in the Z plane is only about 10% of our 2D perception strength.</p>
 
+
<h2>Question 2F: Java/Spring/<span class="caps">AJAX</span> vs. Flex/Flash&#8230; what will rule in 5 years? What are the issues that need to be watched in terms of data visualization?</h2>
                            Current Issues
+
<p>(Tri)<br />
 
+
All of these technologies have problems. Some are heavy-weight that work well for applications, but not on the web (Java). Some are overly complex (<span class="caps">AJAX</span>). Some are proprietary or nearly so (Flash, Silverlight). Often they lack well-defined toolkits for building effective user interfaces. Weak standard support among browsers (such as IE) complicate matters.</p>
Question 1C: You tackle visualization tasks every day. What is the one thing
+
<p>An evolving trend is <span class="caps">HTML</span> 5 and JavaScript, plus JavaScript-based toolkits. These cover up browser quirks, leverage scheduling features in browser JavaScript schedules and sandboxing, and provide a client-side <span class="caps">GUI</span>. These toolkits, as the emerge and mature, may provide a good technology for building vis tools accessed via browsers on platforms from desktops to iPhones.</p>
that you do every day that needs to be different in 5 years for your life to
+
<p>(Square)<br />
improve?
+
Level of programming skill to produce rich application development for novices would be a factor. There’s some relationship to the previous question concerning presentation vs. interactive data analysis. Flash and Java seem well established, respectively, in dealing with these broad user categories. Popularity of small devices and social networking, are Flash/Java others amenable to these environments? With current pace of technological acceleration, 5 years could be a little outside a useful prediction.</p>
 
+
<p>(T3)</p>
  * data search and retrieval
+
<ul>
  * one data format to rule them all -- standards-based in all aspects
+
<li>Both will be around and viable/popular still</li>
    (data structure, metadata)
+
</ul>
 
+
<ul>
Question 2C: Name the three (or more) top obstacles today that limit the
+
<li>Plug-in presence used to be an issue, but muss less so now</li>
effective development/deployment of earth data visualizations.
+
</ul>
 
+
<ul>
Question 3C: State of gridded data fusion capabilities: what are the main
+
<li>Other technologies that may emerge as contenders<br />
obstacles/opportunities?
+
      – Canvas w/video audio <span class="caps">HTML</span> 5 tags<br />
 
+
      – Silverlight (less likely to dominate <span class="caps">IOO</span>)</li>
(T1)
+
</ul>
Error tracking and error propagation during data fusion has to be done
+
<ul>
and provided to the user. For example when resampling data for fusion
+
<li>Issues to watch in terms of data visulalization<br />
we need to keep track of the errors and how it affects the grid
+
      – ability to save/export to alternate formats<br />
fusion,
+
      e.g. export flex app as iPhone app<br />
 
+
      – capabilities that will emerge with Canvas<br />
Knowledge embedded in the tool to facilitate data fusion. Sort of an
+
      <span class="caps">API</span> (specifically relating to data import) enhancements</li>
knowledge based system that recommends to the user how the data fusion
+
</ul>
has to be done.
+
<h2>Question 3F: Standards for viewing earth data: what is the future&#8230; <span class="caps">JPEG</span> 2000? GeoTIFF? <span class="caps">KML</span>? Where are we going?</h2>
 
+
<p>(T3)<br />
Provide raw data,processed re-gridded data and software to data fusion
+
The data <strong>formats</strong> are generally adequate, but the structure of the data is often inadequate. GeoTIFF can easily be abused, <span class="caps">JPEG</span> 2000 isn’t well supported, <span class="caps">KML</span> is useful but primitive (limited support for map projections). Hopefully we’ll refine existing standards, rather than proliferating poorly supported standards.</p>
packaged together. Users have all the pieces needed for data fusion.
+
<h2>Question 4F: Data and video: Thoughts on building animations for video distribution. Where is this going?</h2>
 
+
<p>(Red)<br />
Algorithm/Tools/Gridding are data dependent.What techniques are used
+
Tools need to have more batch processing capabilities.</p>
are dependent on the data brought together.
+
<p>Automated production (creation and serving from web) vs one-off visualizations.</p>
 
+
<p>Time series data vs fly-thoughs – which animations needs human intervention.</p>
Embed data units,scaling, offsets within the data to facilitate
+
<p>There is a distinction between animations vs video with voice-over, music, close captioning. Video takes much more time, resources.</p>
fusion. We have data in different units, for example we have
+
<p>(Green)<br />
temperature data in Fahrenheit and in Celsius. Scaling/ data unit
+
We agree that if the data set is appropriate for time series distribution that applications should contain a component to output animation video.</p>
conversion has to be embedded in the data or interpreted readily by
+
<p>(Blue)<br />
the tool.
+
Question: Data and video: Thoughts on building animations for video distribution. Where is this going?</p>
 
+
<p>Video is dieing. It’s been years since we produced DVDs or tape. When a video is created, it’s in MPEG4, <span class="caps">AVI</span>, or whatever format can be played in PowerPoint or on a web page.</p>
Strengthen interaction between standards for data fusion. OpenDAP to
+
<p>However, the trend is away from these canned video presentations and instead towards live demos of visualization software run on the presenter’s laptop. This is often more credible, and it allows the presenter to adapt their presentation up to the last minute before their talk, or during their talk.</p>
OGC to web services.
+
<h2>Question 5F: Open Source / <span class="caps">COTS</span> vs. custom tools &#8212; issues with not-invented-here and the ability to write plugins for existing packages (or is the future going to continue to be comprised of an ever-expanding repertoire of software)?</h2>
 
+
<p>(T2)<br />
(T2)
+
Significant factors:</p>
How do we get the grid itself – some domains end up creating
+
<ul>
gridded data from point observation data. How does the different
+
<li>Politics are a major driver. Motivation to take credit for tangible results, branding etc.</li>
algorithms and parametrizations affect the outcome? How do we show the
+
</ul>
provenance?
+
<p>Assessment:</p>
 
+
<ul>
How to handle large scale grids – tiling, etc. Its not just a
+
<li>Status quo will likely continue. Proliferation of semi-redundant software is probably overall a positive.<br />
visualization issue – we need integrated data systems to deal
+
      – refinement<br />
with this issue – not just client applications.
+
      – reinforces successes.</li>
 
+
</ul>
Services include – gridding (e.g., Barnes objective analysis),
+
<p>(Red)<br />
varying temporal and spatial resolutions, resampling, irregular and
+
Too often scientists, etc., need to get their work done and do not (or cannot) have the luxury of time to use <span class="caps">COTS</span>, to do effective software engineering and management, i.e., to do the right thing from a software engineering perspective.</p>
unstructured grids, pushing analysis onto the server due to the data
+
<p>This is neither good nor bad – it is just the reality of the way things work.</p>
set size. Need to stream them, etc.
+
<p>(Green)<br />
 
+
Open Source<br />
For example, NCEP’s global 0.5 degree GFS model a single 3D
+
Has well documented code, and <span class="caps">API</span><br />
field has dimensions of 720×361×26 and 61 time steps.
+
Good design architecture for extension for customization ie, has a plug-in capability. No need to change or access system software.<br />
This results in 412233120 points or 1.6 GB of data per field. Lots of
+
Good community involvement, has anyone been able to do this.<br />
data!
+
If above yes, then why not use?</p>
 
+
<p><span class="caps">COTS</span>: issues with propriety data formats a problem, in either case, support for issues with software may be out of the users control.</p>
(T3)
+
<p>Custom tools can be very optimized to do a few tasks very well, from UI level to rendering level. A good visualization system would allow developers to extend the system at these different levels.</p>
Grids are a sampling of an underlying continuous function that is
+
<h1>Book related questions</h1>
reality. That sampling has aliasing artifacts. Those artifacts may
+
<h2>Question 1B: Specific content suggestions</h2>
vary from point to point within the data (imagine a satellite image at
+
<p>Data / meta-data standards, conventions.<br />
an angle across the Earth’s surface – the sample size
+
Standards and conventions for visual display (color map issues / conflicts between domains).<br />
varies from one side of the image to the next).
+
Color maps – selection based upon different intents (perception of different colors, generic conventions: “blue == cold”, “red=hot” ; “rainbow” is problematic, etc).<br />
 
+
Software Interoperability (and relation to file format standards and conventions).<br />
Grid to grid fusion often involves resampling those grids to merge
+
Survey of existing tools.<br />
them into a common grid. This introduces more aliasing artifacts. To
+
What’s the current state of the technology?<br />
reduce those artifacts we should use an interpolation function that
+
What software is available now?<br />
models that underlying continuous function. However, we often do not.
+
What needs to be fixed?<br />
Perhaps the “correct” interpolation function is not
+
Future directions.<br />
known, a subject of debate, or not available in the software.
+
What do we need to do better?<br />
 
+
Common (dimension-independant) issues VS problems specific to 2D or 3D data.</p>
To get a handle on the artifacts introduced, and the interpolation
+
<p>(T3)<br />
function to use (or not use), we need to track the source of the data
+
- How to incentivize sharing of data in usable formats a) requiring b) build an online community, ranking system for contributors c) build the best tool, providers will want to comply (Google Earth is good example – everyone wants to get their data working in GE now) &#8211; The most common current problems with status quo: data retrieval,data formats, etc.<br />
and the propagation of error. This becomes a file format issue because
+
- Current State-of-the-art tool summaries &#8211; Post meeting, have all attendees agree on Book chapter outline</p>
files often store the data result, but not the path to that result and
+
<p>(Tri)<br />
the error function.
+
- should discuss file formats/interoperability issues in abstract way<br />
 
+
- example of successes: can be partial successes such as NetCDF, which is very flexible syntax, but conventions for using syntax were not well established so many variations exist<br />
Question 4C: Gridded data and GIS... getting these to talk to each other.
+
- describes example visualization going through thought behind design decisions – examples of this already are “sound on sound” tutorials, visualization blogs such as http://eagereyes.org/</p>
How to make smart maps of GIS point and polygon data AND gridded satellite
+
<h2>Question 2B: Audience ideas and considerations</h2>
data.
+
<ul>
 
+
<li>non-professional visualizers as a way of communicating the practices, but not &#8220;dumbed-down&#8221;</li>
                      Future Development Ideas
+
</ul>
 
+
<h2>Question 3B: Website issues (links to durable URLs for product info, etc.)</h2>
Question 1F: 2D, 3D, 4D: What's the future of data display?
+
<ul>
 
+
<li>durable links</li>
(T1)
+
<li>Have online examples or not: examples will become dated quickly vs. interaction with examples discussed in book that will facilitate understanding of them</li>
- Data, software, and hardware all can have different dimensionality.
+
</ul>
 
+
<h2>Question 4B: Getting others involved OR not</h2>
- How soon, if ever will 3 spatial dimension displays (stereo glasses,
+
<p>Perhaps, but must be a strong commitment; number still needs to be kept so that logistics do not overwhelm.</p>
etc) become common.
+
<h2>What is the equivalent of pagerank for searching geoscience data?</h2>
 
+
<p>Criteria for computing &#8220;pagerank&#8221;:</p>
- Issues of user interaction/control of displays, augmented reality
+
<ul>
(such iPhone location app) where user location in real world drives
+
<li>Spatial and temporal resolution (assuming high-frequency and -resolution = good)</li>
position in a 3 spatial dimension world shown on 2d hardware.
+
<li>Time distance from desired date</li>
 
+
<li>Quality of metadata (compliant with standards)</li>
- Single user vs. collaborative displays – for single user you
+
<li>Includes calibration / control data</li>
could have level of detail optimizations, where eye look direction
+
<li>Provenance</li>
drives detail of display.
+
<li>Lowest error/trustworthiness</li>
 
+
<li>Data provider (credibility of curator)</li>
(T2)
+
<li>Consistency (outlier detection)</li>
Google earth but with data.
+
<li>Supported data formats (NetCDF vs. custom binary)</li>
Immersive technology:
+
<li>Popularity (links, citations, social ranking)</li>
2D – image, contour
+
<li>Frequency of updates (is dataset current and reliably updated)</li>
3D – volume, surface
+
<li>Access method (direct online access better than placing an order)</li>
4D – 3D + time, animated
+
</ul>
5D – multiple parameters
+
<h2>Common Sources of Error and How Can Visualization Tools Help?<br />
 
+
(<em>not absolutely certain what this actual question was</em>)</h2>
User interacts with data. Showing data in different ways, both
+
<p>Error or uncertainty? Uncertainty can be shown by flagging areas at certain data quality levels, showing confidence intervals, or plotting the results of model ensembles.</p>
geographically and analytically. Not just pretty pictures.
+
<p>Sources of error:</p>
Probing
+
<ul>
Transects through data
+
<li>Floating-point representation as text</li>
Scatter plot
+
<li>Gridding (of points), resampling, changing resolution, reprojecting</li>
Slice &amp; dice
+
<li>Interpolation</li>
Time series analysis
+
<li>Incomplete or inconsistent metadata</li>
Multiple linked views into same data (“4-up”)
+
<li>Undocumented satellite data correction</li>
Geographic displays coupled with charts
+
<li>Incorrect math (e.g., floor vs. ceil vs. trunc)</li>
 
+
</ul>
For doing science, 3D has a problem with perspective
+
<p>What visualization tools can do:</p>
 
+
<ul>
Exploration capability. Educators are adopting 3D, via Google Earth.
+
<li>Represent error in visualization using error bars, or color coding by uncertainty</li>
 
+
<li>Only possible when the error or possible sources of error are captured in the data or metadata</li>
Seems like we have the tools. But on a 2D display, you can control
+
</ul>
pixel color, transparency, and glyph, that’s it.
+
</div>
 
 
(T3)
 
Our consensus is that 3D and beyond are not a good use of resources
 
and that visualizations should concentrate on 2D and 2.5D
 
visualizations. It is desirable but with very little return on the
 
massive efforts.
 
 
 
There are many reasons for this. Those being:
 
 
 
Humans are really poor at perceiving depth. Studies have shown that
 
humans do not perceive more than 6 depths at one time. Humans eyes are
 
rarely of the same strength and contributes to depth perception
 
problems.
 
 
 
Computer technology, at least at this point, does a very poor job or
 
creating the illusion of depth.
 
 
 
Human perception in the Z plane is only about 10% of our 2D perception
 
strength.
 
 
 
Question 2F: Java/Spring/AJAX vs. Flex/Flash... what will rule in 5 years?
 
What are the issues that need to be watched in terms of data visualization?
 
 
 
(Tri)
 
All of these technologies have problems. Some are heavy-weight that
 
work well for applications, but not on the web (Java). Some are overly
 
complex (AJAX). Some are proprietary or nearly so (Flash,
 
Silverlight). Often they lack well-defined toolkits for building
 
effective user interfaces. Weak standard support among browsers (such
 
as IE) complicate matters.
 
 
 
An evolving trend is HTML 5 and JavaScript, plus JavaScript-based
 
toolkits. These cover up browser quirks, leverage scheduling features
 
in browser JavaScript schedules and sandboxing, and provide a
 
client-side GUI. These toolkits, as the emerge and mature, may provide
 
a good technology for building vis tools accessed via browsers on
 
platforms from desktops to iPhones.
 
 
 
(Square)
 
Level of programming skill to produce rich application development for
 
novices would be a factor. There’s some relationship to the
 
previous question concerning presentation vs. interactive data
 
analysis. Flash and Java seem well established, respectively, in
 
dealing with these broad user categories. Popularity of small devices
 
and social networking, are Flash/Java others amenable to these
 
environments? With current pace of technological acceleration, 5 years
 
could be a little outside a useful prediction.
 
 
 
(T3)
 
 
 
  * Both will be around and viable/popular still
 
 
 
  * Plug-in presence used to be an issue, but muss less so now
 
 
 
  * Other technologies that may emerge as contenders
 
    – Canvas w/video audio HTML 5 tags
 
    – Silverlight (less likely to dominate IOO)
 
 
 
  * Issues to watch in terms of data visulalization
 
    – ability to save/export to alternate formats
 
    e.g. export flex app as iPhone app
 
    – capabilities that will emerge with Canvas
 
    – API (specifically relating to data import) enhancements
 
 
 
Question 3F: Standards for viewing earth data: what is the future... JPEG
 
2000? GeoTIFF? KML? Where are we going?
 
 
 
(T3)
 
The data formats are generally adequate, but the structure of the data
 
is often inadequate. GeoTIFF can easily be abused, JPEG 2000
 
isn’t well supported, KML is useful but primitive (limited
 
support for map projections). Hopefully we’ll refine existing
 
standards, rather than proliferating poorly supported standards.
 
 
 
Question 4F: Data and video: Thoughts on building animations for video
 
distribution. Where is this going?
 
 
 
(Red)
 
Tools need to have more batch processing capabilities.
 
 
 
Automated production (creation and serving from web) vs one-off
 
visualizations.
 
 
 
Time series data vs fly-thoughs – which animations needs human
 
intervention.
 
 
 
There is a distinction between animations vs video with voice-over,
 
music, close captioning. Video takes much more time, resources.
 
 
 
(Green)
 
We agree that if the data set is appropriate for time series
 
distribution that applications should contain a component to output
 
animation video.
 
 
 
(Blue)
 
Question: Data and video: Thoughts on building animations for video
 
distribution. Where is this going?
 
 
 
Video is dieing. It’s been years since we produced DVDs or
 
tape. When a video is created, it’s in MPEG4, AVI, or whatever
 
format can be played in PowerPoint or on a web page.
 
 
 
However, the trend is away from these canned video presentations and
 
instead towards live demos of visualization software run on the
 
presenter’s laptop. This is often more credible, and it allows
 
the presenter to adapt their presentation up to the last minute before
 
their talk, or during their talk.
 
 
 
Question 5F: Open Source / COTS vs. custom tools -- issues with
 
not-invented-here and the ability to write plugins for existing packages (or
 
is the future going to continue to be comprised of an ever-expanding
 
repertoire of software)?
 
 
 
(T2)
 
Significant factors:
 
 
 
  * Politics are a major driver. Motivation to take credit for
 
    tangible results, branding etc.
 
 
 
Assessment:
 
 
 
  * Status quo will likely continue. Proliferation of semi-redundant
 
    software is probably overall a positive.
 
    – refinement
 
    – reinforces successes.
 
 
 
(Red)
 
Too often scientists, etc., need to get their work done and do not (or
 
cannot) have the luxury of time to use COTS, to do effective software
 
engineering and management, i.e., to do the right thing from a
 
software engineering perspective.
 
 
 
This is neither good nor bad – it is just the reality of the
 
way things work.
 
 
 
(Green)
 
Open Source
 
Has well documented code, and API
 
Good design architecture for extension for customization ie, has a
 
plug-in capability. No need to change or access system software.
 
Good community involvement, has anyone been able to do this.
 
If above yes, then why not use?
 
 
 
COTS: issues with propriety data formats a problem, in either case,
 
support for issues with software may be out of the users control.
 
 
 
Custom tools can be very optimized to do a few tasks very well, from
 
UI level to rendering level. A good visualization system would allow
 
developers to extend the system at these different levels.
 
 
 
                        Book related questions
 
 
 
Question 1B: Specific content suggestions
 
 
 
Data / meta-data standards, conventions.
 
Standards and conventions for visual display (color map issues /
 
conflicts between domains).
 
Color maps – selection based upon different intents (perception
 
of different colors, generic conventions: “blue ==
 
cold”, “red=hot” ; “rainbow” is
 
problematic, etc).
 
Software Interoperability (and relation to file format standards and
 
conventions).
 
Survey of existing tools.
 
What’s the current state of the technology?
 
What software is available now?
 
What needs to be fixed?
 
Future directions.
 
What do we need to do better?
 
Common (dimension-independant) issues VS problems specific to 2D or 3D
 
data.
 
 
 
(T3)
 
- How to incentivize sharing of data in usable formats a) requiring b)
 
build an online community, ranking system for contributors c) build
 
the best tool, providers will want to comply (Google Earth is good
 
example – everyone wants to get their data working in GE now) -
 
The most common current problems with status quo: data retrieval,data
 
formats, etc.
 
- Current State-of-the-art tool summaries - Post meeting, have all
 
attendees agree on Book chapter outline
 
 
 
(Tri)
 
- should discuss file formats/interoperability issues in abstract way
 
- example of successes: can be partial successes such as NetCDF, which
 
is very flexible syntax, but conventions for using syntax were not
 
well established so many variations exist
 
- describes example visualization going through thought behind design
 
decisions – examples of this already are “sound on
 
sound” tutorials, visualization blogs such as
 
http://eagereyes.org/
 
 
 
Question 2B: Audience ideas and considerations
 
 
 
  * non-professional visualizers as a way of communicating the
 
    practices, but not &quot;dumbed-down&quot;
 
 
 
Question 3B: Website issues (links to durable URLs for product info, etc.)
 
 
 
  * durable links
 
  * Have online examples or not: examples will become dated quickly
 
    vs. interaction with examples discussed in book that will
 
    facilitate understanding of them
 
 
 
Question 4B: Getting others involved OR not
 
 
 
Perhaps, but must be a strong commitment; number still needs to be
 
kept so that logistics do not overwhelm.
 
 
 
What is the equivalent of pagerank for searching geoscience data?
 
 
 
Criteria for computing &quot;pagerank&quot;:
 
 
 
  * Spatial and temporal resolution (assuming high-frequency and
 
    -resolution = good)
 
  * Time distance from desired date
 
  * Quality of metadata (compliant with standards)
 
  * Includes calibration / control data
 
  * Provenance
 
  * Lowest error/trustworthiness
 
  * Data provider (credibility of curator)
 
  * Consistency (outlier detection)
 
  * Supported data formats (NetCDF vs. custom binary)
 
  * Popularity (links, citations, social ranking)
 
  * Frequency of updates (is dataset current and reliably updated)
 
  * Access method (direct online access better than placing an order)
 
 
 
Common Sources of Error and How Can Visualization Tools Help?
 
(not absolutely certain what this actual question was)
 
 
 
Error or uncertainty? Uncertainty can be shown by flagging areas at
 
certain data quality levels, showing confidence intervals, or plotting
 
the results of model ensembles.
 
 
 
Sources of error:
 
 
 
  * Floating-point representation as text
 
  * Gridding (of points), resampling, changing resolution,
 
    reprojecting
 
  * Interpolation
 
  * Incomplete or inconsistent metadata
 
  * Undocumented satellite data correction
 
  * Incorrect math (e.g., floor vs. ceil vs. trunc)
 
 
 
What visualization tools can do:
 
 
 
  * Represent error in visualization using error bars, or color coding
 
    by uncertainty
 
  * Only possible when the error or possible sources of error are
 
    captured in the data or metadata
 

Revision as of 13:18, July 26, 2011

* RESPONSES TO BIG QUESTIONS (edited)

still in progress

Overall questions

Question 1: Imagine the perfect earth data remote sensing visualization tool/system. What are the main components of this system?

  • Needs to closely track needs and abilities of a variety of audiences
  • Support multiple GUIs (meaning API-based back-end to support tools?)
  • Open Source
  • Solves data format/standards problem — new data type? derivative data types?
  • Social web, community-based: support collaborative workflow and data discovery
  • One-stop repository to support aggregate querying — e.g., What is the causality relationship between meningitis outbreaks and precipitation patterns in sub-Saharan Africa?
  • Provide ability to store and track provenance

Question 2: Visualization for understanding vs. conveying — It’s all about the user: How should tools and visualizations be tailored for distinct user groups?

Two different issues:

  • Conveying — you know the answer, need to display it in a good way. “Static”
  • Understanding — trying to find the answer via exploration. Needs more facilities. “Interactive”

Hard to build a general-purpose tool, to satisfy power users vs. non-power users, cross-discipline. Same tool needs to scale with abilities.

Question 3: Are there user groups that you know are under-served by the current data visualization technology? What needs to change to serve these groups?

  • Everyone is underserved, but not equally.
  • PIs/those closest to data are best served by way of their familiarity with the data.
  • Solving this problem is an argument in favor of plug-in based development.
  • Tools need to be developed to support each audience — different audiences have wildly different needs in terms of capabilities and end result.
  • Of course, serving user groups only comes after they have located the data.

Question 4: Delivering data vs. images of data. What are the sweet spots for each? What are the areas where we need to focus or change?

  • It depends on audience: images and data can serve both “public” and scientific audiences
    - In general images are great for the general public or less interested specialists (maybe 80%), data needs to be available for the interested 20%.
    - The <a href="http://www.ifp.illinois.edu/nabhcs/abstracts/shneiderman.html">Ben Shneiderman UI design mantra</a> “overview, zoom & filter, details on demand” applies in this case.

Images sweet spots

  • As a means of discovery and monitoring production
  • Able to meet needs of wider audience [formats and general ease of use]

Data sweet spots

  • Scientist users who create their own visualizations
  • maintaining metadata (provenance)
  • can be used to derive multiple visualizations

Question 5: How can visualizers bridge the science-outreach divide? How to teach the public about science and the scientists about the public?

Visualization is a compelling medium that science communicators can use to make complex scientific ideas approachable to a broad audience. Carl Sagan’s Cosmos series is the prototypical example, weaving visuals with narrative to explain astrophysics. It is crucial to define an audience: there is no such thing as a uniform general public. Concrete visualizations, such as planet walks and painted lines representing sea level rise, can be particularly effective. Effective visualization requires focus: emphasize important elements of a dataset, and de-emphasize or eliminate less important data. Ideally, tools would be designed by visualizers, not computer scientists. Try for verisimilitude: make things appear how the audience expects it to appear (for example Google Earth’s discontinuous boundaries between scenes are very distracting).

(T1)
Perhaps current outreach programs are too top-down, we need a tighter, iterative relationship between viz developers, scientists and the outreach audience. We didn’t like the notion that the public can’t understand science, and consider the question: aren’t scientist part of the public? It’s important to bring scientists and their work into the community, can effective visualization facilitate the needed two-way motivation needed between scientists the community they serve?

Features for tools, easy to read, embedded description with the displays. Simple initial presentation, but allows a progressive disclosure of information and concepts as far as the user desires.

(T2)

  • Tune the visualization to the audience — but how?
    – use of toolkits, plug-ins, component frameworks
    – providing different presentations of the same data
  • Provide HTML/Flash/CMS for data
    – easier mechanism for scientists/educators to convey info
    – equivalent of wiki or “build your own web page” for general public

Part B, How to teach the public about science and the scientists about the
public?

  • Force scientists through NSF, etc, new requirements on publishing
  • Use emerging technologies like social media
  • Enable new ways of publishing
    – e.g. Data CMS, like RAMADDA

General comments:

  • Visualization is conveying information
  • Teacher knows the answer, needs to find way to convey
  • Scientist does not know the answer, needs exploratory analysis
  • There is a declining importance of traditional journal publications

Current Issues

Question 1C: You tackle visualization tasks every day. What is the one thing that you do every day that needs to be different in 5 years for your life to improve?

  • data search and retrieval
  • one data format to rule them all — standards-based in all aspects (data structure, metadata)

Question 2C: Name the three (or more) top obstacles today that limit the effective development/deployment of earth data visualizations.

Question 3C: State of gridded data fusion capabilities: what are the main obstacles/opportunities?

(T1)
Error tracking and error propagation during data fusion has to be done and provided to the user. For example when resampling data for fusion we need to keep track of the errors and how it affects the grid fusion,

Knowledge embedded in the tool to facilitate data fusion. Sort of an knowledge based system that recommends to the user how the data fusion has to be done.

Provide raw data,processed re-gridded data and software to data fusion packaged together. Users have all the pieces needed for data fusion.

Algorithm/Tools/Gridding are data dependent.What techniques are used are dependent on the data brought together.

Embed data units,scaling, offsets within the data to facilitate fusion. We have data in different units, for example we have temperature data in Fahrenheit and in Celsius. Scaling/ data unit conversion has to be embedded in the data or interpreted readily by the tool.

Strengthen interaction between standards for data fusion. OpenDAP to OGC to web services.

(T2)
How do we get the grid itself – some domains end up creating gridded data from point observation data. How does the different algorithms and parametrizations affect the outcome? How do we show the provenance?

How to handle large scale grids – tiling, etc. Its not just a visualization issue – we need integrated data systems to deal with this issue – not just client applications.

Services include – gridding (e.g., Barnes objective analysis), varying temporal and spatial resolutions, resampling, irregular and unstructured grids, pushing analysis onto the server due to the data set size. Need to stream them, etc.

For example, NCEP’s global 0.5 degree GFS model a single 3D field has dimensions of 720×361×26 and 61 time steps. This results in 412233120 points or 1.6 GB of data per field. Lots of data!

(T3)
Grids are a sampling of an underlying continuous function that is reality. That sampling has aliasing artifacts. Those artifacts may vary from point to point within the data (imagine a satellite image at an angle across the Earth’s surface – the sample size varies from one side of the image to the next).

Grid to grid fusion often involves resampling those grids to merge them into a common grid. This introduces more aliasing artifacts. To reduce those artifacts we should use an interpolation function that models that underlying continuous function. However, we often do not. Perhaps the “correct” interpolation function is not known, a subject of debate, or not available in the software.

To get a handle on the artifacts introduced, and the interpolation function to use (or not use), we need to track the source of the data and the propagation of error. This becomes a file format issue because files often store the data result, but not the path to that result and the error function.

Question 4C: Gridded data and GIS… getting these to talk to each other. How to make smart maps of GIS point and polygon data AND gridded satellite data.

Future Development Ideas

Question 1F: 2D, 3D, 4D: What’s the future of data display?

(T1)
- Data, software, and hardware all can have different dimensionality.

- How soon, if ever will 3 spatial dimension displays (stereo glasses, etc) become common.

- Issues of user interaction/control of displays, augmented reality (such iPhone location app) where user location in real world drives position in a 3 spatial dimension world shown on 2d hardware.

- Single user vs. collaborative displays – for single user you could have level of detail optimizations, where eye look direction drives detail of display.

(T2)
Google earth but with data.
Immersive technology:
2D – image, contour
3D – volume, surface
4D – 3D + time, animated
5D – multiple parameters

User interacts with data. Showing data in different ways, both geographically and analytically. Not just pretty pictures.
Probing
Transects through data
Scatter plot
Slice & dice
Time series analysis
Multiple linked views into same data (“4-up”)
Geographic displays coupled with charts

For doing science, 3D has a problem with perspective

Exploration capability. Educators are adopting 3D, via Google Earth.

Seems like we have the tools. But on a 2D display, you can control pixel color, transparency, and glyph, that’s it.

(T3)
Our consensus is that 3D and beyond are not a good use of resources and that visualizations should concentrate on 2D and 2.5D visualizations. It is desirable but with very little return on the massive efforts.

There are many reasons for this. Those being:

Humans are really poor at perceiving depth. Studies have shown that humans do not perceive more than 6 depths at one time. Humans eyes are rarely of the same strength and contributes to depth perception problems.

Computer technology, at least at this point, does a very poor job or creating the illusion of depth.

Human perception in the Z plane is only about 10% of our 2D perception strength.

Question 2F: Java/Spring/AJAX vs. Flex/Flash… what will rule in 5 years? What are the issues that need to be watched in terms of data visualization?

(Tri)
All of these technologies have problems. Some are heavy-weight that work well for applications, but not on the web (Java). Some are overly complex (AJAX). Some are proprietary or nearly so (Flash, Silverlight). Often they lack well-defined toolkits for building effective user interfaces. Weak standard support among browsers (such as IE) complicate matters.

An evolving trend is HTML 5 and JavaScript, plus JavaScript-based toolkits. These cover up browser quirks, leverage scheduling features in browser JavaScript schedules and sandboxing, and provide a client-side GUI. These toolkits, as the emerge and mature, may provide a good technology for building vis tools accessed via browsers on platforms from desktops to iPhones.

(Square)
Level of programming skill to produce rich application development for novices would be a factor. There’s some relationship to the previous question concerning presentation vs. interactive data analysis. Flash and Java seem well established, respectively, in dealing with these broad user categories. Popularity of small devices and social networking, are Flash/Java others amenable to these environments? With current pace of technological acceleration, 5 years could be a little outside a useful prediction.

(T3)

  • Both will be around and viable/popular still
  • Plug-in presence used to be an issue, but muss less so now
  • Other technologies that may emerge as contenders
    – Canvas w/video audio HTML 5 tags
    – Silverlight (less likely to dominate IOO)
  • Issues to watch in terms of data visulalization
    – ability to save/export to alternate formats
    e.g. export flex app as iPhone app
    – capabilities that will emerge with Canvas
    API (specifically relating to data import) enhancements

Question 3F: Standards for viewing earth data: what is the future… JPEG 2000? GeoTIFF? KML? Where are we going?

(T3)
The data formats are generally adequate, but the structure of the data is often inadequate. GeoTIFF can easily be abused, JPEG 2000 isn’t well supported, KML is useful but primitive (limited support for map projections). Hopefully we’ll refine existing standards, rather than proliferating poorly supported standards.

Question 4F: Data and video: Thoughts on building animations for video distribution. Where is this going?

(Red)
Tools need to have more batch processing capabilities.

Automated production (creation and serving from web) vs one-off visualizations.

Time series data vs fly-thoughs – which animations needs human intervention.

There is a distinction between animations vs video with voice-over, music, close captioning. Video takes much more time, resources.

(Green)
We agree that if the data set is appropriate for time series distribution that applications should contain a component to output animation video.

(Blue)
Question: Data and video: Thoughts on building animations for video distribution. Where is this going?

Video is dieing. It’s been years since we produced DVDs or tape. When a video is created, it’s in MPEG4, AVI, or whatever format can be played in PowerPoint or on a web page.

However, the trend is away from these canned video presentations and instead towards live demos of visualization software run on the presenter’s laptop. This is often more credible, and it allows the presenter to adapt their presentation up to the last minute before their talk, or during their talk.

Question 5F: Open Source / COTS vs. custom tools — issues with not-invented-here and the ability to write plugins for existing packages (or is the future going to continue to be comprised of an ever-expanding repertoire of software)?

(T2)
Significant factors:

  • Politics are a major driver. Motivation to take credit for tangible results, branding etc.

Assessment:

  • Status quo will likely continue. Proliferation of semi-redundant software is probably overall a positive.
    – refinement
    – reinforces successes.

(Red)
Too often scientists, etc., need to get their work done and do not (or cannot) have the luxury of time to use COTS, to do effective software engineering and management, i.e., to do the right thing from a software engineering perspective.

This is neither good nor bad – it is just the reality of the way things work.

(Green)
Open Source
Has well documented code, and API
Good design architecture for extension for customization ie, has a plug-in capability. No need to change or access system software.
Good community involvement, has anyone been able to do this.
If above yes, then why not use?

COTS: issues with propriety data formats a problem, in either case, support for issues with software may be out of the users control.

Custom tools can be very optimized to do a few tasks very well, from UI level to rendering level. A good visualization system would allow developers to extend the system at these different levels.

Book related questions

Question 1B: Specific content suggestions

Data / meta-data standards, conventions.
Standards and conventions for visual display (color map issues / conflicts between domains).
Color maps – selection based upon different intents (perception of different colors, generic conventions: “blue == cold”, “red=hot” ; “rainbow” is problematic, etc).
Software Interoperability (and relation to file format standards and conventions).
Survey of existing tools.
What’s the current state of the technology?
What software is available now?
What needs to be fixed?
Future directions.
What do we need to do better?
Common (dimension-independant) issues VS problems specific to 2D or 3D data.

(T3)
- How to incentivize sharing of data in usable formats a) requiring b) build an online community, ranking system for contributors c) build the best tool, providers will want to comply (Google Earth is good example – everyone wants to get their data working in GE now) – The most common current problems with status quo: data retrieval,data formats, etc.
- Current State-of-the-art tool summaries – Post meeting, have all attendees agree on Book chapter outline

(Tri)
- should discuss file formats/interoperability issues in abstract way
- example of successes: can be partial successes such as NetCDF, which is very flexible syntax, but conventions for using syntax were not well established so many variations exist
- describes example visualization going through thought behind design decisions – examples of this already are “sound on sound” tutorials, visualization blogs such as http://eagereyes.org/

Question 2B: Audience ideas and considerations

  • non-professional visualizers as a way of communicating the practices, but not “dumbed-down”

Question 3B: Website issues (links to durable URLs for product info, etc.)

  • durable links
  • Have online examples or not: examples will become dated quickly vs. interaction with examples discussed in book that will facilitate understanding of them

Question 4B: Getting others involved OR not

Perhaps, but must be a strong commitment; number still needs to be kept so that logistics do not overwhelm.

What is the equivalent of pagerank for searching geoscience data?

Criteria for computing “pagerank”:

  • Spatial and temporal resolution (assuming high-frequency and -resolution = good)
  • Time distance from desired date
  • Quality of metadata (compliant with standards)
  • Includes calibration / control data
  • Provenance
  • Lowest error/trustworthiness
  • Data provider (credibility of curator)
  • Consistency (outlier detection)
  • Supported data formats (NetCDF vs. custom binary)
  • Popularity (links, citations, social ranking)
  • Frequency of updates (is dataset current and reliably updated)
  • Access method (direct online access better than placing an order)

Common Sources of Error and How Can Visualization Tools Help?
(not absolutely certain what this actual question was)

Error or uncertainty? Uncertainty can be shown by flagging areas at certain data quality levels, showing confidence intervals, or plotting the results of model ensembles.

Sources of error:

  • Floating-point representation as text
  • Gridding (of points), resampling, changing resolution, reprojecting
  • Interpolation
  • Incomplete or inconsistent metadata
  • Undocumented satellite data correction
  • Incorrect math (e.g., floor vs. ceil vs. trunc)

What visualization tools can do:

  • Represent error in visualization using error bars, or color coding by uncertainty
  • Only possible when the error or possible sources of error are captured in the data or metadata