How so?
To write a parser for a binary file one needs to know the detailed structure of
the file - and if that changes your code is now broken. With GML or any
XML-based encoding the file structure is not known to your application. Where
do you see the problems in parsing and understanding GML?
R
-----Original Message-----
From: Gerry Creager [mailto:gerry.creager@xxxxxxxx]
Sent: August 21, 2009 1:17 PM
To: Ron Lake
Cc: John Caron; Unidata GALEON; wcs-2.0.swg
Subject: Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives
I'm not so sure that's true.
Ron Lake wrote:
> Hi John:
>
> Surely the GML encoding is going to be simpler to parse and "understand" than
> any equivalent binary encoding.
>
> R
>
> -----Original Message-----
> From: galeon-bounces@xxxxxxxxxxxxxxxx
> [mailto:galeon-bounces@xxxxxxxxxxxxxxxx] On Behalf Of John Caron
> Sent: August 21, 2009 12:52 PM
> Cc: Unidata GALEON; wcs-2.0.swg
> Subject: Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives
>
> A few thoughts on these issues:
>
> The WCS/SWE/OGC are making major contributions in elaborating conceptual
> models and defining remote access protocols based on them. More
> problematic is managing complexity and adoption paths of existing
> communities. One of the main weaknesses of WCS is the likely
> proliferation of binary encoding formats. WMS has been lucky to be able
> to use existing, universal formats like JPEG, and partly because of
> that, has been more widely adopted.
>
> A central issue is the relation of the WCS data encoding format to its
> GML envelope. Should the data be "just bits" and let the GML encode all
> the semantics, or should a binary format also encode some/all of the
> semantics? In the first case, any client must understand GML, which is a
> big barrier to use by existing applications, plus the possibility exists
> that some essential information is offline or unavailable, as Tom W
> points out. In the second case, we have a proliferation of data encoding
> formats and the possibility that the semantics are encoded differently
> between the GML and the data.
>
> From Unidata's POV, namely that of supporting a large community of
> users of netcdf and other scientific data formats, using netCDF-CF to
> encode the data is obvious. However, this is not just a matter of one
> more group advocating for their favorite bespoke format. IMHO, netCDF is
> close to being a "maximally simple implementation" of a binary data
> encoding that allows data to be more than "just numbers". Its BNF
> grammer fits on one side of a page of paper. If one chooses to encode
> semantics into the data format, and not require a client to know GML,
> then you need something like netCDF, and you wont find an encoding
> significantly simpler. The CF Conventions group has been slowly adding
> the necessary semantics for 10 years or so. From that motivation we
> decided to recommend netCDF as a WCS encoding format.
>
> The hard part is describing in a formal way the relationship of the
> netCDF-CF binary response to the actual request and to any GML part of
> the response. Stefano et al are doing the heroic work of describing the
> mapping between the two models, but there is and perhaps should be a
> loose coupling between the two. There are always edge cases: what
> happens when the requested lat/lon bounding box crosses a discontinuity
> of the projection plane? One might get different, but reasonable, files
> back from different servers of the same dataset. So what can we
> standardize? None of this is unique to netcdf-CF or binary encodings.
>
> Taking the first path of letting GML carry all the semantics has many
> advantages also, especially for new development and GIS-centric clients,
> and as new libraries are developed that can mitigate some of the
> complexity of GML. In (our) scientific community, these are new
> possibilities, to be experimented with and prototypes tested. Once there
> are real services delivering "gotta have" data we will do whatever we
> need to do to "get me some of that" ;^)
>
> John Caron
> Unidata
>
> Steve Hankin wrote:
>>
>> Robin, Alexandre wrote:
>>> Hi Steve,
>>>
>>> Just to clarify when I said NetCDF was a "NEW standard" I meant a new
>>> standard in OGC.
>>>
>>> As I was telling Ben in an offline email, I am totally conscious of
>>> its penetration and usefulness in certain communities.
>>>
>>> However, _I am not convinced that /having two standards doing the
>>> same thing in OGC /is sending the right message and is the best way
>>> to go for a standardization organization_.
>>>
>> Hi Robin,
>>
>> I confess that I was aware of using a cheap rhetorical device when I
>> twisted your intended meaning of "NEW". (Begging your tolerance.) It
>> was helpful in order to raise more fundamental questions. You have
>> alluded to a key question just above. Is it really best to think of
>> the target of OGC as a the development of a single, definitive
>> standard? one that is more general and more powerful than all existing
>> standards? Or is it better to think of OGC as a process, through which
>> the forces of divergence in geospatial IT systems can be weakened
>> leading towards convergence over time? The notion that there can be a
>> single OGC solution is already patently an illusion. Which one would
>> you pick? WFS? WCS? SOS with SWE Common? SOS with its many other XML
>> schema? (Lets not even look into the profusion of WFS application
>> schema.) I trust that we are not pinning our hopes on a future
>> consolidation of all of these. There is little evidence to indicate
>> that we can sustain the focus necessary to traverse that path. The
>> underlying technology is not standing still.
>>
>> What Ben (and David Arctur and others) have proposed through seeking
>> to put an OGC stamp of approval on netCDF-CF technology is similar to
>> what OGC has achieved through putting its stamp on KML ("There are
>> sound business and policy reasons for doing so.") It is to create a
>> process -- a technical conversation if you will -- which will lead to
>> interoperability pathways that bridge technologies and communities.
>> Real-world interoperability.
>>> There has been a lot of experimentation with SWE technologies as well
>>> that you may not know about and in many communities, especially in
>>> earth science.
>>>
>>> What I'm saying is that perhaps it is worth testing bridging NetCDF
>>> to SWE before we go the way of stamping two 100% overlapping
>>> standards as OGC compliant.
>>>
>> Complete agreement that this sort of testing ought to occur. And
>> interest to hear more about what has been achieved. But great
>> skepticism that there is this degree of overlap between the
>> approaches. And disagreement that this testing ought to be a
>> precondition to OGC recognition of a significant ,community-proven
>> interoperability mechanism like netCDF. OGC standardization of netCDF
>> will provide a forum for testing and experimentation to occur much
>> more rapidly and for a 2-way transfer of the best ideas between
>> approaches. NetCDF & co. (its API, data model, CF, DAP) have a great
>> deal to offer to OGC.
>>
>> - Steve
>>> Regards,
>>>
>>> -------------------------------------------------
>>>
>>> **Alexandre Robin**
>>>
>>> Spot Image, Web and E-Business
>>>
>>> Tel: +33 (0)5 62 19 43 62
>>>
>>> Fax: +33 (0)5 62 19 43 43
>>>
>>> http://www.spotimage.com
>>>
>>> Before printing, think about the environment
>>>
>>> ------------------------------------------------------------------------
>>>
>>> *De :* Steve Hankin [mailto:Steven.C.Hankin@xxxxxxxx]
>>> *Envoyé :* jeudi 20 août 2009 20:58
>>> *À :* Tom Whittaker
>>> *Cc :* Robin, Alexandre; Ben Domenico; Unidata GALEON; wcs-2.0.swg
>>> *Objet :* Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives
>>>
>>> Hi Tom,
>>>
>>> I am grateful to you for opening the door to comments "from 10
>>> thousand feet" -- fundamental truths that we know from many years of
>>> experience, but that we fear may be getting short shrift in
>>> discussions of a new technology. I'd like to offer a comment of that
>>> sort regarding the interplay of ideas today between Robin ("/I hope
>>> we don't have to define a NEW standard .../") and Carl Reed ("/there
>>> are other organizations interested in bringing legacy spatial
>>> encodings into the OGC. There are sound business and policy reasons
>>> for doing so./").
>>>
>>> The NEW standard in this discussion is arguably SWE, rather than
>>> netCDF. NetCDF has decades of practice behind it; huge bodies of data
>>> based upon it; a wide range of applications capable of accessing it
>>> (both locally and remotely); and communities that depend vitally upon
>>> it. As Ben points out, netCDF also has its own de jure pedigree.
>>>
>>> A key peril shared by most IT standards committees -- a lesson that
>>> has been learned, forgotten, relearned and forgotten again so many
>>> times that it is clearly an issue of basic human behavior -- is that
>>> they will try to innovate. Too-common committee behavior is to
>>> propose, discuss and document new and intriguing technologies, and
>>> then advance those documents through a de jure standards process,
>>> despite an insufficient level of testing. The OGC testbed process
>>> exists to address this, but we see continually how large the gap is
>>> between the testbed process and the pace and complexity of
>>> innovations emerging from committees.
>>>
>>> Excellent reading on this subject is the essay by Michi Henning, /The
>>> Rise and Fall of CORBA/ (2006 --
>>> http://queue.acm.org/detail.cfm?id=1142044). Among the many insights
>>> he offers is
>>>
>>> **'Standards consortia need iron-clad rules to ensure that they
>>> standardize existing best practice.** There is no room for innovation
>>> in standards_._ Throwing in "just that extra little feature"
>>> inevitably causes unforeseen technical problems, despite the best
>>> intentions.'
>>>
>>> While it adds weight to an argument to be able to quote from an
>>> in-print source, this is a self-evident truth. We need only reflect
>>> on the recent history of IT. What we need is to work together to find
>>> ways to prevent ourselves from continually forgetting it.
>>>
>>> There is little question in my mind that putting an OGC stamp of
>>> approval on netCDF is a win-win process -- for the met/ocean/climate
>>> community and for the broader geospatial community. It will be a path
>>> to greater interoperability in the long run and it deserves to go
>>> forward. The merits of SWE (or GML) as an alternative approach to the
>>> same functionality also deserve to be explored and tested in
>>> situations of realistic complexity. But this exploration should be
>>> understood initially as a process of R&D -- a required step before a
>>> "standards process" is considered. If that exploration has already
>>> been done it should be widely disseminated, discussed and evaluated.
>>>
>>> - Steve
>>>
>>> ==================================
>>>
>>> Tom Whittaker wrote:
>>>
>>> I may be ignorant about these issues, so please forgive me if I am
>>> completely out-of-line....but when I looked at the examples, I got
>>> very concerned since the metadata needed to interpret the data values
>>> in the "data files" is apparently not actually in the file, but
>>> somewhere else. We've been here before: One of the single biggest
>>> mistakes that the meteorological community made in defining a
>>> distribution format for realtime, streaming data was BUFR -- because
>>> the "tables" needed to interpret the contents of the files are
>>> somewhere else....and sometimes, end users cannot find them!
>>>
>>> NetCDF and ncML maintain the essential metadata within the files:
>>> types, units, coordinates -- and I strongly urge you (or whomever) not
>>> to make the "BUFR mistake" again -- put the metadata into the files!
>>> Do not require the end user to have to have an internet connection to
>>> simply "read" the data....many people download the files and then
>>> "take them along" when traveling, for example.
>>>
>>> If I simply downloaded the file at
>>> <http://schemas.opengis.net/om/1.0.0/examples/weatherObservation.xml>
>>> I would not be able to read it. In fact, it looks like even if I also
>>> got the "metadata" file at:
>>> <http://schemas.opengis.net/om/1.0.0/examples/weatherRecord1.xml>
>>> I would still not be able to read it, since it also refers to other
>>> servers in the universe to obtain essential metadata.
>>>
>>> That is my 2 cents worth....and I hope I am wrong about what I saw in
>>> the examples....
>>>
>>> tom
>>>
>>>
>> ------------------------------------------------------------------------
>>
>> _______________________________________________
>> galeon mailing list
>> galeon@xxxxxxxxxxxxxxxx
>> For list information, to unsubscribe, visit:
>> http://www.unidata.ucar.edu/mailing_lists/
>
> _______________________________________________
> galeon mailing list
> galeon@xxxxxxxxxxxxxxxx
> For list information, to unsubscribe, visit:
> http://www.unidata.ucar.edu/mailing_lists/
>
> _______________________________________________
> galeon mailing list
> galeon@xxxxxxxxxxxxxxxx
> For list information, to unsubscribe, visit:
> http://www.unidata.ucar.edu/mailing_lists/
--
Gerry Creager -- gerry.creager@xxxxxxxx
Texas Mesonet -- AATLT, Texas A&M University
Cell: 979.229.5301 Office: 979.458.4020 FAX: 979.862.3983
Office: 1700 Research Parkway Ste 160, TAMU, College Station, TX 77843