John:
It seems that we are going to a review of MEAD project. MEAD is an NSF 
expedition project back to 4 years ago. At first, we didn't committed to 
provide either an operational parallel or a sequential IO module.s. What we 
did is to do a kind of expedition, and further more to investigate the 
performance of HDF5. However, at the end of the project, John Mialakes 
would like to have our parallel version into WRF release. Since I have been 
spending limited time maintaining the module even after the project ended a 
year ago. Sequential version has never been into WRF release since WRF has 
already had several sequential IO modules. I did remember I checked the 
output carefully about dim. scale at that time; I should realize the 
reverse order of dim-rank. Somehow it just slipped away from my finger.  In 
the future I need to review them again, I just don't have time now. As for 
parallelzation, WRF does parallelism in 
spatially(domain  decompositions).  Time won't affect parallelzation.
As for the last few questions you ask:
We can "extend" the data this way but just not use chunking storage in WRF.
I am not sure your last few questions. The bottom line is to store data in 
contiguous storage won't affect parallelzation at all. It just add a little 
bit more meta-data.
If you want to know more about this, you can go to 
http://hdf.ncsa.uiuc.edu/apps/WRF-ROMS to read more report about this work; 
especially  the article at 
http://hdf.ncsa.uiuc.edu/apps/WRF-ROMS/performance-report.pdf may help 
answer your questions.
Kent
At 09:32 AM 3/31/2005, you wrote:
Hi Kent:
So do you think no one will use this particular format? If so, I will 
wait for new version..
I don't know. Somebody from OU contacted with me about using WRF-PHDF5. 
But I didn't hear from them about this issue.
since its also in the sequential version, i guess the question is,  do you 
know if  WRF is planning on using either version?
BTW, in the parallel output (but not the sequential), you put each time 
step in its own group. was this to optimize parallel performance or something ?
The reason is: we are using contiguous storage for parallel HDF5 because 
of performance issue. So that's why we put each time step in to its own 
group. I hope that in the future I can have time or maybe a little 
funding to modify the WRF-HDF5 IO modules with 1.8 release.
so you use contiguous, not chunked storage. so you cant extend the data ? 
so you create a new group as each time step completes ?
how do you parellelize the tasks? does each thread have its own seperate 
region of the array?