Hi Davide,
> On May 2, 2020, at 1:06 PM, Wei-Keng Liao <wkliao@xxxxxxxxxxxxxxxx> wrote:
>
> The dump information shows there are actually 8 datasets in the file.
> Below is the start offsets, sizes, and end offsets of individual datasets .
> There is not much padding space in between the datasets.
> According to this, your file is expected to be of size 16 GB.
>
> dataset name start offset size end offset
> BS_K_linearized1 2,379 8,100,000,000 8,100,002,379
> BS_K_linearized2 12,150,006,475 3,127,549,440 15,277,555,915
Thanks Wei-kang for preparing this useful information.
These two are netCDF dimensions which in HDF5 (netCDF-4) files are stored as
HDF5 datasets. The above information indicate these HDF5 datasets are taking up
file space although they should not have any actual data. The netCDF library
only needs a specific value in the NAME attribute of these datasets. I, too,
suggest creating your files with the latest version of the netCDF library as it
may improve what kind of create dataset information it passes to the HDF5
library.
-Aleksandar