[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[netCDF #AIQ-275071]: [netcdf-hdf] Unexpected overall file size jump



James,

By the way, it turns out that using compression even with the default chunking
works as well as compression with optimized chunking on your example, so you 
can get the same results of smaller files without even bothering with
customizing the chunk sizes.

Here's the specifics.  With your original netCDF classic file, converting to 
a netCDF-4 or netCDF-4 classic model file turns on default chunking for all 
the variables that use an unlimited dimension, and the resulting chunk sizes 
waste lots of space with some chunks that are mostly empty (missing values):

  $ ls -l Fluid_Meas.snc
  -rw-rw-r-- 1 russ ustaff 42186296 Jan 10 09:52 Fluid_Meas.snc
  $ nccopy -k4 Fluid_Meas.snc tmp.nc; ls -l tmp.nc
  -rw-rw-r-- 1 russ ustaff 95528413 Jan 10 09:53 tmp.nc

But just compressing with deflation level 1 and shuffling results
in a smaller file than the original, even using the default chunk sizes:

  $ nccopy -s -d1 -k4 Fluid_Meas.snc tmp.nc; ls -l tmp.nc
  -rw-rw-r-- 1 russ ustaff 27864220 Jan 10 09:53 tmp.nc

You can do as well with smaller custom chunk sizes, but it may not be 
worth the trouble:

  $ nccopy -s -d1 -k4 -c npoints/76375 Fluid_Meas.snc tmp.nc; ls -l tmp.nc
  -rw-rw-r-- 1 russ ustaff 27864220 Jan 10 09:53 tmp.nc

--Russ

Russ Rew                                         UCAR Unidata Program
address@hidden                      http://www.unidata.ucar.edu



Ticket Details
===================
Ticket ID: AIQ-275071
Department: Support netCDF
Priority: Normal
Status: Closed