Source code for metpy.io.gini

# Copyright (c) 2015,2016,2017,2019 MetPy Developers.
# Distributed under the terms of the BSD 3-Clause License.
# SPDX-License-Identifier: BSD-3-Clause
"""Tools to process GINI-formatted products."""

import contextlib
from datetime import datetime
from enum import Enum
from io import BytesIO
from itertools import repeat  # noqa: I202
import logging
import re

import numpy as np
from xarray import Variable
from xarray.backends.common import AbstractDataStore
try:
    from xarray.core.utils import FrozenDict
except ImportError:
    from xarray.core.utils import FrozenOrderedDict as FrozenDict

from ._tools import Bits, IOBuffer, NamedStruct, open_as_needed, zlib_decompress_all_frames
from ..package_tools import Exporter

exporter = Exporter(globals())
log = logging.getLogger(__name__)


def _make_datetime(s):
    r"""Convert 7 bytes from a GINI file to a `datetime` instance."""
    year, month, day, hour, minute, second, cs = s
    if year < 70:
        year += 100
    return datetime(1900 + year, month, day, hour, minute, second, 10000 * cs)


def _scaled_int(s):
    r"""Convert a 3 byte string to a signed integer value."""
    # Get leftmost bit (sign) as 1 (if 0) or -1 (if 1)
    sign = 1 - ((s[0] & 0x80) >> 6)

    # Combine remaining bits
    int_val = (((s[0] & 0x7f) << 16) | (s[1] << 8) | s[2])
    log.debug('Source: %s Int: %x Sign: %d', ' '.join(hex(c) for c in s), int_val, sign)

    # Return scaled and with proper sign
    return (sign * int_val) / 10000.


def _name_lookup(names):
    r"""Create an io helper to convert an integer to a named value."""
    mapper = dict(zip(range(len(names)), names))

    def lookup(val):
        return mapper.get(val, 'Unknown')
    return lookup


class GiniProjection(Enum):
    r"""Represents projection values in GINI files."""

    mercator = 1
    lambert_conformal = 3
    polar_stereographic = 5


[docs]@exporter.export class GiniFile(AbstractDataStore): """A class that handles reading the GINI format satellite images from the NWS. This class attempts to decode every byte that is in a given GINI file. Notes ----- The internal data structures that things are decoded into are subject to change. """ missing = 255 wmo_finder = re.compile('(T\\w{3}\\d{2})[\\s\\w\\d]+\\w*(\\w{3})\r\r\n') crafts = ['Unknown', 'Unknown', 'Miscellaneous', 'JERS', 'ERS/QuikSCAT', 'POES/NPOESS', 'Composite', 'DMSP', 'GMS', 'METEOSAT', 'GOES-7', 'GOES-8', 'GOES-9', 'GOES-10', 'GOES-11', 'GOES-12', 'GOES-13', 'GOES-14', 'GOES-15', 'GOES-16'] sectors = ['NH Composite', 'East CONUS', 'West CONUS', 'Alaska Regional', 'Alaska National', 'Hawaii Regional', 'Hawaii National', 'Puerto Rico Regional', 'Puerto Rico National', 'Supernational', 'NH Composite', 'Central CONUS', 'East Floater', 'West Floater', 'Central Floater', 'Polar Floater'] channels = ['Unknown', 'Visible', 'IR (3.9 micron)', 'WV (6.5/6.7 micron)', 'IR (11 micron)', 'IR (12 micron)', 'IR (13 micron)', 'IR (1.3 micron)', 'Reserved', 'Reserved', 'Reserved', 'Reserved', 'Reserved', 'LI (Imager)', 'PW (Imager)', 'Surface Skin Temp (Imager)', 'LI (Sounder)', 'PW (Sounder)', 'Surface Skin Temp (Sounder)', 'CAPE', 'Land-sea Temp', 'WINDEX', 'Dry Microburst Potential Index', 'Microburst Day Potential Index', 'Convective Inhibition', 'Volcano Imagery', 'Scatterometer', 'Cloud Top', 'Cloud Amount', 'Rainfall Rate', 'Surface Wind Speed', 'Surface Wetness', 'Ice Concentration', 'Ice Type', 'Ice Edge', 'Cloud Water Content', 'Surface Type', 'Snow Indicator', 'Snow/Water Content', 'Volcano Imagery', 'Reserved', 'Sounder (14.71 micron)', 'Sounder (14.37 micron)', 'Sounder (14.06 micron)', 'Sounder (13.64 micron)', 'Sounder (13.37 micron)', 'Sounder (12.66 micron)', 'Sounder (12.02 micron)', 'Sounder (11.03 micron)', 'Sounder (9.71 micron)', 'Sounder (7.43 micron)', 'Sounder (7.02 micron)', 'Sounder (6.51 micron)', 'Sounder (4.57 micron)', 'Sounder (4.52 micron)', 'Sounder (4.45 micron)', 'Sounder (4.13 micron)', 'Sounder (3.98 micron)', 'Sounder (3.74 micron)', 'Sounder (Visible)'] prod_desc_fmt = NamedStruct([('source', 'b'), ('creating_entity', 'b', _name_lookup(crafts)), ('sector_id', 'b', _name_lookup(sectors)), ('channel', 'b', _name_lookup(channels)), ('num_records', 'H'), ('record_len', 'H'), ('datetime', '7s', _make_datetime), ('projection', 'b', GiniProjection), ('nx', 'H'), ('ny', 'H'), ('la1', '3s', _scaled_int), ('lo1', '3s', _scaled_int) ], '>', 'ProdDescStart') lc_ps_fmt = NamedStruct([('reserved', 'b'), ('lov', '3s', _scaled_int), ('dx', '3s', _scaled_int), ('dy', '3s', _scaled_int), ('proj_center', 'b')], '>', 'LambertOrPolarProjection') mercator_fmt = NamedStruct([('resolution', 'b'), ('la2', '3s', _scaled_int), ('lo2', '3s', _scaled_int), ('di', 'H'), ('dj', 'H') ], '>', 'MercatorProjection') prod_desc2_fmt = NamedStruct([('scanning_mode', 'b', Bits(3)), ('lat_in', '3s', _scaled_int), ('resolution', 'b'), ('compression', 'b'), ('version', 'b'), ('pdb_size', 'H'), ('nav_cal', 'b')], '>', 'ProdDescEnd') nav_fmt = NamedStruct([('sat_lat', '3s', _scaled_int), ('sat_lon', '3s', _scaled_int), ('sat_height', 'H'), ('ur_lat', '3s', _scaled_int), ('ur_lon', '3s', _scaled_int)], '>', 'Navigation')
[docs] def __init__(self, filename): r"""Create an instance of `GiniFile`. Parameters ---------- filename : str or file-like object If str, the name of the file to be opened. Gzip-ed files are recognized with the extension ``'.gz'``, as are bzip2-ed files with the extension ``'.bz2'`` If `filename` is a file-like object, this will be read from directly. """ fobj = open_as_needed(filename) # Just read in the entire set of data at once with contextlib.closing(fobj): self._buffer = IOBuffer.fromfile(fobj) # Pop off the WMO header if we find it self.wmo_code = '' self._process_wmo_header() log.debug('First wmo code: %s', self.wmo_code) # Decompress the data if necessary, and if so, pop off new header log.debug('Length before decompression: %s', len(self._buffer)) self._buffer = IOBuffer(self._buffer.read_func(zlib_decompress_all_frames)) log.debug('Length after decompression: %s', len(self._buffer)) # Process WMO header inside compressed data if necessary self._process_wmo_header() log.debug('2nd wmo code: %s', self.wmo_code) # Read product description start start = self._buffer.set_mark() #: :desc: Decoded first section of product description block #: :type: namedtuple self.prod_desc = self._buffer.read_struct(self.prod_desc_fmt) log.debug(self.prod_desc) #: :desc: Decoded geographic projection information #: :type: namedtuple self.proj_info = None # Handle projection-dependent parts if self.prod_desc.projection in (GiniProjection.lambert_conformal, GiniProjection.polar_stereographic): self.proj_info = self._buffer.read_struct(self.lc_ps_fmt) elif self.prod_desc.projection == GiniProjection.mercator: self.proj_info = self._buffer.read_struct(self.mercator_fmt) else: log.warning('Unknown projection: %d', self.prod_desc.projection) log.debug(self.proj_info) # Read the rest of the guaranteed product description block (PDB) #: :desc: Decoded second section of product description block #: :type: namedtuple self.prod_desc2 = self._buffer.read_struct(self.prod_desc2_fmt) log.debug(self.prod_desc2) if self.prod_desc2.nav_cal not in (0, -128): # TODO: See how GEMPAK/MCIDAS parses # Only warn if there actually seems to be useful navigation data if self._buffer.get_next(self.nav_fmt.size) != b'\x00' * self.nav_fmt.size: log.warning('Navigation/Calibration unhandled: %d', self.prod_desc2.nav_cal) if self.prod_desc2.nav_cal in (1, 2): self.navigation = self._buffer.read_struct(self.nav_fmt) log.debug(self.navigation) # Catch bad PDB with size set to 0 if self.prod_desc2.pdb_size == 0: log.warning('Adjusting bad PDB size from 0 to 512.') self.prod_desc2 = self.prod_desc2._replace(pdb_size=512) # Jump past the remaining empty bytes in the product description block self._buffer.jump_to(start, self.prod_desc2.pdb_size) # Read the actual raster--unless it's PNG compressed, in which case that happens later blob = self._buffer.read(self.prod_desc.num_records * self.prod_desc.record_len) # Check for end marker end = self._buffer.read(self.prod_desc.record_len) if end != b''.join(repeat(b'\xff\x00', self.prod_desc.record_len // 2)): log.warning('End marker not as expected: %s', end) # Check to ensure that we processed all of the data if not self._buffer.at_end(): if not blob: log.debug('No data read yet, trying to decompress remaining data as an image.') from matplotlib.image import imread blob = (imread(BytesIO(self._buffer.read())) * 255).astype('uint8') else: log.warning('Leftover unprocessed data beyond EOF marker: %s', self._buffer.get_next(10)) self.data = np.array(blob).reshape((self.prod_desc.ny, self.prod_desc.nx))
def _process_wmo_header(self): """Read off the WMO header from the file, if necessary.""" data = self._buffer.get_next(64).decode('utf-8', 'ignore') match = self.wmo_finder.search(data) if match: self.wmo_code = match.groups()[0] self.siteID = match.groups()[-1] self._buffer.skip(match.end()) def __str__(self): """Return a string representation of the product.""" parts = [self.__class__.__name__ + ': {0.creating_entity} {0.sector_id} {0.channel}', 'Time: {0.datetime}', 'Size: {0.ny}x{0.nx}', 'Projection: {0.projection.name}', 'Lower Left Corner (Lon, Lat): ({0.lo1}, {0.la1})', 'Resolution: {1.resolution}km'] return '\n\t'.join(parts).format(self.prod_desc, self.prod_desc2) def _make_proj_var(self): proj_info = self.proj_info prod_desc2 = self.prod_desc2 attrs = {'earth_radius': 6371200.0} if self.prod_desc.projection == GiniProjection.lambert_conformal: attrs['grid_mapping_name'] = 'lambert_conformal_conic' attrs['standard_parallel'] = prod_desc2.lat_in attrs['longitude_of_central_meridian'] = proj_info.lov attrs['latitude_of_projection_origin'] = prod_desc2.lat_in elif self.prod_desc.projection == GiniProjection.polar_stereographic: attrs['grid_mapping_name'] = 'polar_stereographic' attrs['straight_vertical_longitude_from_pole'] = proj_info.lov attrs['latitude_of_projection_origin'] = -90 if proj_info.proj_center else 90 attrs['standard_parallel'] = 60.0 # See Note 2 for Table 4.4A in ICD elif self.prod_desc.projection == GiniProjection.mercator: attrs['grid_mapping_name'] = 'mercator' attrs['longitude_of_projection_origin'] = self.prod_desc.lo1 attrs['latitude_of_projection_origin'] = self.prod_desc.la1 attrs['standard_parallel'] = prod_desc2.lat_in else: raise NotImplementedError( 'Unhandled GINI Projection: {}'.format(self.prod_desc.projection)) return 'projection', Variable((), 0, attrs) def _make_time_var(self): base_time = self.prod_desc.datetime.replace(hour=0, minute=0, second=0, microsecond=0) offset = self.prod_desc.datetime - base_time time_var = Variable((), data=offset.seconds + offset.microseconds / 1e6, attrs={'units': 'seconds since ' + base_time.isoformat()}) return 'time', time_var def _get_proj_and_res(self): import pyproj proj_info = self.proj_info prod_desc2 = self.prod_desc2 kwargs = {'a': 6371200.0, 'b': 6371200.0} if self.prod_desc.projection == GiniProjection.lambert_conformal: kwargs['proj'] = 'lcc' kwargs['lat_0'] = prod_desc2.lat_in kwargs['lon_0'] = proj_info.lov kwargs['lat_1'] = prod_desc2.lat_in kwargs['lat_2'] = prod_desc2.lat_in dx, dy = proj_info.dx, proj_info.dy elif self.prod_desc.projection == GiniProjection.polar_stereographic: kwargs['proj'] = 'stere' kwargs['lon_0'] = proj_info.lov kwargs['lat_0'] = -90 if proj_info.proj_center else 90 kwargs['lat_ts'] = 60.0 # See Note 2 for Table 4.4A in ICD kwargs['x_0'] = False # Easting kwargs['y_0'] = False # Northing dx, dy = proj_info.dx, proj_info.dy elif self.prod_desc.projection == GiniProjection.mercator: kwargs['proj'] = 'merc' kwargs['lat_0'] = self.prod_desc.la1 kwargs['lon_0'] = self.prod_desc.lo1 kwargs['lat_ts'] = prod_desc2.lat_in kwargs['x_0'] = False # Easting kwargs['y_0'] = False # Northing dx, dy = prod_desc2.resolution, prod_desc2.resolution return pyproj.Proj(**kwargs), dx, dy def _make_coord_vars(self): proj, dx, dy = self._get_proj_and_res() # Get projected location of lower left point x0, y0 = proj(self.prod_desc.lo1, self.prod_desc.la1) # Coordinate variable for x xlocs = x0 + np.arange(self.prod_desc.nx) * (1000. * dx) attrs = {'units': 'm', 'long_name': 'x coordinate of projection', 'standard_name': 'projection_x_coordinate'} x_var = Variable(('x',), xlocs, attrs) # Now y--Need to flip y because we calculated from the lower left corner, # but the raster data is stored with top row first. ylocs = (y0 + np.arange(self.prod_desc.ny) * (1000. * dy))[::-1] attrs = {'units': 'm', 'long_name': 'y coordinate of projection', 'standard_name': 'projection_y_coordinate'} y_var = Variable(('y',), ylocs, attrs) # Get the two-D lon,lat grid as well x, y = np.meshgrid(xlocs, ylocs) lon, lat = proj(x, y, inverse=True) lon_var = Variable(('y', 'x'), data=lon, attrs={'long_name': 'longitude', 'units': 'degrees_east'}) lat_var = Variable(('y', 'x'), data=lat, attrs={'long_name': 'latitude', 'units': 'degrees_north'}) return [('x', x_var), ('y', y_var), ('lon', lon_var), ('lat', lat_var)] # FIXME: Work around xarray <=0.10.3 docstring for load angering sphinx # That's the only reason this exists.
[docs] def load(self): """ Load the variables and attributes simultaneously. A centralized loading function makes it easier to create data stores that do automatic encoding/decoding. For example:: class SuffixAppendingDataStore(AbstractDataStore): def load(self): variables, attributes = AbstractDataStore.load(self) variables = {'%s_suffix' % k: v for k, v in iteritems(variables)} attributes = {'%s_suffix' % k: v for k, v in iteritems(attributes)} return variables, attributes This function will be called anytime variables or attributes are requested, so care should be taken to make sure its fast. """ return super().load()
[docs] def get_variables(self): """Get all variables in the file. This is used by `xarray.open_dataset`. """ variables = [self._make_time_var()] proj_var_name, proj_var = self._make_proj_var() variables.append((proj_var_name, proj_var)) variables.extend(self._make_coord_vars()) # Now the data name = self.prod_desc.channel if '(' in name: name = name.split('(')[0].rstrip() missing_val = self.missing attrs = {'long_name': self.prod_desc.channel, 'missing_value': missing_val, 'coordinates': 'y x time', 'grid_mapping': proj_var_name} data_var = Variable(('y', 'x'), data=np.ma.array(self.data, mask=self.data == missing_val), attrs=attrs) variables.append((name, data_var)) return FrozenDict(variables)
[docs] def get_attrs(self): """Get the global attributes. This is used by `xarray.open_dataset`. """ return FrozenDict(satellite=self.prod_desc.creating_entity, sector=self.prod_desc.sector_id)
[docs] def get_dimensions(self): """Get the file's dimensions. This is used by `xarray.open_dataset`. """ return FrozenDict(x=self.prod_desc.nx, y=self.prod_desc.ny)