Source code for landlab.data_record.data_record

#!/usr/bin/env python3

import numpy as np
import xarray as xr


[docs] class DataRecord: """Data structure to store variables in time and/or space dimensions. This class uses a xarray Dataset to store variables. This datastructure is located at the property ``dataset``. The DataRecord expands xarray Dataset with additional attributes and functions, including the ability to aggregate values on Landlab grid elements. DataRecord uses the concept of an "item", a physical thing that is located on the grid and has some properties, stored as variables. DataRecord tracks variables through time, variables associated with a Landlab grid across items, or both. Thus data variables can vary along one or both of the following dimensions: - time (model time) - item_id: variables can characterize a set of items (each identified by an individual id) that reside on the grid. If an item or set of items is defined, each item must be defined by the grid element and the element id at which it resides, e.g.: grid_element = 'node' element_id = 9. When items are defined, each item is given a unique id and the underlying Dataset uses a dimension "item_id". **Items are assigned ids beginning with 0 followed by consecutively increasing integers.** Examples: - the variable 'mean_elevation' characterizes the grid and varies with time, - the variable 'clast__rock_type' characterizes a set of items (clasts) and varies with item_id, - the variable 'clast__size' can vary with both time and item_id In the above case, `grid_element` and `element_id` are default data variables (in addition to any user-specified variables). For each item, `element_id` must be less than the number of this item's grid_element that exist on the grid or be one of the dummy element values. For example, if the grid has 100 links, and no dummy link values are indicated, then, no item can live at link 100 or link -3 because only links 0 to 99 exist in this example. Anything that the DataRecord keeps track of is considered a "record", whether it uses one or both of the two standard dimensions (**time** and **item_id**). DataRecord provides two method to assist with adding new records. The method ``add_item`` should be used when no new variables are being added. The method ``add_record`` should be used when new variables are being added or when a variable is only tracked over the **time** dimension. """ _name = "DataRecord"
[docs] def __init__( self, grid, dummy_elements=None, time=None, items=None, data_vars=None, attrs=None, ): """ Parameters ---------- grid : ModelGrid dummy_elements : dict Dictionary indicating valid values for dummy grid elements. For example, if you need an "exit" off of a grid with 100 links, you could indicate `dummy_elements = {"link": [9999]}` to set a link id of 9999 as a dummy link. Multiple dummy elements are possible and we recommend using values larger than the number of grid elements for the dummy values. time : list or 1-D array of float or int (optional) The initial time(s) to add to the record. A time dimension is not created if the value is 'None' (default). items : dict (optional) Generic items that live on grid elements. No item is created if the value is 'None' (default). Otherwise, dictionary describes the position of generic items on the grid. The structure is: .. code-block:: python {"grid_element": [grid_element], "element_id": [element_id]} where: - [grid_element] is a str or number-of-items-long array containing strings of the grid element(s) on which the items live. Valid locations depend on the grid type. If provided as a string it is assumed that all items live on the same type of grid element. - [element_id] is an array of integers identifying the grid element ID on which each item resides. An example argument would be: .. code-block:: python { "grid_element": numpy.array(["node"], ["node"], ["link"]), "element_id": numpy.array([1], [5], [1]), } data_vars : dict (optional) Dictionary of the data variables to be recorded. The structure is: .. code-block:: python { "variable_name_1": (["dimensions"], variable_data_1), "variable_name_2": (["dimensions"], variable_data_2), } where: - 'variable_name...' is a string of the variable name (label) - ['dimensions'] is the dimension(s) over which the variable exists: can be ['time'], ['item_id'] or ['item_id', 'time']. - variable_data is an array containing the data, its size must match that of the variable dimension(s). attrs : dict (optional) Dictionary of global attributes on the DataRecord (metadata). Example: {'time_units' : 'y'} Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with time as the only dimension: >>> dr1 = DataRecord( ... grid, ... time=[0.0], ... data_vars={"mean_elevation": (["time"], np.array([100]))}, ... attrs={"time_units": "y"}, ... ) DataRecord builds off of xarray Dataset, a multi-dimensional, in memory, array database. Dataset implements the mapping interface with keys given by variable names and values given by DataArray objects for each variable name. A DataRecord can have dimensions 'time' and/or 'item_id'. The xarray Dataset is stored in the public attribute ``dataset``. Coordinates are one dimensional arrays used for label-based indexing. DataRecord inherits all the methods and attributes from ``xarray.Dataset``. >>> dr1.dataset.to_dataframe() mean_elevation time 0.0 100 >>> dr1.dataset.time.values array([ 0.]) >>> dr1.variable_names ['mean_elevation'] >>> dr1.dataset["mean_elevation"].values array([100]) >>> list(dr1.dataset.attrs.items()) [('time_units', 'y')] >>> list(dr1.dataset.attrs.items()) [('time_units', 'y')] Example of a DataRecord with item_id as the only dimension: >>> my_items2 = { ... "grid_element": np.array(("node", "link"), dtype=str), ... "element_id": np.array([1, 3]), ... } >>> dr2 = DataRecord(grid, items=my_items2) Note that both arrays (grid_element and element_id) have 1 dimension as they only vary along the dimension 'item_id'. >>> dr2.dataset.to_dataframe()[["grid_element", "element_id"]] grid_element element_id item_id 0 node 1 1 link 3 Example of a DataRecord with dimensions time and item_id: >>> my_items3 = { ... "grid_element": np.array([["node"], ["link"]]), ... "element_id": np.array([[1], [3]]), ... } >>> dr3 = DataRecord(grid, time=[0.0], items=my_items3) Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> dr3.dataset.to_dataframe()[["grid_element", "element_id"]] grid_element element_id item_id time 0 0.0 node 1 1 0.0 link 3 """ # save a reference to the grid self._grid = grid # depending on the grid type, permitted locations for items vary self._permitted_locations = self._grid.groups # save dummy elements reference # check dummies and reformat into {"node": [0, 1, 2]} self._dummy_elements = dummy_elements or {} for at in self._permitted_locations: for item in self._dummy_elements.get(at, []): if (item < self._grid[at].size) and (item >= 0): raise ValueError(f"Dummy id {at} {item} invalid") # set initial time coordinates, if any if isinstance(time, (list, np.ndarray)): self._times = np.array(time) self._number_of_times = len(self._times) elif time is not None: raise TypeError("time must be a list or numpy array") # set initial items, if any if items is not None: try: items.keys() except AttributeError as exc: # items is not a dict raise TypeError( "You must provide an `items` dictionary " "(see documentation for required format)" ) from exc try: _grid_elements, _element_ids = ( items["grid_element"], items["element_id"], ) except KeyError as exc: # grid_element and/or element_id not provided raise TypeError( "You must provide an `items` dictionary," "(see documentation for required format)" ) from exc self._number_of_items = len(_element_ids) if len(_grid_elements) != self._number_of_items: if isinstance(_grid_elements, str): pass else: raise ValueError( "The number of grid_element passed " "to DataRecord must be 1 or equal " "to the number of element_id." ) # check that grid_element and element_id exist on the grid and # have valid format: _grid_elements, _element_ids = self._check_grid_element_and_id( _grid_elements, _element_ids ) # check that element IDs do not exceed number of elements # on the grid: self._check_element_id_values(_grid_elements, _element_ids) # create coordinates for the dimension 'item_id': self._item_ids = np.array(range(self._number_of_items), dtype=int) # create initial dictionaries of variables: if time is not None: data_vars_dict = { "grid_element": (["item_id", "time"], _grid_elements), "element_id": ( ["item_id", "time"], _element_ids, {"dtype": int}, ), } coords = {"time": self._times, "item_id": self._item_ids} else: # no time data_vars_dict = { "grid_element": (["item_id"], _grid_elements), "element_id": (["item_id"], _element_ids, {"dtype": int}), } coords = {"item_id": self._item_ids} else: # no items, initial dictionary of variables is empty: data_vars_dict = {} if time is not None: coords = {"time": self._times} else: # no item and no time = no dimension coords = {} # set variables, if any if data_vars is not None: try: # check format (dict) data_vars.keys() except AttributeError as exc: raise TypeError( "Data variables (data_vars) passed to " "DataRecord must be a dictionary (see " "documentation for valid structure)" ) from exc for key in data_vars.keys(): # check dict structure and dims: if data_vars[key][0] not in ( ["time"], ["item_id"], ["time", "item_id"], ["item_id", "time"], ): raise ValueError( "Data variable dimensions must be " "time and/or item_id" ) # create complete dictionary of variables # (= initial data_vars_dict + additional user-defined data_vars): data_vars_dict.update(data_vars) # set attributes, if any attrs = attrs or {} if not isinstance(attrs, dict): raise TypeError( "Attributes (attrs) passed to DataRecord" "must be a dictionary" ) # create an xarray Dataset: self._dataset = xr.Dataset(data_vars=data_vars_dict, coords=coords, attrs=attrs)
def _check_grid_element_and_id(self, grid_element, element_id): """Check the location and size of grid_element and element_id.""" if isinstance(grid_element, str): # create list of grid_element for all items ge_name = grid_element if hasattr(self, "_number_of_times"): # if time grid_element = np.array( np.empty( (self._number_of_items, self._number_of_times), dtype=object ) ) if element_id.shape != grid_element.shape: element_id = np.broadcast_to(element_id, grid_element.shape) else: # no time grid_element = np.array( np.empty((self._number_of_items,), dtype=object) ) grid_element.fill(ge_name) # verify all grid elements are valid. for loc in grid_element.flatten(): if loc not in self._permitted_locations: raise ValueError( "One or more of the grid elements" " provided is/are not permitted location" " for this grid type" ) return grid_element, element_id def _check_element_id_values(self, grid_element, element_id): """Check that element_id values are valid.""" for at in self._permitted_locations: max_size = self._grid[at].size # this needs to work with 2d arrays (rows, col = np.where (so grid # element always needs to be at least 2d.)) ind = np.nonzero(grid_element == at) selected_elements = element_id[ind] if selected_elements.size > 0: dummy_values = self._dummy_elements.get(at, []) index_values = np.arange(0, max_size) valid_values = np.concatenate((dummy_values, index_values)) valid_elements = np.isin(selected_elements, valid_values) if not np.all(valid_elements): raise ValueError("Invalid element_ids provided.") if not np.issubdtype(element_id.dtype, np.integer): raise ValueError( "You have passed a non-integer element_id to " "DataRecord, this is not permitted" )
[docs] def add_record(self, time=None, item_id=None, new_item_loc=None, new_record=None): """Add a new record to the DataRecord. Unlike add_item, this method can support adding records that include new variables to the DataRecord. It can also support adding records that do not include time. Parameters ---------- time : list or 1-D array of float or int Time step at which the record is to be added. item_id : list or 1-D array of int (optional) ID of the item to which the new record relates. new_item_loc: dict (optional) Dictionary of the new item location. If the new record is a change in the item location (grid_element and/or element_id), this field must be provided as: .. code-block:: python {"grid_element": [grid_element], "element_id": [element_id]} Both must be provided even if only one is being changed. new_record : dict Dictionary containing the new record. Structure should be: {'variable_name_1' : (['dimensions'], variable_data_1)} with: - 'variable_name_1' : name of the (potentially new) variable - ['dimensions'] : dimension(s) along which the new record varies; can be ['time'], ['item_id] or ['item_id', 'time'] - variable_data_1 : new data array, size must match the variable dimension(s) Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with dimensions time and item_id: >>> my_items3 = { ... "grid_element": np.array([["node"], ["link"]]), ... "element_id": np.array([[1], [3]]), ... } Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> dr3 = DataRecord(grid, time=[0.0], items=my_items3) Records relating to pre-existing items can be added to the DataRecord using the method 'add_record': >>> dr3.add_record( ... time=[2.0], ... item_id=[0], ... new_item_loc={ ... "grid_element": np.array([["node"]]), ... "element_id": np.array([[6]]), ... }, ... new_record={"item_size": (["item_id", "time"], np.array([[0.2]]))}, ... ) >>> dr3.dataset["element_id"].values array([[ 1., 6.], [ 3., nan]]) >>> dr3.get_data([2.0], [0], "item_size") array([ 0.2]) The 'add_record' method can also be used to add a non item-related record: >>> dr3.add_record(time=[50.0], new_record={"mean_elev": (["time"], [110])}) >>> dr3.dataset["mean_elev"].to_dataframe() mean_elev time 0.0 NaN 2.0 NaN 50.0 110.0 """ if time is not None: try: # check that time is a dim of the DataRecord self._dataset["time"] except KeyError as exc: raise KeyError("This DataRecord does not record time") from exc if not isinstance(time, (list, np.ndarray)): # check input type raise TypeError( "You have passed a time that is" " not permitted, must be list or array" ) else: if item_id is not None: try: # check that DataRecord holds items self._dataset["item_id"] except KeyError as exc: raise KeyError("This DataRecord does not hold items") from exc try: # check that item_id is list or array len(item_id) except TypeError as exc: raise TypeError("item_id must be a list or a 1D array") from exc if not all(i in self._dataset["item_id"].values for i in item_id): # check that item_id already exist raise ValueError( "One or more of the value(s) you " "passed as item_id is/are not " "currently in the DataRecord. Change" " the input values create a new item" "using the method add_item" ) coords_to_add = {"time": np.array(time), "item_id": item_id} # if item location is changed by this new record, check # that both grid_element and element_id are provided: if new_item_loc is not None: try: new_grid_element = new_item_loc["grid_element"] new_element_id = new_item_loc["element_id"] except KeyError as exc: raise KeyError( "You must provide a " "new_item_loc dictionary with both " "grid_element and element_id" ) from exc # check that grid_element and element_id exist # on the grid and have valid format: ( new_grid_element, new_element_id, ) = self._check_grid_element_and_id( new_grid_element, new_element_id ) # check that element IDs do not exceed number # of elements on this grid: self._check_element_id_values(new_grid_element, new_element_id) _new_data_vars = { "grid_element": (["item_id", "time"], new_grid_element), "element_id": (["item_id", "time"], new_element_id), } else: # new_item_loc is `None` _new_data_vars = {} else: # no item coords_to_add = {"time": np.array(time)} _new_data_vars = {} else: # no time if item_id is not None: if not all(i in self._dataset["item_id"].values for i in item_id): # check that item_id already exist raise ValueError( "One or more of the value(s) you " "passed as item_id is/are not " "currently in the DataRecord. Change" " the input values create a new item" "using the method add_item" ) coords_to_add = {"item_id": np.array(item_id)} _new_data_vars = {} # no time so if item location needs to be changed, # user should use set_data if new_item_loc is not None: raise ValueError( "Use the method set_data to change the " "location of an item in this DataRecord" ) else: # no item _new_data_vars = {} coords_to_add = {} if new_record is not None: # add new_record to dict of variables to add _new_data_vars.update(new_record) # create dataset of new record ds_to_add = xr.Dataset(data_vars=_new_data_vars, coords=coords_to_add) # merge new record and original dataset self._dataset = xr.merge((self._dataset, ds_to_add), compat="no_conflicts")
[docs] def add_item(self, time=None, new_item=None, new_item_spec=None): """Add new item(s) to the current DataRecord. Parameters ---------- time : list or 1-D array of float or int Time step at which the items are to be added. new_item : dict Structure is: .. code-block:: python {"grid_element": [grid_element], "element_id": [element_id]} where: - [grid_element] is str or number-of-items long array containing strings of the grid element(s) on which the items live. Valid locations depend on the grid type. If provided as a string it is assumed that all items live on the same type of grid element. - [element_id] is an array of integers identifying the grid element ID on which each item resides. An example argument would be: .. code-block:: python { "grid_element": numpy.array([["node"], ["node"], ["link"]]), "element_id": numpy.array([[1], [5], [1]]), } new_item_spec : dict (optional) Dictionary containing any data variables (other than 'grid_element' and 'element_id') relating to the new item(s) to be added. Structure is: .. code-block:: python {"variable_name_1": (["dimensions"], variable_data_1)} where: - 'variable_name_1' : name of the (potentially new) variable - ['dimensions'] : dimension(s) along which the new record varies; can be ['time'], ['item_id] or ['item_id', 'time'] - variable_data_1 : new data array, size must match the variable dimension(s) Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with dimensions time and item_id: >>> my_items3 = { ... "grid_element": np.array([["node"], ["link"]]), ... "element_id": np.array([[1], [3]]), ... } Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> dr3 = DataRecord(grid, time=[0.0], items=my_items3) Items can be added to a DataRecord that already holds similar items, using the method 'add_item': >>> dr3.add_item( ... time=[1.0], ... new_item={ ... "grid_element": np.array([["node"], ["node"]]), ... "element_id": np.array([[4], [4]]), ... }, ... new_item_spec={"size": (["item_id", "time"], [[10], [5]])}, ... ) Two items have been added at a new timestep 1.0: >>> dr3.number_of_items 4 >>> dr3.time_coordinates [0.0, 1.0] If a data variable is also added with the new items ('size' in this example), the values for this variable are filled with 'nan' for the pre-existing items: >>> dr3.dataset["size"][:, 1].values array([ nan, nan, 10., 5.]) The previous line calls the values of the variable 'size', for all items, at time=1; the first two items don't have a value for the variable 'size'. """ if time is None and "time" in self._dataset["grid_element"].coords: raise ValueError( "The items previously defined in this DataRecord" ' have dimensions "time" and "item_id", ' 'you must provide a "time" for the new item(s)' ) if not isinstance(new_item, dict): raise TypeError( "You must provide an new_item dictionary " "(see documentation for required format)" ) try: # check that dict contains correct entries _grid_elements, _element_ids = ( new_item["grid_element"], new_item["element_id"], ) except KeyError as exc: raise KeyError( "You must provide a new_item dictionary " "(see documentation for required format)" ) from exc number_of_new_items = len(new_item["element_id"]) # first id of new item = last item in existing datarecord+1 new_first_item_id = self._dataset["item_id"][-1].values + 1 new_item_ids = np.array( range(new_first_item_id, new_first_item_id + number_of_new_items) ) if time is not None: try: self._dataset["time"] except KeyError as exc: raise KeyError("This DataRecord does not record time") from exc if not isinstance(time, (list, np.ndarray)): raise TypeError( "You have passed a time that is not " "permitted, must be list or a 1-D array" ) else: coords_to_add = { "time": np.array(time), "item_id": np.array(new_item_ids), } # check that grid_element and element_id exist # on the grid and have valid format _grid_elements, _element_ids = self._check_grid_element_and_id( _grid_elements, _element_ids ) # check that element IDs do not exceed number # of elements on this grid self._check_element_id_values(_grid_elements, _element_ids) data_vars_dict = { "grid_element": (["item_id", "time"], _grid_elements), "element_id": (["item_id", "time"], _element_ids), } else: # no time coords_to_add = {"item_id": np.array(new_item_ids)} # check that grid_element and element_id exist on # the grid and have valid format: _grid_elements, _element_ids = self._check_grid_element_and_id( _grid_elements, _element_ids ) # check that element IDs do not exceed number of # elements on this grid self._check_element_id_values(_grid_elements, _element_ids) data_vars_dict = { "grid_element": (["item_id"], _grid_elements), "element_id": (["item_id"], _element_ids), } # other variables: if new_item_spec is not None: data_vars_dict.update(new_item_spec) # Dataset of new record: ds_to_add = xr.Dataset(data_vars=data_vars_dict, coords=coords_to_add) # Merge new record and original dataset: self._dataset = xr.merge((self._dataset, ds_to_add), compat="no_conflicts")
[docs] def get_data(self, time=None, item_id=None, data_variable=None): """Get the value of a variable at a model time and/or for an item. Parameters ---------- time : list or 1-D array of float or int (optional) The time coordinate of the record to get. item_id : list or 1-D array of int (optional) The item id of the record to get. data_variable : string The label of the variable to get. Returns ------- object The value of *variable* at *time* and/or for *item_id*. The type of the returned object is dependent on the type of the variable value. Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with dimensions time and item_id: >>> my_items4 = { ... "grid_element": "node", ... "element_id": np.array([[1], [3], [3], [7]]), ... } Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> my_data4 = { ... "item_size": ( ... ["item_id", "time"], ... np.array([[0.3], [0.4], [0.8], [0.4]]), ... ) ... } >>> dr4 = DataRecord(grid, time=[50.0], items=my_items4, data_vars=my_data4) >>> dr4.get_data([50.0], [2], "element_id") array([3]) >>> dr4.get_data(time=[50.0], data_variable="item_size") array([ 0.3, 0.4, 0.8, 0.4]) >>> dr4.get_data(item_id=[1, 2], data_variable="grid_element") array([['node'], ['node']], dtype=object) """ try: self._dataset[data_variable] except KeyError as exc: raise KeyError( f"the variable {data_variable!r} is not in the DataRecord" ) from exc if time is None: if item_id is None: return self._dataset[data_variable].values else: try: self._dataset["item_id"] except KeyError as exc: raise KeyError("This DataRecord does not hold items") from exc try: len(item_id) except TypeError as exc: raise TypeError("item_id must be a list or a 1-D array") from exc try: self._dataset["item_id"].values[item_id] except IndexError as exc: raise IndexError( "The item_id you passed does not exist " "in this DataRecord" ) from exc return self._dataset.isel(item_id=item_id)[data_variable].values else: # time is not None try: self._dataset["time"] except KeyError as exc: raise KeyError("This DataRecord does not record time") from exc try: len(time) except TypeError as exc: raise TypeError("time must be a list or a 1-D array") from exc try: time_index = int(self.time_coordinates.index(time[0])) except ValueError as exc: raise IndexError( "The time you passed is not currently" " in the DataRecord, you must change the value" " you pass or first create the new time " " coordinate using the add_record method" ) from exc if item_id is None: return self._dataset.isel(time=time_index)[data_variable].values else: try: self._dataset["item_id"] except KeyError as exc: raise KeyError("This DataRecord does not hold items") from exc try: len(item_id) except TypeError as exc: raise TypeError("item_id must be a list or a 1-D array") from exc try: self._dataset["item_id"].values[item_id] except IndexError as exc: raise IndexError( "The item_id you passed does not exist " "in this DataRecord" ) from exc return self._dataset.isel(time=time_index, item_id=item_id)[ data_variable ].values
[docs] def set_data(self, time=None, item_id=None, data_variable=None, new_value=np.nan): """Set a variable value at a model time and/or an item to a new value. The value of only one variable can be changed at a time using this method. Parameters ---------- time : list or 1-D array of float or int The time coordinate of the record to set. item_id : list or 1-D array of int The item id of the record to set. data_variable : string The label of the variable to set. new_value : list or 1-D array The new value to give to the variable data. Returns ------- DataRecord with updated data. Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with dimensions time and item_id: >>> my_items4 = { ... "grid_element": "node", ... "element_id": np.array([[1], [3], [3], [7]]), ... } Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> my_data4 = { ... "item_size": ( ... ["item_id", "time"], ... np.array([[0.3], [0.4], [0.8], [0.4]]), ... ) ... } >>> dr4 = DataRecord(grid, time=[50.0], items=my_items4, data_vars=my_data4) >>> dr4.dataset["item_size"].values array([[ 0.3], [ 0.4], [ 0.8], [ 0.4]]) >>> dr4.set_data([50.0], [2], "item_size", [0.5]) >>> dr4.dataset["item_size"].values array([[ 0.3], [ 0.4], [ 0.5], [ 0.4]]) """ if data_variable not in self.variable_names: raise KeyError( "the variable '{}' is not in the " "DataRecord".format(data_variable) ) # If record to be changed is 'grid_element' or 'element_id', # check that provided grid_element is valid and that new # grid_element+element_id combination exist on the grid and # have valid format: if data_variable in ("grid_element", "element_id"): if data_variable == "grid_element": assoc_grid_element = new_value assoc_element_id = self.get_data(time, item_id, "element_id")[0] if data_variable == "element_id": if not isinstance(new_value, int): raise ValueError( "You have passed a non-integer " "element_id to DataRecord, this is not " "permitted" ) if new_value < 0: raise ValueError( "You have passed an element id below " "zero. This is not permitted" ) assoc_element_id = new_value assoc_grid_element = self.get_data(time, item_id, "grid_element")[0] self._check_grid_element_and_id(assoc_grid_element, assoc_element_id) if assoc_element_id >= self._grid[assoc_grid_element].size: raise ValueError( "The location " + assoc_grid_element + " " + str(assoc_element_id) + " does not exist on this grid" ) if time is None: self._dataset[data_variable].values[item_id] = new_value else: try: len(time) except TypeError as exc: raise TypeError("time must be a list or a 1-d array") from exc try: # check that time coordinate already exists time_index = np.where(self._dataset.time.values == time)[0][0] except IndexError as exc: raise IndexError( "The time you passed is not currently" " in the DataRecord, you must change the value" " you pass or first create the new time " " coordinate using the add_record method" ) from exc if item_id is None: self._dataset[data_variable].values[time_index] = new_value else: try: len(item_id) except TypeError as exc: raise TypeError("item_id must be a list or a 1-d array") from exc try: self._dataset["item_id"] self._dataset[data_variable].values[item_id, time_index] = new_value except KeyError as exc: raise KeyError("This DataRecord does not hold items") from exc
[docs] def calc_aggregate_value( self, func, data_variable, at="node", filter_array=None, fill_value=np.nan, args=(), **kwargs, ): """Apply a function to a variable aggregated at grid elements. Parameters ---------- func : function Function to apply to be aggregated. data_variable : str Name of variable on which to apply the function. at : str, optional Name of grid element at which to apply the function. Default is "node". filter_array: boolean array with dimensions matching that of the DataRecord (optional) Array to filter the DataRecord before aggregation. fill_value: float Fill value for array. Default is np.nan. args : tuple (optional) Additional positional arguments to pass to the function. **kwargs : key value pairs (optional) Additional keyword arguments to pass to func. Returns ------- out : ndarray Array of size number-of-grid_elements (grid_elements is the group passed as 'at' argument). Examples -------- >>> import numpy as np >>> from landlab.data_record import DataRecord >>> from landlab import RasterModelGrid >>> grid = RasterModelGrid((3, 3)) >>> element_id = [0, 0, 0, 0, 1, 2, 3, 4, 5, 9999] >>> volumes = [4, 5, 1, 2, 3, 4, 5, 6, 7, 1234] >>> ages = [10, 11, 12, 13, 14, 15, 16, 8, 10, 3456] >>> grid_element = "node" >>> data = {"ages": ages, "volumes": volumes} >>> dr = DataRecord( ... grid, ... dummy_elements={"node": [9999]}, ... items={"grid_element": "node", "element_id": np.array(element_id)}, ... data_vars={ ... "ages": (["item_id"], np.array(ages)), ... "volumes": (["item_id"], np.array(volumes)), ... }, ... ) >>> s = dr.calc_aggregate_value(func=xr.Dataset.sum, data_variable="ages") >>> s array([ 46., 14., 15., 16., 8., 10., nan, nan, nan]) >>> len(s) == grid.number_of_nodes True If you want to first filter the DataRecord and then aggregate, first create a filter array with dimensions matching that of the DataRecord and has `True` for entries that should be retained and False for entries that should be ignored. For example, if we wanted to aggregate volume for items with an age greater than 10 we would to the following: >>> f = dr.dataset["ages"] > 10.0 >>> v_f = dr.calc_aggregate_value( ... func=xr.Dataset.sum, data_variable="volumes", filter_array=f ... ) >>> v_f array([ 8., 3., 4., 5., nan, nan, nan, nan, nan]) If we wanted the value for elements with no volume to be zero instead of np.nan we could use the keyword argument ``fill_value``. >>> f = dr.dataset["ages"] > 10.0 >>> v_f = dr.calc_aggregate_value( ... func=xr.Dataset.sum, ... data_variable="volumes", ... filter_array=f, ... fill_value=0.0, ... ) >>> v_f array([ 8., 3., 4., 5., 0., 0., 0., 0., 0.]) An array of ``fill_value`` is returned when ``filter_array`` is all ``False`` (np.nan is the default value). >>> f = dr.dataset["ages"] > 4000.0 >>> v_f = dr.calc_aggregate_value( ... func=xr.Dataset.sum, data_variable="volumes", filter_array=f ... ) >>> v_f array([ nan, nan, nan, nan, nan, nan, nan, nan, nan]) Other values can be specified for ``fill_value``. >>> f = dr.dataset["ages"] > 4000.0 >>> v_f = dr.calc_aggregate_value( ... func=xr.Dataset.sum, ... data_variable="volumes", ... filter_array=f, ... fill_value=0.0, ... ) >>> v_f array([ 0., 0., 0., 0., 0., 0., 0., 0., 0.]) """ filter_at = self._dataset["grid_element"] == at filter_valid_element = (self._dataset["element_id"] >= 0) * ( self._dataset["element_id"] < self._grid[at].size ) if filter_array is None: my_filter = filter_at * filter_valid_element else: my_filter = filter_at * filter_valid_element * filter_array if np.any(my_filter): # Filter DataRecord with my_filter and groupby element_id: filtered = self._dataset.where(my_filter).groupby("element_id") # Calculate values vals = filtered.map(func, *args, **kwargs) # .reduce # Create a nan array that we will fill with the results of the sum # this should be the size of the number of elements, even if there are # no items living at some grid elements. out = fill_value * np.ones(self._grid[at].size) # put the values of the specified variable into the correct location # of the out array. out[vals.element_id.values.astype(int)] = vals[data_variable] return out else: return np.repeat(fill_value, self._grid[at].size)
[docs] def ffill_grid_element_and_id(self): """Fill NaN values of the fields 'grid_element' and 'element_id'. Fields are filled by propagating values forward in time. Examples -------- >>> import numpy as np >>> from landlab import RasterModelGrid >>> from landlab.data_record import DataRecord >>> grid = RasterModelGrid((3, 3)) Example of a DataRecord with dimensions time and item_id: >>> my_items3 = { ... "grid_element": np.array([["node"], ["link"]]), ... "element_id": np.array([[1], [3]]), ... } Note that both arrays have 2 dimensions as they vary along dimensions 'time' and 'item_id'. >>> dr3 = DataRecord(grid, time=[0.0], items=my_items3) Records relating to pre-existing items can be added to the DataRecord using the method 'add_record': >>> dr3.add_record( ... time=[2.0, 3.0], ... new_record={"mean_elevation": (["time"], np.array([200.0, 250.0]))}, ... ) Adding this data record created two new time coordinates. The grid_element and element_id of the items has been filled with 'nan' for these time coordinates. >>> dr3.dataset["grid_element"].values array([['node', nan, nan], ['link', nan, nan]], dtype=object) >>> dr3.dataset["element_id"].values array([[ 1., nan, nan], [ 3., nan, nan]]) To fill these values with the last valid value, use the method ffill_grid_element_and_id: >>> dr3.ffill_grid_element_and_id() >>> dr3.dataset["grid_element"].values array([['node', 'node', 'node'], ['link', 'link', 'link']], dtype=object) >>> dr3.dataset["element_id"].values array([[ 1., 1., 1.], [ 3., 3., 3.]]) In some applications, there may be no prior valid value. Under these circumstances, those values will stay as NaN. That is, this only forward fills, and does not backfill. >>> my_items3 = { ... "grid_element": np.array([["node"], ["link"]]), ... "element_id": np.array([[1], [3]]), ... } >>> dr3 = DataRecord(grid, time=[0.0], items=my_items3) >>> dr3.dataset["element_id"].values array([[1], [3]]) >>> dr3.dataset["grid_element"].values array([['node'], ['link']], dtype='<U4') Next add some new items at a new time. >>> dr3.add_item( ... time=[1.0], ... new_item={ ... "grid_element": np.array([["node"], ["node"]]), ... "element_id": np.array([[4], [4]]), ... }, ... new_item_spec={"size": (["item_id", "time"], [[10], [5]])}, ... ) Two items have been added at a new timestep 1.0: >>> dr3.number_of_items 4 >>> dr3.time_coordinates [0.0, 1.0] >>> dr3.dataset["element_id"].values array([[ 1., nan], [ 3., nan], [ nan, 4.], [ nan, 4.]]) >>> dr3.dataset["grid_element"].values array([['node', nan], ['link', nan], [nan, 'node'], [nan, 'node']], dtype=object) We expect that the NaN's to the left of the 4.s will stay NaN. And they do. >>> dr3.ffill_grid_element_and_id() >>> dr3.dataset["element_id"].values array([[ 1., 1.], [ 3., 3.], [ nan, 4.], [ nan, 4.]]) >>> dr3.dataset["grid_element"].values array([['node', 'node'], ['link', 'link'], [nan, 'node'], [nan, 'node']], dtype=object) Finally, if we add a new time, we see that we need to fill in the full time column. >>> dr3.add_record(time=[2]) >>> dr3.dataset["element_id"].values array([[ 1., 1., nan], [ 3., 3., nan], [ nan, 4., nan], [ nan, 4., nan]]) >>> dr3.dataset["grid_element"].values array([['node', 'node', nan], ['link', 'link', nan], [nan, 'node', nan], [nan, 'node', nan]], dtype=object) And that forward filling fills everything as expected. >>> dr3.ffill_grid_element_and_id() >>> dr3.dataset["element_id"].values array([[ 1., 1., 1.], [ 3., 3., 3.], [ nan, 4., 4.], [ nan, 4., 4.]]) >>> dr3.dataset["grid_element"].values array([['node', 'node', 'node'], ['link', 'link', 'link'], [nan, 'node', 'node'], [nan, 'node', 'node']], dtype=object) """ ei = self._dataset["element_id"].values for i in range(ei.shape[0]): for j in range(1, ei.shape[1]): if np.isnan(ei[i, j]): ei[i, j] = ei[i, j - 1] self._dataset["element_id"] = (["item_id", "time"], ei) ge = self._dataset["grid_element"].values for i in range(ge.shape[0]): for j in range(1, ge.shape[1]): if ge[i, j] not in self._permitted_locations: ge[i, j] = ge[i, j - 1] self._dataset["grid_element"] = (["item_id", "time"], ge)
@property def dataset(self): """The xarray Dataset that serves as the core datastructure.""" return self._dataset @property def variable_names(self): """Return the name(s) of the data variable(s) in the record as a list.""" _keys = [] for key in self._dataset.to_dataframe().keys(): _keys.append(key) return _keys @property def number_of_items(self): """Return the number of items in the DataRecord.""" return len(self._dataset.item_id) @property def item_coordinates(self): """Return a list of the item_id coordinates in the DataRecord.""" return self._dataset.item_id.values.tolist() @property def number_of_timesteps(self): """Return the number of time steps in the DataRecord.""" return len(self._dataset.time) @property def time_coordinates(self): """Return a list of the time coordinates in the DataRecord.""" return self._dataset.time.values.tolist() @property def earliest_time(self): """Return the earliest time coordinate in the DataRecord.""" return min(self._dataset.time.values) @property def latest_time(self): """Return the latest time coordinate in the DataRecord.""" return max(self._dataset.time.values) @property def prior_time(self): """Return the penultimate time coordinate in the DataRecord.""" if self.number_of_timesteps < 2: return np.nan else: return sorted(self.time_coordinates)[-2]