Source code for jwst.model_blender.blendrules

""" blendmeta - Merge metadata from multiple models to create
                   a new metadata instance and table

from collections import OrderedDict

import numpy as np
from datetime import time
from import fits
from astropy.time import Time
from stdatamodels import schema as dm_schema
from stdatamodels import properties

from jwst import __version__
from . import blender

# Version of rules file format supported by this version of the code
# All changes should be backwards compatible to older rules versions
# so any rules file with Version >= __rules_version__ should work
# with this code
__rules_version__ = 2.1

# Custom blending functions
[docs]def multi(vals): """ This will either return the common value from a list of identical values or 'MULTIPLE' """ uniq_vals = list(set(vals)) num_vals = len(uniq_vals) if num_vals == 0: return None if num_vals == 1: return uniq_vals[0] if num_vals > 1: return "MULTIPLE"
[docs]def multi1(vals): """ This will either return the common value from a list of identical values or the single character '?' """ uniq_vals = list(set(vals)) num_vals = len(uniq_vals) if num_vals == 0: return None if num_vals == 1: return uniq_vals[0] if num_vals > 1: return "?"
[docs]def float_one(vals): """ Return a constant floating point value of 1.0 """ return 1.0
[docs]def int_one(vals): """ Return an integer value of 1 """ return int(1)
[docs]def zero(vals): """ Return a value of 0 """ return 0
[docs]def first(items): """ Return first item from list of values""" if len(items): return items[0] return None
[docs]def last(items): """ Return last item from list of values""" if len(items): return items[-1] return None
[docs]def mindate(items): """Return the minimum date from a list of date strings in yyyy-mm-dd format.""" time_list = Time(items, format="iso", in_subfmt="date", out_subfmt="date") return str(time_list.min())
[docs]def maxdate(items): """Return the maximum date from a list of date strings in yyyy-mm-dd format.""" time_list = Time(items, format="iso", in_subfmt="date", out_subfmt="date") return str(time_list.max())
[docs]def mindatetime(items): """Return the minimum datetime from a list of datetime strings in ISO-8601 format.""" time_list = Time(items, format="isot") return str(time_list.min())
[docs]def maxdatetime(items): """Return the maximum datetime from a list of datetime strings in ISO-8601 format.""" time_list = Time(items, format="isot") return str(time_list.max())
[docs]def mintime(items): times = [_isotime(time_str) for time_str in items] return min(times).isoformat()
[docs]def maxtime(items): times = [_isotime(time_str) for time_str in items] return max(times).isoformat()
def _isotime(time_str): hms = [float(i) for i in time_str.split(':')] sec_ms = hms[2] - int(hms[2]) isotime = time(int(hms[0]), int(hms[1]), int(hms[2]), int(sec_ms * 1000000)) return isotime # translation dictionary for function entries from rules files blender_funcs = {'first': first, 'last': last, 'float_one': float_one, 'int_one': int_one, 'zero': zero, 'multi': multi, 'multi?': multi1, 'mean': np.mean, 'sum': np.sum, 'max': np.max, 'min': np.min, 'stddev': np.std, 'mintime': mintime, 'maxtime': maxtime, 'mindate': mindate, 'maxdate': maxdate, 'mindatetime': mindatetime, 'maxdatetime': maxdatetime} # Classes for managing keyword rules
[docs]class KeywordRules(): def __init__(self, model): """ Read in the rules used to interpret the keywords from the specified instrument image header. """ self.instrument = self.new_header = None self.rules_version = __rules_version__ self.rules_file = model._schema self.rule_specs = _build_schema_rules_dict(self.rules_file) self.rule_objects = [] self.rules = [] self.section_names = []
[docs] def interpret_rules(self, hdrs): """ Convert specifications for rules from rules file into specific rules for this header(instrument/detector). Notes ----- This allows for expansion rules to be applied to rules from the rules files (such as any wildcards or section titles). Output will be 'self.rules' that contains a list of tuples: - a tuple of 2 values for each column in the table - a tuple of 4 values for each attribute identified in metadata Partial sample from HST to show format: [('CTYPE1O', 'CTYPE1O'), ('CTYPE2O', 'CTYPE2O'), ('CUNIT1O', 'CUNIT1O'), ('CUNIT2O', 'CUNIT2O'), ('APERTURE', 'APERTURE', <function fitsblender.blendheaders.multi>, 'ignore'), ('DETECTOR', 'DETECTOR', <function fitsblender.blender.first>, 'ignore'), ('EXPEND', 'EXPEND', <function numpy.core.fromnumeric.amax>, 'ignore'), ('EXPSTART', 'EXPSTART', <function numpy.core.fromnumeric.amin>, 'ignore'), ('EXPTIME', 'TEXPTIME', <function numpy.core.fromnumeric.sum>, 'ignore'), ('EXPTIME', 'EXPTIME', <function numpy.core.fromnumeric.sum>, 'ignore')] This rules format will allow the algorithm, logic and code from the original fitsblender to be used with as little change as possible. It will need to be derived (as with HST) from the input models metadata for expansion of attribute sections or wildcards in attributes specified in the rules. """ if isinstance(hdrs, tuple): hdrs = list(hdrs) if not isinstance(hdrs, list): hdrs = [hdrs] # apply rules to headers for attr in self.rule_specs: speclist = self.rule_specs[attr] for rule in speclist: # Create KwRule input equivalent to HST rules input kwr = KwRule(rule) duplicate_rule = False for robj in self.rule_objects: if kwr.rule_spec == robj.rule_spec: duplicate_rule = True break if not duplicate_rule: for hdr in hdrs: kwr.interpret(hdr) self.rules.extend(kwr.rules)
[docs] def merge(self, kwrules): """ Merge a new set of interpreted rules into the current set The new rules, kwrules, can either be a new class or a whole new set of rules (like those obtained from using self.interpret_rules with a new header). """ if isinstance(kwrules, KeywordRules): kwrules = kwrules.rules # Determine what rules are specified in kwrules that # are NOT in self.rules k = [] # Delete these extraneous rules from input kwrules for r in kwrules: if r not in self.rules: k.append(r) # extend self.rules with additional rules self.rules.extend(k)
[docs] def apply(self, models, tabhdu=False): """ For a full list of metadata objects, apply the specified rules to generate a dictionary of new values and a table using blender. This method returns the new metadata object and summary table as `datamodels.model.ndmodel` and fits.binTableHDU objects. """ # Apply rules to headers fbdict, fbtab = blender.metablender(models, self.rules) # Determine which keywords are included in the table but not # the new dict(header). These will be removed from the output # header altogether tabcols = fbtab.dtype.names # Start with a copy of the template as the new header # This will define what keywords need to be updated, as the rules # and input headers often include headers for multiple extensions in # order to build the complete table for all the keywords in the file # in one run new_model = models[0].copy() # Apply updated/blended values into new header, but only those # keywords which are already present in the 'template' new header # this allows the rules to be used on all extensions at once yet # update each extension separately without making copies of kws from # one extension to another. # new_model.update(fbdict) for attr in new_model.to_flat_dict(): if 'meta' in attr and attr in fbdict: new_model[attr] = fbdict[attr] # Create summary table if len(tabcols) > 0: if tabhdu: new_table = fits.BinTableHDU.from_columns(fbtab) new_table.header['EXTNAME'] = 'HDRTAB' else: new_table = fbtab else: new_table = None return new_model, new_table
[docs] def add_rules_kws(self, hdr): """ Update metadata with WARNING ------- Needs to be modified to work with metadata. Update PRIMARY header with HISTORY cards that report the exact rules used to create this header. Only non-comment lines from the rules file will be reported. """ hdr['meta.rules_version'] = (__rules_version__, 'Version ID for header kw rules file') hdr['meta.blender_version'] = (__version__, 'Version of blendheader software used')
[docs] def index_of(self, kw): """ Reports the index of the specified kw.""" indx = [] for r, i in zip(self.rules, list(range(len(self.rules)))): if r[0] == kw: indx.append(i) return indx
[docs]class KwRule(): """ This class encapsulates the logic needed for interpreting a single keyword rule from a text file. Notes ----- The ``.rules`` attribute contains the interpreted set of rules that corresponds to this line. Example:: Interpreting rule from {'meta.attribute': { 'rule': 'first', 'output': 'meta.attribute'}} --or-- {'meta.attribute': 'meta.attribute'} # Table column specification into rule [('meta.attribute', 'meta.attribute', <function first at 0x7fe505db7668>, 'ignore')] and sname None """ def __init__(self, line): """Initialize new keyword rule. Parameters ========== line : dict Line should be dict with attribute name as the key, and a dict as the value specifying 'rule' and (optionally)'output'. """ self.rule_spec = line # dict read in from rules file self.rules = [] self.delete_kws = [] self.section_name = []
[docs] def interpret(self, hdr): """Use metadata to interpret rule.""" if self.rules: # If self.rules has already been defined for this rule, do not try # to interpret it any further with additional headers return irules, sname = interpret_entry(self.rule_spec, hdr) # keep track of any section name identified for this rule if sname: self.section_name.append(sname) # Now, interpret rule based on presence of kw in hdr if irules: self.rules = irules
# Utility functions. def _build_schema_rules_dict(schema): """ Create a dict that extracts blend rules from an input schema. Parameters ---------- schema : JSON schema fragment The schema in which to search. Returns ------- results : OrderedDict Dictionary with schema attributes as keys and blend rules as values """ def build_rules_dict(subschema, path, combiner, ctx, recurse): # Only interpret elements of the meta component of the model if len(path) > 1 and path[0] == 'meta' and 'items' not in path: for combiner in ['anyOf', 'oneOf']: if combiner in path: path = path[:path.index(combiner)] break attr = '.'.join(path) if subschema.get('properties'): return # Ignore ObjectNodes # Get blending info kwrule = subschema.get('blend_rule') kwtab = subschema.get('blend_table') kwname = subschema.get('fits_keyword', attr) # If rules had already been set, only modify # the rules if there are explicit settings. rule_spec = None result = results.get(attr, []) if kwrule: rule_spec = {attr: {'rule': kwrule}} elif not results.get(attr): rule_spec = {attr: {'rule': 'first'}} if rule_spec: result.append(rule_spec) # Add keyword to table if specified. if kwtab: result.append({attr: kwname}) # Add the results back. if len(result): results[attr] = result else: return results = OrderedDict() dm_schema.walk_schema(schema, build_rules_dict, results) return results
[docs]def interpret_entry(line, hdr): """ Generate the rule(s) specified by the entry from the rules file. Notes ----- The entry should always be a dict with format: {attribute_name : {'rule':'some_rule', 'output':''}} -- or (for table column specification)-- {attribute_name: attribute_name} where 'output' is assumed to be the same as attribute_name if not present """ # Interpret raw input line attr = list(line.keys())[0] line_spec = line[attr] attr_column = True # Determine whether this rule defines a table column if isinstance(line_spec, dict): attr_column = False # If not, turn this off # Initialize output values rules = [] section_name = None # Parse the line if isinstance(hdr[attr], properties.ObjectNode): section_name = attr # Datamodel sections are just parent Nodes for each attribute keys = hdr[section_name].instance.keys() kws = ['{}.{}'.format(section_name, k) for k in keys] # if kws is not None: for kw1 in kws: if attr_column: # Interpret each attribute as a new column in Table rules.append((kw1, kw1)) else: # Interpret each attribute in Node with same rule attr_rules = interpret_attr_line(kw1, line_spec) rules.extend(attr_rules) else: attr_rules = interpret_attr_line(attr, line_spec) rules.extend(attr_rules) return rules, section_name
[docs]def interpret_attr_line(attr, line_spec): """ Generate rule for single attribute from input line from rules file.""" rules = [] kws = [attr] if isinstance(line_spec, dict): if 'output' in line_spec: kws2 = [line_spec['output']] else: kws2 = kws else: kws2 = [line_spec] lrule = None if 'rule' in line_spec: lrule = line_spec['rule'] # Interpret short-hand rules using dict if lrule is not None and len(lrule) > 0: if lrule in blender_funcs: lrule = blender_funcs[lrule] else: lrule = None # build separate rule for each kw for kw1, kw2 in zip(kws, kws2): new_rule = (kw1, kw2, lrule, "ignore") if new_rule not in rules: rules.append(new_rule) else: for kw1, kw2 in zip(kws, kws2): new_rule = (kw1, kw2) if new_rule not in rules: rules.append(new_rule) return rules
[docs]def find_keywords_in_section(hdr, title): """Return a list of keyword names. The list will be derived from the section with the specified section title identified in the hdr. """ # Indentify card indices of start and end of specified section sect_start = None sect_end = None for i, kw in enumerate( if sect_start is None: if title in str(hdr[i]): sect_start = i else: if '/' in str(hdr[i]) and hdr[i] not in ['N/A', ' ', '']: sect_end = i break if sect_end is None: sect_end = len(hdr) if sect_start is None: return None # Now, extract the keyword names from this section # section_keys = hdr.ascard[sect_start+1:sect_end-1].keys() section_keys = list(hdr[sect_start + 1:sect_end - 1].keys()) # remove any blank keywords while section_keys.count('') > 0: section_keys.remove('') return section_keys