"""Digital Micrograph (.dm3/.dm4) extractor plugin."""
import contextlib
import logging
from datetime import UTC
from datetime import datetime as dt
from pathlib import Path
from struct import error
from typing import Any, ClassVar, Dict, List
import numpy as np
from hyperspy.io import load as hs_load
from rsciio.utils.exceptions import (
DM3DataTypeError,
DM3FileVersionError,
DM3TagError,
DM3TagIDError,
DM3TagTypeError,
)
from nexusLIMS.extractors.base import ExtractionContext
from nexusLIMS.extractors.plugins.basic_metadata import BasicFileInfoExtractor
from nexusLIMS.extractors.plugins.profiles import register_all_profiles
from nexusLIMS.extractors.profiles import get_profile_registry
from nexusLIMS.extractors.utils import (
_coerce_to_list,
_find_val,
_parse_filter_settings,
_set_acquisition_device_name,
_set_camera_binning,
_set_eds_meta,
_set_eels_meta,
_set_eels_processing,
_set_eels_spectrometer_meta,
_set_exposure_time,
_set_gms_version,
_set_image_processing,
_set_si_meta,
_try_decimal,
add_to_extensions,
)
from nexusLIMS.instruments import get_instr_from_filepath
from nexusLIMS.schemas.units import ureg
from nexusLIMS.utils.dicts import (
remove_dict_nones,
remove_dtb_element,
set_nested_dict_value,
sort_dict,
try_getting_dict_value,
)
from nexusLIMS.utils.time import current_system_tz
_logger = logging.getLogger(__name__)
def _apply_profile_to_metadata(metadata: dict, instrument, file_path: Path) -> dict:
"""
Apply instrument profile to metadata dictionary.
This is a helper function used by get_dm3_metadata() to maintain backward
compatibility with code that calls it directly.
Parameters
----------
metadata
Metadata dictionary with 'nx_meta' key
instrument
Instrument object
file_path
Path to the file being processed
Returns
-------
dict
Modified metadata dictionary with profile transformations applied
"""
# Ensure profiles are loaded
register_all_profiles()
profile = get_profile_registry().get_profile(instrument)
if profile is None:
return metadata
_logger.debug("Applying profile for instrument: %s", instrument.name)
# Create a mock context for profile application
context = ExtractionContext(file_path=file_path, instrument=instrument)
# Apply custom parsers in order
for parser_name, parser_func in profile.parsers.items():
try:
metadata = parser_func(metadata, context)
except Exception as e:
_logger.warning(
"Profile parser '%s' failed: %s",
parser_name,
e,
)
# Apply transformations
for key, transform_func in profile.transformations.items():
try:
if key in metadata:
metadata[key] = transform_func(metadata[key])
except Exception as e:
_logger.warning(
"Profile transformation '%s' failed: %s",
key,
e,
)
# Inject extension fields
if profile.extension_fields:
for key, value in profile.extension_fields.items():
try:
add_to_extensions(metadata["nx_meta"], key, value)
except Exception as e:
_logger.warning(
"Profile extension field injection '%s' failed: %s",
key,
e,
)
return metadata
[docs]
def get_pre_path(mdict: Dict) -> List[str]:
"""
Get the appropriate pre-path in the metadata tag structure for a given signal.
Get the path into a dictionary where the important DigitalMicrograph metadata is
expected to be found. If the .dm3/.dm4 file contains a stack of images, the
important metadata for NexusLIMS is not at its usual place and is instead under a
`plan info` tag, so this method will determine if the stack metadata is present and
return the correct path.
Parameters
----------
mdict : dict
A metadata dictionary as returned by :py:meth:`get_dm3_metadata`
Returns
-------
A list containing the subsequent keys that need to be traversed to
get to the point in the `mdict` where the important metadata is stored
"""
# test if we have a stack
stack_val = try_getting_dict_value(
mdict,
["ImageList", "TagGroup0", "ImageTags", "plane info"],
)
if stack_val is not None:
# we're in a stack
pre_path = [
"ImageList",
"TagGroup0",
"ImageTags",
"plane info",
"TagGroup0",
"source tags",
]
else:
pre_path = ["ImageList", "TagGroup0", "ImageTags"]
return pre_path
def _migrate_to_schema_compliant_metadata(mdict: dict) -> dict: # noqa: PLR0912
"""
Migrate metadata to schema-compliant format.
This function reorganizes metadata extracted from DM3/DM4 files to conform
to the type-specific metadata schemas. It:
1. Maps display names to EM Glossary field names for core fields
2. Moves vendor-specific fields to the extensions section
3. Converts Stage Position dict to proper StagePosition structure
Parameters
----------
mdict : dict
Metadata dictionary with 'nx_meta' key
Returns
-------
dict
Metadata dictionary with schema-compliant nx_meta
"""
nx_meta = mdict.get("nx_meta", {})
dataset_type = nx_meta.get("DatasetType", "Image")
# Field mappings from display names to EM Glossary names
# These are core schema fields that just need renaming
# Note: dataset_type-specific fields are handled conditionally below
field_mappings = {
# Common mappings for all types
"Voltage": "acceleration_voltage",
"Horizontal Field Width": "horizontal_field_width",
"Vertical Field Width": "vertical_field_width",
"Acquisition Device": "acquisition_device",
"Sample Time": "dwell_time",
}
# Conditional mappings based on dataset type
if dataset_type == "Diffraction":
field_mappings["STEM Camera Length"] = "camera_length"
if dataset_type in ("Image", "SpectrumImage"):
# magnification is only a core field for image-like datasets;
# for others (e.g. Diffraction) it routes to extensions via the
# fall-through below
field_mappings["Indicated Magnification"] = "magnification"
# Fields that should ALWAYS go to extensions (vendor/instrument-specific)
extension_fields = {
# Gatan-specific
"GMS Version",
"Microscope",
"Operator",
"Specimen",
# Operation modes
"Illumination Mode",
"Imaging Mode",
"Operation Mode",
# Apertures
"Condenser Aperture",
"Objective Aperture",
"Selected Area Aperture",
# Vendor-specific settings
"Cs", # Spherical aberration
# Signal/Analytic metadata
"Signal Name",
"Analytic Format",
"Analytic Label",
"Analytic Signal",
# Nested vendor metadata (will be moved as-is)
"EELS",
"EDS",
# STEM-specific fields that should be extensions for non-Diffraction types
"STEM Camera Length", # Only core for Diffraction
}
# NOTE: "NexusLIMS Extraction" is added AFTER this migration function runs
# by add_extraction_details in __init__.py, so we don't need to handle it here
# Create new nx_meta dict with schema-compliant structure
new_nx_meta = {}
# Preserve any existing extensions (e.g., from instrument profiles)
extensions = nx_meta.get("extensions", {}).copy() if "extensions" in nx_meta else {}
# Copy required fields as-is
required_fields = {"Creation Time", "Data Type", "DatasetType"}
for field in required_fields:
if field in nx_meta:
new_nx_meta[field] = nx_meta[field]
# Copy common optional fields
common_fields = {
"Data Dimensions",
"Instrument ID",
"warnings",
"Extractor Warnings",
}
for field in common_fields:
if field in nx_meta:
new_nx_meta[field] = nx_meta[field]
# Process all other fields
for key, value in nx_meta.items():
# Skip if already processed
if key in required_fields or key in common_fields:
continue
# Check if it's a core field that needs renaming
if key in field_mappings:
new_key = field_mappings[key]
new_nx_meta[new_key] = value
# Check if it should go to extensions
elif key in extension_fields:
extensions[key] = value
# Handle Stage Position specially
elif key == "Stage Position":
# DM3 files have Stage Position as a dict with keys
# like 'X', 'Y', 'α', etc. # noqa: RUF003
# Convert to snake_case keys for StagePosition schema
if isinstance(value, dict):
stage_pos = {}
key_map = {
"X": "x",
"Y": "y",
"Z": "z",
"α": "tilt_alpha", # noqa: RUF001
"β": "tilt_beta",
}
for old_key, new_key in key_map.items():
if old_key in value:
# Convert to Pint Quantity if needed
val = value[old_key]
if new_key in ("x", "y") and not isinstance(val, ureg.Quantity):
# X/Y in micrometers
val = ureg.Quantity(val, "micrometer")
elif new_key == "z" and not isinstance(val, ureg.Quantity):
# Z in millimeters
val = ureg.Quantity(val, "millimeter")
elif new_key in (
"tilt_alpha",
"tilt_beta",
) and not isinstance(val, ureg.Quantity):
# Tilts in degrees
val = ureg.Quantity(val, "degree")
stage_pos[new_key] = val
# Only emit stage_position when non-empty and the dataset
# type declares the field (Image / SpectrumImage); route
# non-empty values to extensions for other types and drop
# empty dicts entirely.
if stage_pos:
if dataset_type in ("Image", "SpectrumImage"):
new_nx_meta["stage_position"] = stage_pos
else:
extensions["Stage Position"] = stage_pos
else:
# If it's not a dict, move to extensions (this is not expected)
extensions["Stage Position"] = value # pragma: no cover
# Everything else goes to extensions
else:
extensions[key] = value
# Add extensions if any
for key, value in extensions.items():
add_to_extensions(new_nx_meta, key, value)
mdict["nx_meta"] = new_nx_meta
return mdict
[docs]
def parse_dm3_microscope_info(mdict): # noqa: PLR0912
"""
Parse the "microscope info" metadata.
Parse the "important" metadata that is saved at specific places within the DM3 tag
structure into a consistent place in the metadata dictionary returned by
:py:meth:`get_dm3_metadata`. Specifically looks at the "Microscope Info",
"Session Info", and "Meta Data" nodes (these are not present on every microscope).
Parameters
----------
mdict : dict
A metadata dictionary as returned by :py:meth:`get_dm3_metadata`
Returns
-------
mdict : dict
The same metadata dictionary with some values added under the
root-level ``nx_meta`` key
"""
if "nx_meta" not in mdict:
mdict["nx_meta"] = {} # pragma: no cover
pre_path = get_pre_path(mdict)
# General "microscope info" .dm3 tags (not present on all instruments):
for meta_key in [
"Indicated Magnification",
"Actual Magnification",
"Cs(mm)",
"STEM Camera Length",
"Voltage",
"Operation Mode",
"Specimen",
"Microscope",
"Operator",
"Imaging Mode",
"Illumination Mode",
"Name",
"Field of View (\u00b5m)",
"Facility",
"Condenser Aperture",
"Objective Aperture",
"Selected Area Aperture",
["Stage Position", "Stage Alpha"],
["Stage Position", "Stage Beta"],
["Stage Position", "Stage X"],
["Stage Position", "Stage Y"],
["Stage Position", "Stage Z"],
]:
base = [*pre_path, "Microscope Info"]
meta_key = _coerce_to_list(meta_key) # noqa: PLW2901
val = try_getting_dict_value(mdict, base + meta_key)
# only add the value to this list if we found it, and it's not one of
# the "facility-wide" set values that do not have any meaning:
if val is not None and val not in ["DO NOT EDIT", "DO NOT ENTER"] and val != []:
# Store original field name for unit mapping
field_name = meta_key[-1] if isinstance(meta_key, list) else meta_key
# Convert to Pint Quantity if the field has units
unit_map = {
"Cs(mm)": "millimeter",
"STEM Camera Length": "millimeter",
"Voltage": "volt", # Will auto-convert to kilovolt
"Field of View (\u00b5m)": "micrometer",
}
if field_name in unit_map:
with contextlib.suppress(ValueError, TypeError):
val = ureg.Quantity(val, unit_map[field_name])
# Remove unit suffix from field name
if field_name == "Cs(mm)":
meta_key = ["Cs"] # noqa: PLW2901
elif field_name == "Field of View (\u00b5m)":
meta_key = ["Horizontal Field Width"] # noqa: PLW2901
# change output of "Stage Position" to unicode characters
if "Stage Position" in meta_key:
meta_key[-1] = (
meta_key[-1]
.replace("Alpha", "α") # noqa: RUF001
.replace("Beta", "β")
.replace("Stage ", "")
)
set_nested_dict_value(mdict, ["nx_meta", *meta_key], val)
# General "session info" .dm3 tags (sometimes this information is stored
# here instead of under "Microscope Info":
for meta_key in ["Detector", "Microscope", "Operator", "Specimen"]:
base = [*pre_path, "Session Info"]
meta_key = _coerce_to_list(meta_key) # noqa: PLW2901
val = try_getting_dict_value(mdict, base + meta_key)
# only add the value to this list if we found it, and it's not
# one of the "facility-wide" set values that do not have any meaning:
if val is not None and val not in ["DO NOT EDIT", "DO NOT ENTER"] and val != []:
set_nested_dict_value(mdict, ["nx_meta", *meta_key], val)
# General "Meta Data" .dm3 tags
for meta_key in [
"Acquisition Mode",
"Format",
"Signal",
# this one is seen sometimes in EDS signals:
["Experiment keywords", "TagGroup1", "Label"],
]:
base = [*pre_path, "Meta Data"]
meta_key = _coerce_to_list(meta_key) # noqa: PLW2901
val = try_getting_dict_value(mdict, base + meta_key)
# only add the value to this list if we found it, and it's not
# one of the "facility-wide" set values that do not have any meaning:
if val is not None and val not in ["DO NOT EDIT", "DO NOT ENTER"] and val != []:
if "Label" in meta_key:
set_nested_dict_value(mdict, ["nx_meta", "Analytic Label"], val)
else:
set_nested_dict_value(
mdict,
["nx_meta"] + [f"Analytic {lbl}" for lbl in meta_key],
val,
)
# acquisition device name:
_set_acquisition_device_name(mdict, pre_path)
# exposure time:
_set_exposure_time(mdict, pre_path)
# GMS version:
_set_gms_version(mdict, pre_path)
# camera binning:
_set_camera_binning(mdict, pre_path)
# image processing:
_set_image_processing(mdict, pre_path)
# Signal Name (from DataBar):
signal_name = try_getting_dict_value(mdict, [*pre_path, "DataBar", "Signal Name"])
if signal_name is not None:
set_nested_dict_value(mdict, ["nx_meta", "Signal Name"], signal_name)
# DigiScan Sample Time (dwell time per pixel in microseconds):
sample_time = try_getting_dict_value(mdict, [*pre_path, "DigiScan", "Sample Time"])
if sample_time is not None:
with contextlib.suppress(ValueError, TypeError):
sample_time = ureg.Quantity(sample_time, "microsecond")
set_nested_dict_value(
mdict,
["nx_meta", "Sample Time"],
sample_time,
)
if (
"Illumination Mode" in mdict["nx_meta"]
and "STEM" in mdict["nx_meta"]["Illumination Mode"]
):
mdict["nx_meta"]["Data Type"] = "STEM_Imaging"
return mdict
[docs]
def parse_dm3_eels_info(mdict):
"""
Parse EELS information from the metadata.
Parses metadata from the DigitalMicrograph tag structure that concerns any
EELS acquisition or spectrometer settings, placing it in an ``EELS``
dictionary underneath the root-level ``nx_meta`` node.
Parameters
----------
mdict : dict
A metadata dictionary as returned by :py:meth:`get_dm3_metadata`
Returns
-------
mdict : dict
The metadata dict with all the "EELS-specific" metadata added under ``nx_meta``
"""
pre_path = get_pre_path(mdict)
# EELS .dm3 tags of interest:
base = [*pre_path, "EELS"]
for meta_key in [
["Acquisition", "Exposure (s)"],
["Acquisition", "Integration time (s)"],
["Acquisition", "Number of frames"],
["Experimental Conditions", "Collection semi-angle (mrad)"],
["Experimental Conditions", "Convergence semi-angle (mrad)"],
]:
_set_eels_meta(mdict, base, meta_key)
# different instruments have the spectrometer information in different
# places...
if mdict["nx_meta"]["Instrument ID"] == "FEI-Titan-TEM":
base = [*pre_path, "EELS", "Acquisition", "Spectrometer"]
elif mdict["nx_meta"]["Instrument ID"] == "FEI-Titan-STEM":
base = [*pre_path, "EELS Spectrometer"]
else:
base = None
if base is not None:
for meta_key in [
"Aperture label",
"Dispersion (eV/ch)",
"Energy loss (eV)",
"Instrument name",
"Drift tube enabled",
"Drift tube voltage (V)",
"Slit inserted",
"Slit width (eV)",
"Prism offset (V)",
"Prism offset enabled ",
]:
meta_key = [meta_key] # noqa: PLW2901
_set_eels_spectrometer_meta(mdict, base, meta_key)
_set_eels_processing(mdict, pre_path)
# Set the dataset type to Spectrum if any EELS tags were added
if "EELS" in mdict["nx_meta"]:
_logger.info("Detected file as Spectrum type based on EELS metadata")
mdict["nx_meta"]["DatasetType"] = "Spectrum"
if "STEM" in mdict["nx_meta"]["Illumination Mode"]:
mdict["nx_meta"]["Data Type"] = "STEM_EELS"
else:
mdict["nx_meta"]["Data Type"] = "TEM_EELS"
return mdict
[docs]
def parse_dm3_eds_info(mdict):
"""
Parse EDS information from the dm3 metadata.
Parses metadata from the DigitalMicrograph tag structure that concerns any
EDS acquisition or spectrometer settings, placing it in an ``EDS``
dictionary underneath the root-level ``nx_meta`` node. Metadata values
that are commonly incorrect or may be placeholders are specified in a
list under the ``nx_meta.warnings`` node.
Parameters
----------
mdict : dict
A metadata dictionary as returned by :py:meth:`get_dm3_metadata`
Returns
-------
mdict : dict
The metadata dictionary with all the "EDS-specific" metadata
added as sub-node under the ``nx_meta`` root level dictionary
"""
pre_path = get_pre_path(mdict)
# EELS .dm3 tags of interest:
base = [*pre_path, "EDS"]
for meta_key in [
["Acquisition", "Continuous Mode"],
["Acquisition", "Count Rate Unit"],
["Acquisition", "Dispersion (eV)"],
["Acquisition", "Energy Cutoff (V)"],
["Acquisition", "Exposure (s)"],
["Count rate"],
["Detector Info", "Active layer"],
["Detector Info", "Azimuthal angle"],
["Detector Info", "Dead layer"],
["Detector Info", "Detector type"],
["Detector Info", "Elevation angle"],
["Detector Info", "Fano"],
["Detector Info", "Gold layer"],
["Detector Info", "Incidence angle"],
["Detector Info", "Solid angle"],
["Detector Info", "Stage tilt"],
["Detector Info", "Window thickness"],
["Detector Info", "Window type"],
["Detector Info", "Zero fwhm"],
["Live time"],
["Real time"],
]:
_set_eds_meta(mdict, base, meta_key)
# test to see if the SI attribute is present in the metadata dictionary.
# If so, then some relevant EDS values are located there, rather
# than in the root-level EDS tag (all the EDS.Acquisition tags from
# above)
if try_getting_dict_value(mdict, [*pre_path, "SI"]) is not None:
for meta_key in [
["Acquisition", "Continuous Mode"],
["Acquisition", "Count Rate Unit"],
["Acquisition", "Dispersion (eV)"],
["Acquisition", "Energy Cutoff (V)"],
["Acquisition", "Exposure (s)"],
]:
_set_si_meta(mdict, pre_path, meta_key)
# for an SI EDS dataset, set "Live time", "Real time" and "Count rate"
# to the averages stored in the ImageList.TagGroup0.ImageTags.EDS.Images
# values
im_dict = try_getting_dict_value(mdict, [*pre_path, "EDS", "Images"])
if isinstance(im_dict, dict):
for k, v in im_dict.items():
if k in mdict["nx_meta"]["EDS"]:
del mdict["nx_meta"]["EDS"][k]
# this should work for 2D (spectrum image) as well as 1D
# (linescan) datasets since DM saves this information as a 1D
# list regardless of original data shape
avg_val = np.array(v).mean()
set_nested_dict_value(
mdict,
["nx_meta", "EDS", f"{k} (SI Average)"],
avg_val,
)
# Add the .dm3 EDS values to the warnings list, since they might not be
# accurate
for meta_key in [
["Count rate"],
["Detector Info", "Active layer"],
["Detector Info", "Azimuthal angle"],
["Detector Info", "Dead layer"],
["Detector Info", "Detector type"],
["Detector Info", "Elevation angle"],
["Detector Info", "Fano"],
["Detector Info", "Gold layer"],
["Detector Info", "Incidence angle"],
["Detector Info", "Solid angle"],
["Detector Info", "Stage tilt"],
["Detector Info", "Window thickness"],
["Detector Info", "Window type"],
["Detector Info", "Zero fwhm"],
["Live time"],
["Real time"],
]:
if try_getting_dict_value(mdict, base + meta_key) is not None:
mdict["nx_meta"]["warnings"].append(
["EDS", meta_key[-1] if len(meta_key) > 1 else meta_key[0]],
)
# Set the dataset type to Spectrum if any EDS tags were added
if "EDS" in mdict["nx_meta"]:
_logger.info("Detected file as Spectrum type based on presence of EDS metadata")
mdict["nx_meta"]["DatasetType"] = "Spectrum"
if "STEM" in mdict["nx_meta"]["Illumination Mode"]:
mdict["nx_meta"]["Data Type"] = "STEM_EDS"
else:
# no known files match this mode, so skip for coverage
mdict["nx_meta"]["Data Type"] = "TEM_EDS" # pragma: no cover
return mdict
[docs]
def parse_dm3_spectrum_image_info(mdict):
"""
Parse "spectrum image" information from the metadata.
Parses metadata that concerns any spectrum imaging information (the "SI" tag) and
places it in a "Spectrum Imaging" dictionary underneath the root-level ``nx_meta``
node. Metadata values that are commonly incorrect or may be placeholders are
specified in a list under the ``nx_meta.warnings`` node.
Parameters
----------
mdict : dict
A metadata dictionary as returned by :py:meth:`get_dm3_metadata`
Returns
-------
mdict : dict
The metadata dictionary with all the "EDS-specific" metadata
added as sub-node under the ``nx_meta`` root level dictionary
"""
pre_path = get_pre_path(mdict)
# Spectrum imaging .dm3 tags of interest:
base = [*pre_path, "SI"]
for m_in, m_out in [
(["Acquisition", "Pixel time (s)"], ["Pixel time (s)"]),
(["Acquisition", "SI Application Mode", "Name"], ["Scan Mode"]),
(
["Acquisition", "Spatial Sampling", "Height (pixels)"],
["Spatial Sampling (Vertical)"],
),
(
["Acquisition", "Spatial Sampling", "Width (pixels)"],
["Spatial Sampling (Horizontal)"],
),
(
["Acquisition", "Scan Options", "Sub-pixel sampling"],
["Sub-pixel Sampling Factor"],
),
]:
val = try_getting_dict_value(mdict, base + m_in)
# only add the value to this list if we found it, and it's not
# one of the "facility-wide" set values that do not have any meaning:
if val is not None:
# Convert to Pint Quantity if the field has units
output_key = m_out[0] if len(m_out) == 1 else m_out
if output_key == "Pixel time (s)":
with contextlib.suppress(ValueError, TypeError):
val = ureg.Quantity(val, "second")
output_key = ["Pixel time"]
# add last value of each parameter to the "Spectrum Imaging" sub-tree
key_list = [output_key] if isinstance(output_key, str) else output_key
set_nested_dict_value(
mdict, ["nx_meta", "Spectrum Imaging", *key_list], val
)
# Check spatial drift correction separately:
drift_per_val = try_getting_dict_value(
mdict,
[*base, "Acquisition", "Artefact Correction", "Spatial Drift", "Periodicity"],
)
drift_unit_val = try_getting_dict_value(
mdict,
[*base, "Acquisition", "Artefact Correction", "Spatial Drift", "Units"],
)
if drift_per_val is not None and drift_unit_val is not None:
val_to_set = f"Spatial drift correction every {drift_per_val} {drift_unit_val}"
# make sure statement looks gramatically correct
if drift_per_val == 1:
val_to_set = val_to_set.replace("(s)", "")
else:
val_to_set = val_to_set.replace("(s)", "s")
# fix for "seconds(s)" (*********...)
if val_to_set[-2:] == "ss":
val_to_set = val_to_set[:-1]
set_nested_dict_value(
mdict,
["nx_meta", "Spectrum Imaging", "Artefact Correction"],
val_to_set,
)
start_val = try_getting_dict_value(mdict, [*base, "Acquisition", "Start time"])
end_val = try_getting_dict_value(mdict, [*base, "Acquisition", "End time"])
if start_val is not None and end_val is not None:
start_dt = dt.strptime(start_val, "%I:%M:%S %p").replace(tzinfo=UTC)
end_dt = dt.strptime(end_val, "%I:%M:%S %p").replace(tzinfo=UTC)
duration = (end_dt - start_dt).seconds # Calculate acquisition duration
with contextlib.suppress(ValueError, TypeError):
duration = ureg.Quantity(duration, "second")
set_nested_dict_value(
mdict,
["nx_meta", "Spectrum Imaging", "Acquisition Duration"],
duration,
)
# Set the dataset type to SpectrumImage if it is already a Spectrum ( otherwise it's
# just a STEM image) and any Spectrum Imaging tags were added
if (
"Spectrum Imaging" in mdict["nx_meta"]
and mdict["nx_meta"]["DatasetType"] == "Spectrum"
):
_logger.info(
"Detected file as SpectrumImage type based on "
"presence of spectral metadata and spectrum imaging "
"info",
)
mdict["nx_meta"]["DatasetType"] = "SpectrumImage"
mdict["nx_meta"]["Data Type"] = "Spectrum_Imaging"
if "EELS" in mdict["nx_meta"]:
mdict["nx_meta"]["Data Type"] = "EELS_Spectrum_Imaging"
if "EDS" in mdict["nx_meta"]:
mdict["nx_meta"]["Data Type"] = "EDS_Spectrum_Imaging"
return mdict
def _parse_stage_position(tecnai_info):
"""
Parse stage position from Tecnai metadata.
Parameters
----------
tecnai_info : list
Split metadata strings
Returns
-------
dict
Dictionary with stage position x, y, z, theta, phi values
"""
tmp = _find_val("Stage ", tecnai_info).split(",")
tmp = [_try_decimal(t.strip(" umdeg")) for t in tmp]
return {
"Stage_Position_x": tmp[0],
"Stage_Position_y": tmp[1],
"Stage_Position_z": tmp[2],
"Stage_Position_theta": tmp[3],
"Stage_Position_phi": tmp[4],
}
def _parse_apertures(tecnai_info):
"""
Parse aperture settings from Tecnai metadata.
Parameters
----------
tecnai_info : list
Split metadata strings
Returns
-------
dict
Dictionary with C1, C2, Obj, and SA aperture values
"""
def _read_aperture(val, tecnai_info_):
"""Test if aperture has value or is retracted."""
try:
value = _find_val(val, tecnai_info_).strip(" um")
return int(value)
except (ValueError, AttributeError):
return None
return {
"C1_Aperture": _read_aperture("C1 Aperture: ", tecnai_info),
"C2_Aperture": _read_aperture("C2 Aperture: ", tecnai_info),
"Obj_Aperture": _read_aperture("OBJ Aperture: ", tecnai_info),
"SA_Aperture": _read_aperture("SA Aperture: ", tecnai_info),
}
[docs]
def process_tecnai_microscope_info(
microscope_info,
delimiter="\u2028",
):
"""
Process the Microscope_Info metadata string into a dictionary of key-value pairs.
This method is only relevant for FEI Titan TEMs that write additional metadata into
a unicode-delimited string at a certain place in the DM3 tag structure
Parameters
----------
microscope_info : str
The string of data obtained from the Tecnai.Microscope_Info leaf of the metadata
delimiter : str
The value (a unicode string) used to split the ``microscope_info`` string.
Returns
-------
info_dict : dict
The information contained in the string, in a more easily-digestible form.
"""
info_dict = {}
tecnai_info = microscope_info.split(delimiter)
info_dict["Microscope_Name"] = _find_val("Microscope ", tecnai_info) # String
info_dict["User"] = _find_val("User ", tecnai_info) # String
tmp = _find_val("Gun ", tecnai_info)
info_dict["Gun_Name"] = tmp[: tmp.index(" Extr volt")]
tmp = tmp[tmp.index(info_dict["Gun_Name"]) + len(info_dict["Gun_Name"]) :] # String
tmp = tmp.replace("Extr volt ", "")
info_dict["Extractor_Voltage"] = int(tmp.split()[0]) # Integer (volts)
tmp = tmp[tmp.index("Gun Lens ") + len("Gun Lens ") :]
info_dict["Gun_Lens_No"] = int(tmp.split()[0]) # Integer
tmp = tmp[tmp.index("Emission ") + len("Emission ") :]
info_dict["Emission_Current"] = _try_decimal(tmp.split("uA")[0]) # Decimal (microA)
tmp = _find_val("Mode ", tecnai_info)
info_dict["Mode"] = tmp[: tmp.index(" Defocus")] # String
# 'Mode' should be five terms long, and the last term is either 'Image',
# 'Diffraction', (or maybe something else)
# Decimal val (micrometer)
if "Magn " in tmp: # Imaging mode
info_dict["Defocus"] = _try_decimal(tmp.split("Defocus (um) ")[1].split()[0])
elif "CL " in tmp: # Diffraction mode
info_dict["Defocus"] = _try_decimal(tmp.split("Defocus ")[1].split()[0])
# This value changes based on whether in image or diffraction mode (mag or CL)
# Integer
if info_dict["Mode"].split()[4] == "Image":
info_dict["Magnification"] = int(tmp.split("Magn ")[1].strip("x"))
# Decimal
elif info_dict["Mode"].split()[4] == "Diffraction":
info_dict["Camera_Length"] = _try_decimal(tmp.split("CL ")[1].strip("m"))
# Integer (1 to 5)
info_dict["Spot"] = int(_find_val("Spot ", tecnai_info))
# Decimals - Lens strengths expressed as a "%" value
info_dict["C2_Strength"] = _try_decimal(_find_val("C2 ", tecnai_info).strip("%"))
info_dict["C3_Strength"] = _try_decimal(_find_val("C3 ", tecnai_info).strip("%"))
info_dict["Obj_Strength"] = _try_decimal(_find_val("Obj ", tecnai_info).strip("%"))
info_dict["Dif_Strength"] = _try_decimal(_find_val("Dif ", tecnai_info).strip("%"))
# Decimal values (micrometers)
tmp = _find_val("Image shift ", tecnai_info).strip("um")
info_dict["Image_Shift_x"] = _try_decimal(tmp.split("/")[0])
info_dict["Image_Shift_y"] = _try_decimal(tmp.split("/")[1])
# Parse stage position and apertures using helper functions
info_dict.update(_parse_stage_position(tecnai_info))
info_dict.update(_parse_apertures(tecnai_info))
# Nested dictionary
info_dict = _parse_filter_settings(info_dict, tecnai_info)
return _parse_filter_settings(info_dict, tecnai_info)