{
"cells": [
{
"cell_type": "markdown",
"metadata": {
"tags": []
},
"source": [
"# Visualizing and reducing WFM imaging data with Scipp\n",
"\n",
"\n",
"## How to start\n",
"\n",
"Before starting you must:\n",
"\n",
"- Have conda installed\n",
"- `conda env create -f ess-notebooks-stable.yml python=3.7` . The yaml environment file is part of this repository.\n",
"- fetch the data `git clone git@github.com:scipp/ess-notebooks-data.git` somewhere local \n",
"- Generate the `dataconfig.py` file using `make_config.py` located in same directory as this notebook. In general, you simply need to point `make_config.py` to the root directory of data you cloned above. Refer to the help `make_config.py --help` for more information. \n",
"\n",
"## Experimental Summary\n",
"\n",
"This script has been developed to measure local strain $\\varepsilon$ defined as $\\varepsilon = \\Delta L/L_{0}$ in a FCC steel sample under elastic strain in a stress rig.\n",
"The measurements were measured at V20, HZB, Berlin, in September 2018 by Peter Kadletz.\n",
"\n",
"$\\lambda = 2 d \\sin\\theta$, where $2\\theta = \\pi$ (transmission), edges characterise the Bragg condition and hence $\\lambda = 2 d$.\n",
"Therefore strain is easily computed from the wavelength measurement of a Bragg edge directly, using un-loaded vs loaded experimental runs (and reference mesurements).\n",
"\n",
"The known Miller indices of the crystal structure (FCC) are used to predict the wavelength of the Bragg edges, which is bound by the reachable wavelength extents of the instrument.\n",
"This provides an approximate region to apply a fit.\n",
"A complement error function is used to fit each Bragg edge, and a refined centre location ($\\lambda$) for the edge is used in the strain measurement.\n",
"Because each Bragg edge can be identified individually, one can determine anisotropic strain across the unit cell in the reachable crystallographic directions.\n",
"\n",
"\n",
"\n",
"# Setup"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"import numpy as np\n",
"import matplotlib.pyplot as plt\n",
"plt.ioff()\n",
"\n",
"import scipp as sc\n",
"import scippneutron as sn\n",
"from scipp import plot\n",
"\n",
"import os\n",
"import sys\n",
"\n",
"import ess.v20.imaging as imaging\n",
"import ess.wfm as wfm\n",
"\n",
"import dataconfig\n",
"\n",
"local_data_path = os.path.join('ess-notebooks', 'v20', 'imaging',\n",
" 'gp2-stress-experiments')\n",
"data_dir = os.path.join(dataconfig.data_root, local_data_path)\n",
"instrument_file = os.path.join(data_dir, 'V20_Definition_GP2.xml')\n",
"tofs_path = os.path.join(data_dir, 'GP2_Stress_time_values.txt')\n",
"raw_data_dir = os.path.join(data_dir)"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"def load_component_info(file, advanced_geometry=False):\n",
" instrument = sn.load(filename=file,\n",
" mantid_alg='LoadEmptyInstrument',\n",
" mantid_args={'StoreInADS': False})\n",
" geometry = {key: value for key, value in instrument.coords.items()}\n",
" # Assume the detector is square\n",
" npix = int(np.sqrt(instrument.sizes[\"spectrum\"]))\n",
" geometry[\"position\"] = sc.fold(geometry[\"position\"],\n",
" dim='spectrum', dims=['y', 'x'], shape=[npix, npix])\n",
" geometry[\"x\"] = geometry[\"position\"].fields.x['y', 0]\n",
" geometry[\"y\"] = geometry[\"position\"].fields.y['x', 0]\n",
" del geometry[\"spectrum\"]\n",
" del geometry[\"empty\"]\n",
" return geometry"
]
},
{
"cell_type": "markdown",
"metadata": {
"pycharm": {
"name": "#%% md\n"
}
},
"source": [
"# Reduction\n",
"\n",
"## Load the data files and instrument geometry"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"def load_and_scale(folder_name, scale_factor):\n",
" to_load = os.path.join(raw_data_dir, folder_name)\n",
" variable = imaging.tiffs_to_variable(to_load, dtype=np.float32, with_variances=False)\n",
" variable *= scale_factor\n",
" return variable"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"# Number of pulses for each run, to scale data according to integration times\n",
"pulse_number_reference = 1.0 / 770956\n",
"pulse_number_sample = 1.0 / 1280381\n",
"pulse_number_sample_elastic = 1.0 / 2416839\n",
"\n",
"# Create Dataset\n",
"ds = sc.Dataset()\n",
"\n",
"# Load tiff stack\n",
"ds[\"reference\"] = load_and_scale(folder_name=\"R825-open-beam\",\n",
" scale_factor=pulse_number_reference)\n",
"ds[\"sample\"] = load_and_scale(folder_name=\"R825\",\n",
" scale_factor=pulse_number_sample)\n",
"ds[\"sample_elastic\"] = load_and_scale(folder_name=\"R825-600-Mpa\",\n",
" scale_factor=pulse_number_sample_elastic)\n",
"\n",
"# Load time bins from 1D text file\n",
"ds.coords[\"t\"] = sc.array(\n",
" dims=[\"t\"],\n",
" unit=sc.units.us,\n",
" values=imaging.read_x_values(tofs_path,\n",
" skiprows=1,\n",
" usecols=1,\n",
" delimiter='\\t') * 1.0e3)\n",
"\n",
"# Instrument geometry\n",
"geometry = load_component_info(instrument_file)\n",
"for key, val in geometry.items():\n",
" ds.coords[key] = val\n",
"\n",
"# Chopper cascade\n",
"beamline = imaging.make_beamline()\n",
"ds.coords['choppers'] = sc.scalar(beamline[\"choppers\"])\n",
"for key, value in beamline[\"source\"].items():\n",
" ds.coords[key] = value"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"ds"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Raw data visualization"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(ds[\"sample\"]['t', 600:], norm='log') # Slice at 600 gets us to an interesting t value"
]
},
{
"cell_type": "markdown",
"metadata": {
"pycharm": {
"name": "#%% md\n"
}
},
"source": [
"## Converting time coordinate to TOF\n",
"\n",
"Use the instrument geometry and chopper cascade parameters to compute time-distance diagram."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"frames = wfm.get_frames(ds)\n",
"frames"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"wfm.plot.time_distance_diagram(ds)"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"The `wfm.plot` provides another helper function to inspect the individual frames on a spectrum of integrated counts:"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"wfm.plot.frames_before_stitching(data=ds['reference'], frames=frames, dim='t')"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"We then use the `wfm.stitch` function to convert the time dimension to time-of-flight\n",
"(see [here](https://scipp.github.io/ess/techniques/wfm/introduction-to-wfm.html)\n",
"for more details on working with WFM data).\n",
"The contributions from each frame are automatically rebinned onto a single time-of-flight axis."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"stitched = wfm.stitch(data=ds, dim=\"t\", frames=frames)\n",
"stitched"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(sc.sum(sc.sum(stitched, 'x'), 'y'))"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Crop to relevant Tof section\n",
"\n",
"We take a subset of the Tof range which contains the Bragg edges of interest:"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"tof_start = 9.0e3*sc.units.us\n",
"tof_end = 2.75e4*sc.units.us\n",
"stitched = stitched[\"tof\", tof_start:tof_end].copy()"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Transmission Masking\n",
"\n",
"Divides the integrated sample counts with an open beam reference. Any values > masking threshold will be masked. The adj pixels step checks for random pixels which were left unmasked or masked with all their neighbours having the same mask value. These are forced to True or false depending on their neighbour value."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"integrated = sc.sum(stitched, 'tof')\n",
"integrated /= integrated[\"reference\"]\n",
"del integrated[\"reference\"]"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"masking_threshold = 0.80 * sc.units.one\n",
"\n",
"for key in [\"sample\", \"sample_elastic\"]:\n",
" mask = integrated[key].data > masking_threshold\n",
" # Apply some neighbour smoothing to the masks\n",
" mask = imaging.operations.mask_from_adj_pixels(mask)\n",
" stitched[key].masks[\"non-sample-region\"] = mask\n",
"stitched"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(sc.sum(stitched[\"sample\"], \"tof\"))"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Convert to wavelength\n",
"\n",
"Scipp's `neutron` submodule contains utilities specific to neutron science, and in particular unit conversions."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"stitched = sn.convert(stitched, origin=\"tof\", target=\"wavelength\", scatter=False)\n",
"# Rebin to common wavelength axis\n",
"edges = sc.array(dims=[\"wavelength\"],\n",
" values=np.linspace(2.0, 5.5, 129), unit=sc.units.angstrom)\n",
"wavelength = sc.rebin(stitched, \"wavelength\", edges)\n",
"\n",
"wavelength"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Apply mean filter"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"for key in wavelength:\n",
" wavelength[key].data = imaging.operations.mean_from_adj_pixels(wavelength[key].data)"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(wavelength[\"sample\"])"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Other visualizations\n",
"\n",
"### Show difference between sample and elastic"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(sc.sum(wavelength[\"sample\"] - wavelength[\"sample_elastic\"], 'wavelength'),\n",
" vmin=-0.002*sc.units.counts, vmax=0.002*sc.units.counts, cmap=\"RdBu\")"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Group detector pixels to increase signal to noise"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"nbins=27\n",
"grouped = imaging.operations.groupby2D(wavelength, nx_target=nbins, ny_target=nbins)\n",
"for key, item in grouped.items():\n",
" item.masks[\"zero-counts\"] = item.data == 0.0*sc.units.counts\n",
"del wavelength"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(grouped[\"sample\"])"
]
},
{
"cell_type": "markdown",
"metadata": {
"pycharm": {
"name": "#%% md\n"
}
},
"source": [
"## Normalization"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {
"pycharm": {
"name": "#%%\n"
}
},
"outputs": [],
"source": [
"# Normalize by open beam\n",
"normalized = grouped / grouped[\"reference\"]\n",
"del normalized[\"reference\"]\n",
"summed = sc.nansum(sc.nansum(normalized, 'x'), 'y') "
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"plot(summed, grid=True, title='I/I0')"
]
},
{
"cell_type": "markdown",
"metadata": {},
"source": [
"