Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create new 3D bioimaging example. #7309

Open
wants to merge 18 commits into
base: main
Choose a base branch
from
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
205 changes: 205 additions & 0 deletions doc/examples/applications/plot_3d_segmentation_embryo.py
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Taking a step back, I'm curious about your (teaching) goals with this example.

At first glance, the segmentation approach taken here seems very similar to the one taken in Segment human cells (in mitosis). Is the goal to compare the approach to a machine-learning-based one?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

At first glance, the segmentation approach taken here seems very similar to the one taken in Segment human cells (in mitosis).

Yes, but in 3D. I also had in mind the comparison between this 'native' 3D segmentation and a 2D version along z ('stitching' back the xy sections together afterwards): I tried quickly, and the direct 3D segmentation works much better. Also, I'd like to extend this example one day (or write a new one which would refer to this one) with the tracking challenge, since the (original/full) dataset is not only 3D in space but has a time dimension.

Is the goal to compare the approach to a machine-learning-based one?

Yes!

Original file line number Diff line number Diff line change
@@ -0,0 +1,205 @@
"""
==================================================
Segment 3D image sample of developing mouse embryo
==================================================

In this example, we look at a microscopy image of a developing mouse embryo.
We use sample data from [1]_, more precisely from embryo B at time point 184.
"""

import io
import requests

import matplotlib.pyplot as plt
import numpy as np
from scipy import ndimage as ndi

import skimage as ski


#####################################################################
# We downloaded the original data in KLB format and sliced a particular
# subvolume, which we saved into a compressed Numpy format:
#
# .. code-block:: python
#
# import numpy as np
# import pyklb as klb
#
# data = klb.readfull('Mmu_E1_CAGTAG1.TM000184_timeFused_blending/SPM00_TM000184_CM00_CM01_CHN00.fusedStack.corrected.shifted.klb')
# sample = data[400:450, 1000:1750, 400:900]
# np.savez_compressed('sample_3D_frame_184.npz', sample)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I was originally looking at the rendered example and was wondering / confused on why this code was passing even though pyklb isn't available when building the docs.

This might confuse readers in a similar manner; is this really relevant here? Or how we can make it more clear that this is not part of the examples workflow?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The rendering of such code blocks used to make it clear that they weren't part of the current workflow... They would appear as 'commented out,' sort of. I had used this formatting in a previous example ("Estimate anisotropy in a 3D microscopy image"). I can see that it's not the case anymore (with the new theme, I guess).

Yes, it would be best to find some alternative formatting...

Copy link
Member

@lagru lagru May 13, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe a code block within a citation block? 🤔


#####################################################################
# View 3D image data
# ==================

sample_url = 'https://gitlab.com/scikit-image/data/-/raw/30a6bf082e5a91a2ee97e003465537224ffad216/Embryo2/sample_3D_frame_184.npz'
response = requests.get(sample_url)
im3d_dict = np.load(io.BytesIO(response.content))
im3d = im3d_dict['arr_0']

print(f'The shape of the image is: {im3d.shape}')

#####################################################################
# The sample dataset is a 3D image with 50 `xy` sections stacked along `z`. Let us
# visualize it by picking every fifth section.

data_montage = ski.util.montage(im3d[::5], grid_shape=(2, 5), padding_width=5)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Always surprised when discovering another small convenience function. 😄

fig, ax = plt.subplots(figsize=(10, 5))
ax.imshow(data_montage, interpolation="none")
ax.set_axis_off()

#####################################################################
# Apply thresholding techniques
# =============================

# global thresholding vs local thresholding
global_thresh = ski.filters.threshold_otsu(im3d)
binary_global = im3d > global_thresh

block_size = 31
local_thresh = ski.filters.threshold_local(im3d, block_size)
binary_local = im3d > local_thresh

#####################################################################
# Let us view the mid-stack `xy` section.

fig, axes = plt.subplots(ncols=3, figsize=(12, 4))
ax = axes.ravel()

ax[0].imshow(im3d[25, :, :], interpolation="none")
ax[0].set_title('Original')

ax[1].imshow(binary_global[25, :, :], interpolation="none")
ax[1].set_title('Global thresholding (Otsu)')

ax[2].imshow(binary_local[25, :, :], interpolation="none")
ax[2].set_title('Local thresholding')

for a in ax:
a.axis('off')

#####################################################################
# We smooth out the locally thresholded image (which is binary), so we can
# in turn threshold it globally.

smooth = ski.filters.gaussian(binary_local, sigma=1.5)
thresholds = ski.filters.threshold_multiotsu(smooth, classes=3)
regions = np.digitize(smooth, bins=thresholds)

fig, ax = plt.subplots(ncols=2, figsize=(8, 4))
ax[0].imshow(smooth[25, :, :], interpolation="none")
ax[0].set_title('Smoothing out')
ax[0].axis('off')
ax[1].imshow(regions[25, :, :], interpolation="none")
ax[1].set_title('Multi-Otsu thresholding')
ax[1].axis('off')

#####################################################################
# We identify nuclei to be the brightest of the three classes and we remove
# small objects.

cells_noisy = smooth > thresholds[1]
cells = ski.morphology.opening(cells_noisy, footprint=np.ones((3, 5, 5)))

#####################################################################
# Use watershed algorithm
# =======================
# We use the watershed algorithm to separate nuclei when they are touching
# or overlapping.

distance = ndi.distance_transform_edt(cells)

local_max_coords = ski.feature.peak_local_max(
distance, min_distance=12, exclude_border=False
)
local_max_mask = np.zeros(distance.shape, dtype=bool)
local_max_mask[tuple(local_max_coords.T)] = True
markers = ski.measure.label(local_max_mask)

segmented_cells = ski.segmentation.watershed(-distance, markers, mask=cells)

fig, ax = plt.subplots(ncols=2, figsize=(10, 5))
ax[0].imshow(cells[25, :, :], cmap='gray')
ax[0].set_title('Touching nuclei')
ax[0].axis('off')
ax[1].imshow(ski.color.label2rgb(segmented_cells[25, :, :], bg_label=0))
ax[1].set_title('Segmented nuclei')
ax[1].axis('off')

#####################################################################
# With the naked eye, we can see a slight over-segmentation... How does this
# segmentation compare with that obtained in [1]_?

#####################################################################
# Compare segmentation results
# ============================
# We used the software developed for the original research, "TGMM paper," and made available
# [online](https://bitbucket.org/fernandoamat/tgmm-paper/src/master/doc/new/docs/user-guide/quickstart.md).
# We edited the TGMM config file to apply the hierarchical segmentation on
# the sample data, which we saved 'back' in KLB format:
#
# .. code-block:: python
#
# klb.writefull(np.ascontiguousarray(sample), 'sample_3D_frame_184.klb')
#
# We installed the software following the [instructions](https://bitbucket.org/fernandoamat/tgmm-paper/src/master/doc/new/docs/dev-guide/building.md)
# and ran it:
#
# .. code-block:: bash
#
# ProcessStack config.md 184
# ProcessStack sample_3D_frame_184_seg_conn74_rad2.bin 14 50
#
# We chose tau=14 because persistanceSegmentationTau=14 in the TGMM config file.
# A value of tau=2 clearly yields over-segmentation (yields 1597 nuclei).
# We chose minSuperVoxelSzPx=50 because minNucleiSize=50 in the TGMM config file.
# There is not much difference between minSuperVoxelSzPx=50 (yields 517 nuclei) and
# minSuperVoxelSzPx=14 (yields 541 nuclei).
# The output (segmentation result) is a KLB file; we save it into a Numpy archive,
# which we can easily load here:

res_url = 'https://gitlab.com/scikit-image/data/-/raw/30a6bf082e5a91a2ee97e003465537224ffad216/Embryo2/tgmm_conn74_tau14.npz'
resp = requests.get(res_url)
gt_dict = np.load(io.BytesIO(resp.content))
gt = gt_dict['arr_0']

assert gt.shape == im3d.shape

# Ensure TGMM result is an image of type "labeled"
assert gt.dtype in [np.uint16, np.uint32, np.uint64]
assert gt.min() == 0
assert gt.max() == np.unique(gt).shape[0] - 1
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@lagru maybe, once we type the API, the above could turn into a one-liner looking like assert isinstance(gt, Labels)? 💪

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure, isinstance() works with class inheritance, typing is not necessarily about that but might be about structural typing and such. And I find it difficult to imagine that running a type checker would check the actual content like it's done here.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What is structural typing? 👀

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Have a look at protocols and PEP 544 -- Protocols: Structural subtyping (static duck typing) if you are interested. Basically it's trying to address the problem that typing doesn't support a lot of the more dynamic duck typing capabilities of Python. It's also referred to as "static duck typing".


fig, ax = plt.subplots(ncols=2, figsize=(10, 5))
ax[0].imshow(ski.color.label2rgb(gt[25, :, :], bg_label=0))
ax[0].set_title('TGMM output')
ax[0].axis('off')
ax[1].imshow(ski.color.label2rgb(segmented_cells[25, :, :], bg_label=0))
ax[1].set_title('Our output')
ax[1].axis('off')

#####################################################################
# Although the TGMM segmentation looks cleaner than ours, it seems to be missing
# quite a few nuclei in the upper half of the `xy` section.

print(f'TGMM finds {gt.max()} nuclei.')
print(f'We find {segmented_cells.max()} nuclei.')

#####################################################################
# Our *local* thresholding seems to be making the difference here.
# When enhancing the contrast of the original image, the nuclei in this darker
# area can be clearly seen:

enhanced_image = ski.exposure.equalize_hist(im3d[25, :, :])
fig, ax = plt.subplots()
ax.imshow(enhanced_image, interpolation="none")

#####################################################################
# .. [1] McDole K, Guignard L, Amat F, Berger A, Malandain G, Royer LA,
# Turaga SC, Branson K, Keller PJ (2018) "In Toto Imaging and
# Reconstruction of Post-Implantation Mouse Development at the
# Single-Cell Level" Cell, 175(3):859-876.e33.
# ISSN: 0092-8674
# :DOI:`10.1016/j.cell.2018.09.031`