Cell segmentation

Description

Fiji plugin to segment oocyte and zona pellucida contours from transmitted light images and extract hundreds of morphological features to describe numerically the oocyte. Segmentation is based on trained neural networks (U-Net) that were trained on both mouse and human oocytes (in prophase and meiosis I) acquired in different conditions. They are freely avaialable on the github repository and can be retrained if necessary. Oocytor also have options to extract hundreds of morphological/intensity features to characterize manually the oocyte (eg perimeter, texture...). These features can also be used in machine learning pipeline for automatic phenotyping.

Description

While a quickly retrained cellpose network (only on xy slices, no need to train on xz or yz slices) is giving good results in 2D, the anisotropy of the SIM image prevents its usage in 3D. Here the workflow consists in applying 2D cellpose segmentation and then using the CellStich libraries to optimize the 3D labelling of objects from the 2D independant labels.

Here the provided notebook is fully compatible with Google Collab and can be run by uploading your own images to your gdrive. A model is provided to be replaced by your own (create by CellPose 2.0)

has function
example of usage
Description

CellStich proposes a set of tools for 3D segmentation from 2D segmentation: it reassembles 2D labels obtained from cell in slices in unique 3D labels across slices. It isparticularly robust to anisotropy, and is the ideal companion to cellpose 2D models or other 2D deep learning based models. One could also think about using it for cell tracking by overlap (using time as a third dimension).

cellstitch
Description

SuperDSM is a globally optimal segmentation method based on superadditivity and deformable shape models for cell nuclei in fluorescence microscopy images and beyond.

Description

btrack is a Python library for multi object tracking, used to reconstruct trajectories in crowded fields. btrack implemented a residual U-Net model coupledd with a classification CNN to allow accurate instance segmentation of the cell nuclei. To track the cells over time and through cell divisions, btrack developed a Bayesian cell tracking methodology that uses input features from the images to enable the retrieval of multi-generational lineage information from a corpus of thousands of hours of live-cell imaging data.

need a thumbnail
Description

The method proposed in this paper is a robust combination of multi-task learning and unsupervised domain adaptation for segmenting amoeboid cells in microscopy. This end-to-end framework provides a consolidated mechanism to harness the potential of multi-task learning to isolate and segment clustered cells from low contrast brightfield images, and it simultaneously leverages deep domain adaptation to segment fluorescent cells without explicit pixel-level re- annotation of the data.

The entry-point to the codebase is the main.py file. The user has the option to

  • Train the network on their own dataset
  • Load a pre-trained model and use that for inference on their own data
  • NoteThe provided pretrained model was trained on 256x256 images. Results on different resolutions could require fine-tuning This model is trained (supervised) on brightfield, and domain adapted to fluorescence data. The results are saved as 'inference.png'
has function
daman
Description

This workflow describes a deep-learning based pipeline for reliable single-organoid segmentation and tracking in 2D+t high-resolution brightfield microscopy of mouse mammary epithelial organoids. The pipeline involves a four-layer U-Net to infer semantic segmentation predictions, adaptive morphological filtering to establish candidate organoid instances, and a shape-similarity-constrained, instance-segmentation-correcting tracking step to associate the corresponding organoid instances in time.

It is particularly focused on automatically detecting an organoid located approximately in the center of the first frame and track all its subsequent instances in the remaining frames, emphasizing on accurate organoid boundary delineation. Furthermore, segmentation network was trained using plausible pix2pixHD-generated bioimage data. Syntheric image simulator code and data are also available here.

Adapted from https://cbia.fi.muni.cz/research/spatiotemporal/organoids.html
Description

OrganoSeg is an open-source software that integrates segmentation, filtering, and analysis for breast-cancer spheroid and colon and colorectal-cancer organoid morphologies.

Figure 2 in OrganoSeg Scientific Reports publication
Description

OrganoID is an image analysis platform that automatically recognizes, labels, and tracks single organoids, pixel-by-pixel, in brightfield and phase-contrast microscopy experiments. The platform was trained on images of pancreatic cancer organoids and validated on separate images of pancreatic, lung, colon, and adenoid cystic carcinoma organoids.

need a thumbnail

Introduction to 3D Analysis with 3D ImageJ Suite

The 3D ImageJ Suite is a set of algorithms and tools (mostly ImageJ plugins) developed since 2010, originally for 3D analysis of fluorescence microscopy. Since then, the plugins have been widely used and cited more than 200 times in biological journals. In this presentation we will give a general introduction to the tools available in the 3D ImageJ Suite : filtering, 3D segmentation for spots and nuclei, and 3D analysis. A graphical interface to manage 3D objects, the 3DManager, was also developed and will be presented.

GPU Accelerated Image Processing with CLIJ2

The NEUBIAS Academy at home about CLIJ2 gives an introduction to accelerated image processing using Graphics Processing Units (GPUs) in ImageJ/Fiji. Core concepts are explained as well as usage of the tools with the ImageJ Macro recorder and auto-completion in Fijis script editor. Furthermore, an outlook is provided of how the CLIJ project will develop in the coming years to provide long-term maintained access to GPU-acceleration in the Bio-Image Analysis context.

Image Analysis of Biological Data using CellProfiler

After the session you will be able to built your own CellProfiler pipeline, including:

  • Image data import
  • Object segmentation (e.g. detect nuclei in an image) using the modules "IdentifyPrimaryObjects" and "IdentifySecondaryObjects"
  • Object feature measurements (e.g. measure size, shape and intensity of cells)
  • Measurements export to a spreadsheet
  • Creating and saving quality control images
Description

This workflow applies a Stardist pre-trained model (versatile_fluo or versatile_HE) depending on the input images ie. uses both models for a dataset including both fluorescence (grayscale or RGB where all channels are equal) and H&E stained (RGB where channels are not equal) images.

This version uses tensorflow CPU version (See Dockerfile) to ensure compatibility with a larger number of computers. A GPU version should be possible by adapting the Dockerfile with tensorflow-gpu and/or nvidia-docker images.

has topic
has function
need a thumbnail
Description

This workflow processes a group of images containing cells with discernible nuclei and segments the nuclei and outputs a binary mask that show where nuclei were detected. It performs 2D nuclei segmentation using pre-trained nuclei segmentation models of Cellpose. And it was developed as a test workflow for Neubias BIAFLOWS Benchmarking tool.

has topic
has function
need a thumbnail
Description

The Incucyte® Base Analysis Software provides a guided interface and purpose-built tools, which include the process of acquiring, viewing, analyzing and sharing images of living cells.

need a thumbnail
Description

The authors present an ImageJ-based, semi-automated phagocytosis workflow to rapidly quantitate three distinct stages during the early engulfment of opsonized beads.

Description

SMLM is a mature but still growing field, which still lacks efficient and user-friendly analysis and visualization software platform adapted for both users and developers. We here introduce PoCA, a powerful open-source software platform dedicated to the visualization and analysis of 2D and 3D point-cloud data. PoCA allows manipulating large datasets, and integrates a plugin architecture, a native batch analysis engine and a Python code interpreter, facilitating both the analysis of data and the integration of new methods.

Visualization, segmentation and exploration of 3D SMLM data
Description

The empanada-napari plugin is built to democratize deep learning image segmentation for researchers in electron microscopy (EM). It ships with MitoNet, a generalist model for the instance segmentation of mitochondria. There are also tools to quickly build and annotate training datasets, train generic panoptic segmentation models, finetune existing models, and scalably run inference on 2D or 3D data. To make segmentation model training faster and more robust, CEM pre-trained weights are used by default. These weights were trained using an unsupervised learning algorithm on over 1.5 million EM images from hundreds of unique EM datasets making them remarkably general.

Empanada-napari
Description

ASTEC stands for Adaptive Segmentation and Tracking of Embryonic Cells. It proposes a full workflow for time lapse light sheet imaging analysis, including drift/motion compensation before the segmentation itself, and the capacity to correct for it.  It was used to process 3D+t movies acquired by the MuViSPIM light-sheet microscope in particular.

Astec embryon
Description

ClearMap is a toolbox for the analysis and registration of volumetric data from cleared tissues.

It was initially developed to map brain activity at cellular resolution in whole mouse brains using immediate early gene expression. It has since then been extended as a tool for the qunatification of whole mouse brain vascualtur networks at capilary resolution.

It is composed of sevral specialized modules or scripts: tubemap, cellmap, WobblyStitcher.

ClearMap has been designed to analyze O(TB) 3d datasets obtained via light sheet microscopy from iDISCO+ cleared tissue samples immunolabeled for proteins. The ClearMap tools may also be useful for data obtained with other types of microscopes, types of markers, clearing techniques, as well as other species, organs, or samples.

ClearMap SCreenshot
Description

Machine Learning made easy

APEER ML provides an easy way to train your own machine learning
models and segment your microscopy images. No expertise or coding required.

APEER

Description

The tool allows to measure the area of the invading spheroïd in a 3D cell invasion assay. It can also count and measure the area of the nuclei within the spheroïd.

need a thumbnail