Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/saalfeldlab/n5-ij
ImageJ convenience layer for N5
https://github.com/saalfeldlab/n5-ij
fiji fiji-plugin hdf5 n5 ome-zarr-converter zarr
Last synced: 3 months ago
JSON representation
ImageJ convenience layer for N5
- Host: GitHub
- URL: https://github.com/saalfeldlab/n5-ij
- Owner: saalfeldlab
- License: bsd-2-clause
- Created: 2018-02-27T16:23:57.000Z (almost 7 years ago)
- Default Branch: master
- Last Pushed: 2024-10-30T12:05:12.000Z (3 months ago)
- Last Synced: 2024-10-30T13:19:13.449Z (3 months ago)
- Topics: fiji, fiji-plugin, hdf5, n5, ome-zarr-converter, zarr
- Language: Java
- Homepage:
- Size: 2.76 MB
- Stars: 15
- Watchers: 11
- Forks: 11
- Open Issues: 22
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- awesome-janelia-software - n5-ij - Fiji plugin for loading and saving image data as N5 data sets (Fiji Plugins)
- awesome-janelia-software - n5-ij - Fiji plugin for loading and saving image data as N5 data sets (Fiji Plugins)
README
# n5-ij [![Build Status](https://github.com/saalfeldlab/n5-ij/actions/workflows/build-main.yml/badge.svg)](https://github.com/saalfeldlab/n5-ij/actions/workflows/build-main.yml)
A Fiji plugin for loading and saving image data to N5 containers. Supports [HDF5](https://www.hdfgroup.org/solutions/hdf5/), [Zarr](https://zarr.readthedocs.io/en/stable/#), [Amazon S3](https://aws.amazon.com/s3/), and [Google cloud storage](https://cloud.google.com/storage).
## Contents
1. [Open HDF5/N5/Zarr/OME-NGFF](#open-n5)
1. [Virtual](#virtual)
2. [Cropping](#cropping)
2. [Export N5](#export-n5)
3. [Container types](#container-types)
4. [Metadata](#metadata)
5. [For developers](#for-developers)
6. [Details](#details)
1. [Cloud writing benchmarks](#cloud-writing-benchmarks)## Open HDF5/N5/Zarr/OME-NGFF
Open HDF5/N5/Zarr/OME-NGFF datasets from Fiji with `File > Import > HDF5/N5/Zarr/OME-NGFF ... `.
Quickly open a dataset by pasting the full path to the dataset and press `OK`.
For example, try `gs://example_multi-n5_bucket/mitosis.n5/raw` to open the sample mitosis image from Google
cloud storage.Click the `Browse` button to select a folder on your filesystem.
The detected datasets will be displayed in the dialog. Select (highlight) the datasets you would like to open
and press `OK`. In the example below, we will open the datasets `/blobs`, and `/t1-head/c0/s0`.### Virtual
Check the `Open as virtual` box to open the dataset as a [virtual stack in ImageJ](https://imagej.nih.gov/ij/docs/guide/146-8.html#toc-Section-8).
This enable the opening and viewing of image data that do not fit in RAM. Image slices are loaded on-the-fly, so
navigation will be slow when parts of the images are loaded.### Cropping
Subsets of datasets can be opened by checking the `Crop` box in the dialog, then pressing `OK`.
A separate dialog will appear for each selected dataset as shown below.Give the min and max values for the field-of-view to open **in pixel / voxel units** a particular
subset. The opened interval includes both min and max values, so the image will be of size `max - min + 1` along
each dimension. In the example shown above, the resulting image will be of size `101 x 111 x 2 x 51`.## Export HDF5/N5/Zarr/OME-NGFF
Save full images opened in Fiji as HDF5/N5/Zarr/OME-NGFF datasets with `File > Save As > Export HDF5/N5/Zarr/OME-NGFF ...`, and patch images into an existing dataset using `File > Save As > Export HDF5/N5/Zarr/OME-NGFF (patch)`. The patch export exports the current image into an existing dataset with a given offset. This offset and the size of the image do not need to align with the block raster of the dataset.
Parameters
* `Root url` - the root location of the n5 (see also [Container types](#container-types))
* `Dataset` - the name of the dataset.
* `Format` - the storage format to use: one of `Auto`, `Zarr`, `N5`, or `HDF5`
* `Auto` : try to infer the storage format from the url (see below)
* `Chunk size` - chunk/block size as comma-separated list.
* ImageJ's axis order is X,Y,C,Z,T. The chunk size must be specified in this order. You must skip any axis whose size is `1`, e.g. a 2D time-series without channels may have a chunk size of `1024,1024,1` (X,Y,T).
* You may provide fewer values than the data dimension. In that case, the list will be expanded to necessary size with the last value, for example `64`, will expand to `64,64,64` for 3D data.
* `Create Pyramid` - If checked, a multiscale pyramid will be created (if possible). See below for details.
* `Downsampling method` - The downsampling method to be used if a multiscale pyramid can be created. See below for details.
* `Compression` - The compression method to be used for chunks / blocks.
* `metadata type` - style and type of metadata to store (see also [Metadata](#metadata))
* `Thread count` - number of threads used for parallel writing (see also [Cloud writing benchmarks](#cloud-writing-benchmarks))
* `Overwrite` - If checked, existing data may be deleted and overwritten without warning.### Container types
If the "Format" option is set to `Auto`, the export plugin infers the storage format from the given url given. First, it checks scheme
of the URL, next it checks the directory or file extension. Note that URLs may have two schemes [(as in neuroglancer)](https://connectomics.readthedocs.io/en/latest/external/neuroglancer.html#basic-usage),
for example: `zarr://s3://my-bucket/my-key`* Filesystem N5
* Specify a URL starting with `n5:`
* Specify a directory ending in `.n5`
* example `/path/to/my/data.n5`
* Zarr
* Specify a URL starting with `zarr:`
* Specify a directory ending in `.zarr`
* example `/Users/user/Documents/sample.zarr`
* HDF5
* Specify a URL starting with `hdf5:`
* Specify a file ending in `.h5` ,`.hdf5`, or `.hdf`
* example `C:\user\docs\example.h5`
### BackendSpecify the backend by protocol, "file:" or not protocol indicate the local file system:
* Amazon S3
* Specify one of two url styles:
* `s3://bucket-name/path/to/root.n5`
* `https://bucket-name.s3.amazonaws.com/path/to/root.n5`
* Google cloud storage (one of two url styles)
* Specify one of two url styles:
* `gs://bucket-name/path/inside/bucket/root.n5`
* `https://bucket-name.s3.amazonaws.com/path/to/root.n5`
## Multi-scale pyramids### How many scale levels will be created
The number of scale levels is determined by the image size and specified block size.
The exporter will downsample an image only if the block size is striclty smaller than
the image size in every dimension.#### Example 1
If image size is `100 x 100 x 100` and the block size is `100 x 100 x 100`
will write one scale level because the whole image fits into one block
at the first scale level.#### Example 2
If image size is `100 x 100 x 100` and the block size is `64 x 64 x 64`
will write two scale levels: The first scale level will have a `2 x 2 x 2` grid
of blocks.#### Example 3
If image size is `100 x 100 x 32` and the block size is `64 x 64 x 64`
will write one scale level because the third dimension of the image is
smaller than the third dimension of the block.### Downsampling
The N5 exporter always downsamples images by factors of two in all dimensions.
There are two downsampling methods:#### Sample
N5 will take even-indexed samples and discard odd-indexed samples along
each dimension.#### Averaging
N5 will average adjacent samples along each dimension. This results in a
"half-pixel" shift, which will be reflected in the metadata.## Overwriting
### Warning messages
If the `overwrite` option is not selected in the dialog, the exporter will determine if the
write operation would overwrite or invalidate existing data. If so, it prompts the user with a
warning dialog, asking if data should be overwritten.Data could be overwritten if:
* the path where data will be written already exists and contains data.
* a parent of the path where data will be written already exists and contains data.
* here, the newly written data would be inaccessible, because data arrays must be leafs of the hierarchy tree.If the `overwrite` option is selected in the initial dialog, the user will not be prompted, but
data will be overwritten if needed as explained below.
#### Example 1A dataset exists at `/image`. User attempts to write data into `/image`. This warns the user
about overwriting because an array already exists at that location.#### Example 2
A dataset exists at `/data/s0`. User attempts to write data into `/data` using N5Viewer metadata.
This warns the user about overwriting because when writing N5Viewer metadata, the plugin will
write the full resolution level of the multiscale pyramid at location `/data/s0`, but an array
already exists at that location.### Overwriting removes existing data
If the user decides to overwite data the N5 exporter will completely (array data and metadata)
remove any groups that cause conflicts before writing the new data.* If a dataset already exists at a path where new data will be written, then all data at that path will be removed.
* If a dataset already exists at a parent path where new data will be written, then all data at that parent path will be removed.#### Example 3
A dataset exists at `/image`. User attempts to write data into `/image/channel0`. This warns
the user about overwriting because the newly written data would be a child path of existing
data, and therefore be invalid. If the user decides to overwrite, all data at `/image` will be
removed before writing the new data to `/image/channel0`.## Metadata
This plugin supports three types of image metadata:
1) ImageJ-style metadata
2) [N5-viewer](https://github.com/saalfeldlab/n5-viewer) metadata
3) [COSEM](https://github.com/janelia-cosem/schemas/blob/master/multiscale.md) metadata
4) [OME-NGFF v0.4](https://ngff.openmicroscopy.org/0.4/) metadata
5) Custom metadata. [Read details here](CustomMetadata.md)The metadata style for exported HDF5/N5/Zarr/OME-NGFF datasets is customizable, more detail coming soon.
## For developers
ImageJ convenience layer for N5
Build into your Fiji installation:
```bash
mvn -Dscijava.app.directory=/home/saalfelds/packages/Fiji.app -Ddelete.other.versions=true clean install
```Then, in Fiji's Script Interpreter (Plugins > Scripting > Script Interpreter), load an N5 dataset into an `ImagePlus`:
```java
import org.janelia.saalfeldlab.n5.*;
import org.janelia.saalfeldlab.n5.ij.*;imp = N5IJUtils.load(new N5FSReader("/home/saalfelds/example.n5"), "/volumes/raw");
```or save an `ImagePlus` into an N5 dataset:
```java
import ij.IJ;
import org.janelia.saalfeldlab.n5.*;
import org.janelia.saalfeldlab.n5.ij.*;N5IJUtils.save(
IJ.getImage(),
new N5FSWriter("/home/saalfelds/example.n5"),
"/volumes/raw",
new int[] {128, 128, 64},
new GzipCompression()
);
```Save an image stored locally to cloud storage (using four threads):
```java
final ImagePlus imp = IJ.openImage( "/path/to/some.tif" );
final ExecutorService exec = Executors.newFixedThreadPool( 4 );
N5IJUtils.save( imp,
new N5Factory().openWriter( "s3://myBucket/myContainer.n5" ),
"/myDataset",
new int[]{64, 64, 64},
new GzipCompression(),
exec );
exec.shutdown();
```See also scripts demonstrating
* [how to read and write imglib2 images with the methods in `N5Utils`](https://github.com/saalfeldlab/n5-imglib2/blob/master/scripts/readProcessWriteDemo.bsh)
* [how to read and write ImageJ images with the methods in `N5IJUtils`](https://github.com/saalfeldlab/n5-ij/blob/master/scripts/readProcessWriteIJDemo.bsh)## Details
* This plugin supports images of up to 5 dimensions, and the datatypes supported by ImageJ (`uint8`, `uint16`, `float32`) For higher dimensions and other datatypes, we recommend [n5-imglib2](https://github.com/saalfeldlab/n5-imglib2).
* This plugin supports only the datatypes supported by ImageJ, namely uint8, uint16, and float32. For other datatypes, use [n5-imglib2](https://github.com/saalfeldlab/n5-imglib2).
### Cloud writing benchmarks
Below are benchmarks for writing images of various sizes, block sizes, and with
increasing amount of parallelism.#### Amazon S3
Time in seconds to write the image data. Increased parallelism speeds
up writing substantially when the total number of blocks is high.| Image size | Block size | 1 thread | 2 threads | 4 threads | 8 threads | 16 threads |
| ------------ | ---------- | ---------- | ----------- | ----------- | ----------- | ----------- |
| 64x64x64 | 32x32x32 | 0.98 | 0.60 | 0.45 | 0.50 | 0.51 |
| 128x128x128 | 32x32x32 | 4.72 | 2.64 | 1.62 | 1.00 |
| 256x256x256 | 32x32x32 | 37.09 | 19.11 | 9.09 | 5.20 | 3.2 |
| 256x256x256 | 64x64x64 | 10.56 | 5.04 | 3.23 | 2.17 | 1.86 |
| 512x512x512 | 32x32x32 | 279.28 | 156.89 | 74.72 | 37.15 | 19.77 |
| 512x512x512 | 64x64x64 | 76.63 | 38.16 | 19.86 | 10.16 | 6.14 |
| 512x512x512 | 128x128x128 | 27.16 | 14.32 | 8.01 | 4.70 | 3.31 |
| 1024x1024x1024 | 32x32x32 | 2014.73 | 980.66 | 483.04 | 249.83 | 122.36 |
| 1024x1024x1024 | 64x64x64 | 579.46 | 289.53 | 149.98 | 75.85 | 38.18 |
| 1024x1024x1024 | 128x128x128 | 203.47 | 107.23 | 55.11 | 27.41 | 15.33 |