README first commit
#2
by
egrandgi
- opened
README.md
CHANGED
|
@@ -1,3 +1,53 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# NucGen3D: A Synthetic Framework for Large-Scale 3D Nuclear Segmentation with Open-Source Training Data and Models
|
| 2 |
+
|
| 3 |
+
**NucGen3D** is an open dataset of realistically simulated, annotated 3D microscopy-like images of cell nuclei.
|
| 4 |
+
It was generated using a procedural simulation framework designed to reproduce the structural and visual complexity of real fluorescence microscopy data.
|
| 5 |
+
|
| 6 |
+
The dataset provides paired 3D images and ground truth masks, enabling training and benchmarking of 3D segmentation models.
|
| 7 |
+
It addresses the scarcity of large, well-annotated 3D datasets in bioimage analysis by offering controllable, unbiased, and reproducible training data.
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
## 📂 Dataset structure
|
| 11 |
+
This repository hosts:
|
| 12 |
+
- The NucGen3D dataset (40 000 3D images ≈ 10 M nuclei)
|
| 13 |
+
- The noise templates used for augmentation and realistic image generation (Perlin and anisotropic noise)
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
## 🚀 What you can do
|
| 17 |
+
- Use the dataset for training or benchmarking 3D nuclear segmentation models
|
| 18 |
+
- Generate new synthetic 3D images using the NucGen3D simulator (code on GitHub)
|
| 19 |
+
- Augment existing data with our provided noise templates (Perlin / anisotropic) for more realistic training
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
## 💻 Example: loading a NucGen3D image and applying random noise
|
| 23 |
+
Code available on [GitHub](https://github.com/mathieuserr/nucgen3D/tree/main).
|
| 24 |
+
```python
|
| 25 |
+
from torch.utils.data import DataLoader
|
| 26 |
+
from nucgen3d.dataset.loader import SimImageNoiseDataset
|
| 27 |
+
|
| 28 |
+
# Example use
|
| 29 |
+
ds = SimImageNoiseDataset(
|
| 30 |
+
img_dir="data/simulated/images", # directory containing simulated images
|
| 31 |
+
noise_dirs=["data/noise2", "data/noise3", "data/noise_aniso1"], # str or list[str] - noise templates directory (.tif) - Perlin/anisotropic
|
| 32 |
+
crop_size=256, # example: 256, None = full image
|
| 33 |
+
z_slices=8, # example: 8 -> number of z slices
|
| 34 |
+
quant_prob=0.3, # Noisator quantification parameter
|
| 35 |
+
background_prob=0.7,
|
| 36 |
+
background_coeff_max=0.6,
|
| 37 |
+
readout_max=0.03,
|
| 38 |
+
random_shot=0.5,
|
| 39 |
+
)
|
| 40 |
+
|
| 41 |
+
dl = DataLoader(ds, batch_size=4, shuffle=True, num_workers=0)
|
| 42 |
+
noisy, clean, names = next(iter(dl))
|
| 43 |
+
```
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
## Publication
|
| 47 |
+
|
| 48 |
+
📄 Preprint will be available on bioRxiv by October 2025.
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
## 🙏 Acknowledgements
|
| 52 |
+
|
| 53 |
+
This work was developed within the RESTORE (INSERM, Université de Toulouse) and IRIT (Université de Toulouse) laboratories, as part of the ANITI program and the CALM research chair.
|