Safetensors
tapct
custom_code
TimVeenboer commited on
Commit
63bb3bd
·
1 Parent(s): 3758fa2

docs(tap-hf): Update README and processor config

Browse files
Files changed (2) hide show
  1. README.md +23 -1
  2. preprocessor_config.json +1 -0
README.md CHANGED
@@ -3,6 +3,7 @@ license: cc-by-nc-4.0
3
  ---
4
 
5
  # TAP-CT: 3D Task-Agnostic Pretraining of CT Foundation Models
 
6
 
7
  TAP-CT is a suite of foundation models for computed tomography (CT) imaging, pretrained in a task-agnostic manner through an adaptation of DINOv2 for volumetric data. These models learn robust 3D representations from CT scans without requiring task-specific annotations.
8
 
@@ -53,6 +54,14 @@ with torch.no_grad():
53
 
54
  ### Usage with Preprocessor, loading CT volumes & slice-wise inference
55
 
 
 
 
 
 
 
 
 
56
  ```python
57
  import numpy as np
58
  import SimpleITK as sitk
@@ -69,7 +78,7 @@ volume = sitk.DICOMOrient(volume, 'LPS')
69
 
70
  # Get array, expand to (B, C, D, H, W) and preprocess
71
  array = sitk.GetArrayFromImage(volume)
72
- array = np.expand_dims(array, axis(0, 1))
73
  x = preprocessor(array)['pixel_values']
74
 
75
  # Forward pass
@@ -104,3 +113,16 @@ The model returns a `BaseModelOutputWithPooling` object from the transformers li
104
  - **Input Shape**: `(batch_size, 1, height, width)`
105
  - **Example Input**: `(16, 1, 224, 224)` - batch of 16 CT slices at 224×224 resolution
106
  - **License**: CC-BY-NC-4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
 
5
  # TAP-CT: 3D Task-Agnostic Pretraining of CT Foundation Models
6
+ [![arXiv](https://img.shields.io/badge/arXiv-TAP--CT-b31b1b.svg)](https://arxiv.org/abs/2512.00872)
7
 
8
  TAP-CT is a suite of foundation models for computed tomography (CT) imaging, pretrained in a task-agnostic manner through an adaptation of DINOv2 for volumetric data. These models learn robust 3D representations from CT scans without requiring task-specific annotations.
9
 
 
54
 
55
  ### Usage with Preprocessor, loading CT volumes & slice-wise inference
56
 
57
+ **Recommended environment:**
58
+ - Python >= 3.11
59
+ - torch >= 2.8
60
+ - numpy >= 2.35
61
+ - SimpleITK >= 2.52
62
+ - monai >= 1.4.0
63
+ - xformers >= 0.0.32 (optional, recommended for CUDA)
64
+
65
  ```python
66
  import numpy as np
67
  import SimpleITK as sitk
 
78
 
79
  # Get array, expand to (B, C, D, H, W) and preprocess
80
  array = sitk.GetArrayFromImage(volume)
81
+ array = np.expand_dims(array, axis=(0, 1))
82
  x = preprocessor(array)['pixel_values']
83
 
84
  # Forward pass
 
113
  - **Input Shape**: `(batch_size, 1, height, width)`
114
  - **Example Input**: `(16, 1, 224, 224)` - batch of 16 CT slices at 224×224 resolution
115
  - **License**: CC-BY-NC-4.0
116
+
117
+ ## Citation
118
+
119
+ If you find this work useful, please cite:
120
+
121
+ ```bibtex
122
+ @article{veenboer2025tapct,
123
+ title={TAP-CT: 3D Task-Agnostic Pretraining of Computed Tomography Foundation Models},
124
+ author={Veenboer, Tim and Yiasemis, George and Marcus, Eric and Van Veldhuizen, Vivien and Snoek, Cees G. M. and Teuwen, Jonas and Groot Lipman, Kevin B. W.},
125
+ journal={arXiv preprint arXiv:2512.00872},
126
+ year={2025}
127
+ }
128
+ ```
preprocessor_config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "image_processor_type": "TAPCTProcessor",
 
3
  "resize_dims": [224, 224],
4
  "divisible_pad_z": 1,
5
  "clip_range": [-1008.0, 822.0],
 
1
  {
2
  "image_processor_type": "TAPCTProcessor",
3
+ "use_fast": false,
4
  "resize_dims": [224, 224],
5
  "divisible_pad_z": 1,
6
  "clip_range": [-1008.0, 822.0],