2026 - Thesis - Accelerate 3D Brain Mapping

Alzheimer's Disease, warp


Readings


Comparison between Conventional Structural MRI and Quantitative Susceptibility Mapping (QSM MRI)

Property Conventional Structural MRI (e.g., T1/T2) Quantitative Susceptibility Mapping (QSM MRI)
Signal origin Measures proton relaxation properties (T1 and T2 relaxation times) following RF excitation. Measures phase shifts induced by local variations in the proton Larmor precession frequency.
Primary contrast mechanism Differences in longitudinal and transverse relaxation of hydrogen protons. Local magnetic field perturbations caused by tissue magnetic susceptibility.
What it depicts Macroscopic anatomical structures (e.g., gray matter–white matter boundaries, cortical thickness). The absolute magnetic susceptibility (Ο‡) distribution of tissue.
Physical nature Qualitative or semi-quantitative signal intensity (relative brightness or darkness). Quantitative physical parameter expressed in parts per million (ppm).
Biophysical interpretability Indirect and non-specific; contrast reflects multiple tissue properties simultaneously. Directly linked to underlying biophysical sources of magnetism in tissue.
Sensitivity to pathology Sensitive to gross structural changes such as atrophy or lesions, but largely insensitive to early molecular pathology. Highly sensitive to paramagnetic substances (e.g., iron deposition) and diamagnetic components (e.g., calcification or protein aggregates).


Topics


Coding


Alzheimer’s Disease Neuroimaging Initiative (ADNI)

  • A large-scale longitudinal multi-center study initiated in 2004. The dataset includes 3D brain MRI and PET images with associated diagnostic labels and clinical metadata, and is publicly available via the ADNI Image and Data Archive under a data use agreement
  • ADNI Database
  • The essence of Alzheimer’s disease (AD) is the breakdown of neuronal connections caused by the deposition of amyloid plaques at the microscopic level, PATHFINDER (bioRxiv 2025) addresses how to precisely reconstruct damaged neurons, QSM/MRI Framework (Arxiv 2503) addresses how to quantify plaque burden in vivo using imaging
  • Data alignment: Microscopic data (PATHFINDER) and MRI data (ADNI) differ in spatial scale by several orders of magnitude. Instead of directly feeding them into the same model, you need to learn their representation mapping, 3D U-Net or A Medical GAN
  • Python + PyTorch (deep learning) + ANTs (image registration) + MEDI (QSM reconstruction)
  • Based on 3D deep learning, Spatial Mapping Reconstruction from QSM magnetic signals to Amyloid pathological signals is achieved, Why:
    • PET scan: Can directly visualize amyloid plaques in the brain, but it is expensive, involves radiation, and is not available in many hospitals
    • QSM MRI (Input): A newer MRI technique, highly sensitive to magnetic materials in the brain (such as iron deposits and plaques). It is inexpensive and safe
    • Thesis task: Use AI to find patterns between QSM signals and PET plaque distribution.


ADNI Cohort
β”‚
β”œβ”€β”€ QSM MRI (in vivo)
β”‚     β”œβ”€β”€ QSM reconstruction & normalization
β”‚     β”œβ”€β”€ Spatial registration to PET space
β”‚     └── 3D volume cropping / resampling
β”‚
β”œβ”€β”€ Amyloid PET (reference standard)
β”‚     β”œβ”€β”€ ADNI-standard preprocessing
β”‚     β”œβ”€β”€ Intensity normalization
β”‚     └── Co-registration with QSM
β”‚
β–Ό
3D QSM Volume
β”‚
β–Ό
Encoder: BrainIAC-Pretrained 3D Vision Transformer
β”‚   (global contextual representation learning)
β”‚
β–Ό
Latent Cross-Modal Representation
β”‚
β–Ό
Alignment Module
β”‚   (Conditional Diffusion or GAN-based refinement)
β”‚
β–Ό
Decoder
β”‚
β–Ό
Predicted Amyloid Burden Map
(continuous voxel-wise 3D estimate)
β”‚
β–Ό
Loss Optimization
β”‚   β”œβ”€β”€ Structural Similarity (SSIM)
β”‚   β”œβ”€β”€ Perceptual Loss (VGG-based)
β”‚   └── Intensity Consistency Loss
β”‚
β–Ό
Voxel-wise Quantification of Cerebral Amyloid Plaque Burden


1. Overview of the ADNI Dataset

Item Description
Study Name Alzheimer’s Disease Neuroimaging Initiative (ADNI)
Start Year 2004
Current Phase ADNI4
Phases ADNI1, ADNIGO, ADNI2, ADNI3, ADNI4
Study Type Longitudinal, multi-center, multi-modal
Primary Goal Early detection and progression modeling of Alzheimer’s disease
Access IDA portal (login + Data Use Agreement required)


2. Participant Identifiers and Longitudinal Indexing

Field Description Usage
PTID Participant ID (format: XXX_S_XXXXX) Primary key across all tables
RID Numeric subject ID derived from PTID Easier joins and indexing
VISDATE / EXAMDATE / SCANDATE Visit / exam / scan date Temporal alignment for longitudinal analysis
Phase Indicator ADNI1 / GO / 2 / 3 / 4 Cohort and protocol stratification


3. Diagnostic Group Distribution

Group Description Number of Subjects
CN Cognitively Normal 1,272
SMC Significant Memory Concern 97
EMCI Early Mild Cognitive Impairment 315
LMCI Late Mild Cognitive Impairment 180
MCI (total) EMCI + LMCI 1,006
AD Alzheimer’s Disease 523
Total Patients All non-CN subjects 141


4. Neuroimaging Data (Raw and Processed)

Modality Access Path Format Dimensionality Typical Use
Structural MRI Advanced Image Search DICOM / NIfTI 3D Brain atrophy analysis, 3D CNN
Functional MRI Advanced Image Search NIfTI 4D Functional connectivity
Amyloid PET Advanced Image Search DICOM / NIfTI 3D Amyloid burden estimation
FDG-PET Advanced Image Search DICOM / NIfTI 3D Glucose metabolism analysis
Pathology Slides Advanced Image Search Whole-slide images 2D/3D Neuropathological validation


Brain Signals (Why Median + MAD)

Property Meaning Impact
Non-stationary The mean varies across time and sessions Mean and standard deviation become unstable
Heavy-tailed distribution Strong artifacts or high-amplitude spikes Standard deviation is inflated by outliers
Weak signal + mixed noise High-frequency oscillations + low-frequency drift Large mean variation, clear skewness
Inter-channel variation Each sensor has different sensitivity Requires independent per-channel normalization



References


Multi-sensor Input Fusion From Space, Safety Detection


Topics

0. Sensor Modalities and Data Types

Modality Sensor Type Data Representation
Optical Visible-light satellite camera 3-channel RGB image (8-bit)
SAR Synthetic Aperture Radar 1-channel SAR image (32-bit float)


1. Maritime Search and Rescue

Optical satellite images
+ SAR satellite images
β†’ Ship Detection
β†’ Ship Re-Identification (ReID)
β†’ Trajectory generation & route prediction
Platform Strength Fundamental Limitation
GEO satellites Wide coverage, high temporal resolution Low spatial resolution
Video satellites High spatial & temporal resolution Short duration, small coverage
AIS-based systems Accurate identity info Only works for cooperative targets
Axis Examples
Sensors Optical, SAR, LiDAR, multispectral
Tasks Detection, ReID, tracking, mapping
Scale Local β†’ Global
Time Snapshot β†’ Long-term monitoring


2. Input Data Type

Modality Data Type Format
Optical RGB image 3-channel, 8-bit TIF
SAR Radar backscatter 1-channel, 32-bit float TIF
Geometry Ship size (derived) Numeric vector (length, width, aspect ratio)


3. Fusion Space

Optical image ─┐
               β”œβ”€ Dual-head tokenizer β†’ Shared Transformer Encoder β†’ Unified embedding
SAR image     β”€β”˜


4. Output Data

Stage Output Used
ReID Feature distance matrix
Tracking Identity association
Trajectory Time-ordered identity matches


A Dynamic Camera with Multi-modal Input Signal Fusion

          Human perception
        β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
        β”‚  Vestibular      β”‚
        β”‚  Vision          β”‚
        β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                 β–²
                 β”‚
    β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
    β”‚ Wearable System Estimation  β”‚
    β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                 β”‚
 β”Œβ”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”
 β”‚ Cameraβ”‚  IMU   β”‚ Eye    β”‚ Depth  β”‚ Others β”‚
 β”‚       β”‚        β”‚ trackerβ”‚ / ToF  β”‚        β”‚
 β””β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜


Core Formulation: Bayesian Multi-Modal Sensor Fusion

Latent State Definition

  • At time step t, the latent state is defined as: $x_t = { T_t, \theta, \psi_t }$

  • where $T_t$ denotes the device pose, $\theta$ represents the calibration parameters shared across time, and $\psi_t$ denotes user-centric latent variables.

Multi-Modal Observations

  • Given heterogeneous sensor measurements at time t: $z_t = { z_t^{cam}, z_t^{imu}, z_t^{eye} }$

  • where observations are obtained from the camera, IMU, and eye-tracking modalities.

Bayesian Fusion Objective

  • Multi-modal fusion is defined as inference over the joint posterior: $p(x_{1:T} \mid z_{1:T})$

  • Using the Markov assumption and conditional independence of observations, the posterior factorizes as: $p(x_{1:T} \mid z_{1:T}) \propto \prod_{t=1}^{T} p(z_t \mid x_t)\, p(x_t \mid x_{t-1})$

Multi-Modal Likelihood Factorization

  • Assuming conditional independence between sensor modalities given the latent state: $p(z_t \mid x_t) = p(z_t^{cam} \mid x_t)\, p(z_t^{imu} \mid x_t)\, p(z_t^{eye} \mid x_t)$

State Transition Model

  • The temporal evolution of the latent state is modeled as: $p(x_t \mid x_{t-1}) = p(T_t \mid T_{t-1})\, p(\psi_t \mid \psi_{t-1})\, p(\theta)$

  • where $\theta$ is treated as a time-invariant latent variable, $p(\theta)$ enforces temporal consistency of calibration parameters.

Interpretation

  • Fusion thus corresponds to Bayesian state estimation under uncertainty, where heterogeneous sensor observations impose probabilistic constraints on a shared latent state evolving over time. Calibration parameters are inferred jointly with pose and user states, enabling online self-calibration.

Sensor Models

  • $z_t^{imu} = h_{imu}(T_{t-1}, T_t) + \epsilon_{imu}$
  • $z_t^{cam} = h_{cam}(T_t, \theta) + \epsilon_{cam}$
  • $z_t^{eye} = h_{eye}(T_t, \psi_t) + \epsilon_{eye}$

Filtering Approximation

For online inference, we approximate the posterior using Bayesian filtering.

  • Prediction: $p(x_t \mid z_{1:t-1}) = \int p(x_t \mid x_{t-1}) p(x_{t-1} \mid z_{1:t-1}) dx_{t-1}$
  • Update: $p(x_t \mid z_{1:t}) \propto p(z_t \mid x_t) p(x_t \mid z_{1:t-1})$


Multiple sensors = multiple Gaussian constraints on the same state

                    z_t^cam
                 (camera likelihood)
                        β”‚
                        β–Ό
                   β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”
z_t^imu ───────▢   β”‚  x_t   β”‚   ◀────── z_t^eye
(IMU likelihood)   β”‚ latent β”‚   (eye-tracking likelihood)
                   β”‚ state  β”‚
                   β””β”€β”€β”€β”€β”€β”€β”€β”€β”˜


4D and LiDAR Free


References