Lung Cancer Screening Radiomics

A comprehensive framework for lung cancer screening radiomics using LIDC-IDRI and LUNGx dataset.

  • Data preprocessing – download data, conversion, etc.
  • Radiomics feature extraction including spiculation features
  • AutoML model building and validation

Source code


  1. Wookjin Choi, Jung Hun Oh, Sadegh Riyahi, Chia-Ju Liu, Feng Jiang, Wengen Chen, Charles White, Andreas Rimner, James G. Mechalakos, Joseph O. Deasy, and Wei Lu, “Radiomics analysis of pulmonary nodules in low-dose CT for early detection of lung cancer”, Medical Physics, Vol. 45, No. 4, pp. 1537-1549, April 2018.
  2. Wookjin Choi, Saad Nadeem, Sadegh Riyahi, Joseph O. Deasy, Allen Tannenbaum, Wei Lu, “Reproducible and Interpretable Spiculation Quantification for Lung Cancer Screening.” Computer methods and programs in biomedicine. 200 2021.

Hiring a Postdoctoral Fellow

Postdoctoral Fellow – Developing Clinically Interpretable Medical Imaging AI in Radiation Therapy

  • PI: Wookjin Choi, Ph.D. <>
    Assistant Professor of Radiation Oncology, Thomas Jefferson University
  • 2 Years


POST-DOCTORAL POSITION, DEPARTMENT OF RADIATION ONCOLOGY: Thomas Jefferson University is now accepting applications for a post-doctoral fellow in the Department of Radiation Oncology with the Choi lab.  The post-doctoral position is for developing AI techniques for image-guided radiation therapy and clinical outcome prediction and decision-making using radiomics, deep learning, and other computationally intensive techniques. Trainees must have the opportunity to carry out supervised biomedical research with the primary objective of developing or extending their research skills and knowledge in preparation for an independent research career.

This is an exciting opportunity to work on an emerging research project funded by Sidney Kimmel Cancer Center on the incorporation of medical imaging AI into radiation therapy and clinical care for cancer patients. The research focuses on the development of image analysis tools for cancer imaging, as well as clinically interpretable radiomics features and deep learning models for clinical outcome prediction and decision making. The post-doctoral fellow will have an opportunity to collaborate with Thomas Jefferson faculty, national and international collaborators, and work alongside investigators at the NCI Quantitative Imaging Network.

The entire Thomas Jefferson Medical Physics Division currently consists of 22 physicists and eight physics residents. The Jefferson physics faculty lead a highly-impactful and diverse research program. The group currently has ongoing funded projects by the National Cancer Institute, vendor-funded research, as well as Sidney Kimmel Cancer Center funding. System-wide equipment includes a ViewRay MRI-Linac, Varian TrueBeams, and Elekta Agility Linacs. The department has advanced scripting capabilities with multiple Radiation Oncology software packages including Mosaiq, Eclipse, and MIM Software.

Thomas Jefferson University is an Equal Opportunity Employer.  Jefferson values diversity and encourages applications from women, members of minority groups, LGBTQ individuals, disabled individuals, and veterans. Applicants should forward a curriculum vitae and a statement of interest to the administrative assistant, Juli Johnson at


Candidates must have a Ph.D. in Computer Science, Medical Physics, Physics, Mathematics, Electrical Engineering, Biomedical Engineering, or related field required. The ideal candidate will be highly interested in an academic Medical Physics career, have strong computational skills, and seek out a highly collaborative environment. Based on the interest of the post-doctoral fellow; opportunities will be provided to obtain clinical experience, treatment planning experience, as well as mentorship on clinical trial design and statistical modeling methods.

Conditions of Employment

Covid Vaccination is a requirement for employment at Jefferson for employees working at Jefferson’s clinical entities or at the University.  If you are not currently vac

PathCNN: interpretable convolutional neural networks for survival prediction and pathway analysis applied to glioblastoma

Jung Hun Oh, Wookjin Choi, Euiseong Ko, Mingon Kang, Allen Tannenbaum, Joseph O Deasy

The authors wish it to be known that, in their opinion, Jung Hun Oh and Wookjin Choi should be regarded as Joint First Authors.

An illustration of biological interpretation. (A) Grad-CAM procedure to generate class activation maps. The two images on the left bottom represent an example of the class activation maps for a sample in the cohort, which were generated from Grad-CAM procedure; (B) statistical analysis to identify significantly different pathways between the LTS and non-LTS groups. LTS, long-term survival; CNN, convolutional neural network; ReLU, rectified linear unit



Convolutional neural networks (CNNs) have achieved great success in the areas of image processing and computer vision, handling grid-structured inputs and efficiently capturing local dependencies through multiple levels of abstraction. However, a lack of interpretability remains a key barrier to the adoption of deep neural networks, particularly in predictive modeling of disease outcomes. Moreover, because biological array data are generally represented in a non-grid structured format, CNNs cannot be applied directly.


To address these issues, we propose a novel method, called PathCNN, that constructs an interpretable CNN model on integrated multi-omics data using a newly defined pathway image. PathCNN showed promising predictive performance in differentiating between long-term survival (LTS) and non-LTS when applied to glioblastoma multiforme (GBM). The adoption of a visualization tool coupled with statistical analysis enabled the identification of plausible pathways associated with survival in GBM. In summary, PathCNN demonstrates that CNNs can be effectively applied to multi-omics data in an interpretable manner, resulting in promising predictive power while identifying key biological correlates of disease.Availability and implementation

The source code is freely available at:

Fourth place Winner on AI Tracks at Sea Challenge

We won 4th place in the Artificial Intelligence (AI) Tracks at Sea Challenge.
This national competition is organized by the U.S. Navy.

VSU TrojanOne Team: Jose Diaz, Curtrell Trott, Advisor: Ju Wang, Wookjin Choi

The $200,000 prize was distributed among five winning teams, which submitted full working solutions, and three runners-up, which submitted partial working solutions. The monetary prize will be awarded to the school the corresponding team attends:

Challenge Winners

Teams participating in the AI Tracks at Sea Challenge spanned collegiate institutions from east to west U.S. coasts, from both public and private colleges and universities. Collectively, the student submissions for the challenge represent various types of STEM research institutions, Ivy League Schools, Historically Black Colleges and Universities (HBCU) and Hispanic Serving Institutes (HSI). Of the challenge teams, 26% were comprised of students from HBCUs and 16% of the teams attend HSIs.

“With 94% of the competitors attending colleges and universities outside of California, this challenge served as an avenue to make broader impacts in STEM,” said Yolanda Tanner, Naval Information Warfare Systems Command (NAVWAR) STEM Federal Action Officer and NIWC Pacific Internship and Fellowship project manager. “It was also a means by which students could further develop their STEM skills while working collaboratively to solve a real-world naval problem.”

Florida, North Carolina, and Texas had the largest population of participating collegiate teams.

Automatic motion tracking system for analysis of insect behavior

Darrin Gladman, Jehu Osegbe, Wookjin Choi*, and Joon Suk Lee “Automatic motion tracking system for analysis of insect behavior”, Proc. SPIE 11510, Applications of Digital Image Processing XLIII, 115102W (21 August 2020);

*Corresponding author


We present a multi-object tracking system to track small insects such as ants and bees. Motion-based object tracking recognizes the movements of objects in videos using information extracted from the given video frames. We applied several computer vision techniques, such as blob detection and appearance matching, to track ants. Moreover, we discussed different object detection methodologies and investigated the various challenges of object detection, such as illumination variations and blob merge/split. The proposed system effectively tracked multiple objects in various environments.

Reproducible and Interpretable Spiculation Quantification for Lung Cancer Screening

Choi, W., Nadeem, S., Alam, S. R., Deasy, J. O., Tannenbaum, A., & Lu, W. (2020). Reproducible and Interpretable Spiculation Quantification for Lung Cancer Screening. Computer Methods and Programs in Biomedicine, 105839.


  • A novel interpretable spiculation feature is presented, computed using the area distortion metric from spherical conformal (angle-preserving) parameterization.
  • A simple one-step feature and prediction model is introduced which only uses our interpretable features (size, spiculation, lobulation, vessel/wall attachment) and has the added advantage of using weak-labeled training data.
  • A semi-automatic segmentation algorithm is also introduced for more accurate and reproducible lung nodule as well as vessel/wall attachment segmentation. This leads to more accurate spiculation quantification because the attachments can be excluded from spikes on the lung nodule surface (triangular mesh) data.
  • Using just our interpretable features (size, attachment, spiculation, lobulation), we were able to achieve AUC=0.82 on public Lung LIDC dataset and AUC=0.76 on public LUNGx dataset (the previous LUNGx best being AUC=0.68).
  • State-of-the-art correlation is achieved between our spiculation score (the number of spiculations, Ns) and radiologists spiculation score (ρ = 0.44).


Spiculations are important predictors of lung cancer malignancy, which are spikes on the surface of the pulmonary nodules. In this study, we proposed an interpretable and parameter-free technique to quantify the spiculation using area distortion metric obtained by the conformal (angle-preserving) spherical parameterization. We exploit the insight that for an angle-preserved spherical mapping of a given nodule, the corresponding negative area distortion precisely characterizes the spiculations on that nodule. We introduced novel spiculation scores based on the area distortion metric and spiculation measures. We also semi-automatically segment lung nodule (for reproducibility) as well as vessel and wall attachment to differentiate the real spiculations from lobulation and attachment. A simple pathological malignancy prediction model is also introduced. We used the publicly-available LIDC-IDRI dataset pathologists (strong-label) and radiologists (weak-label) ratings to train and test radiomics models containing this feature, and then externally validate the models. We achieved AUC = 0.80 and 0.76, respectively, with the models trained on the 811 weakly-labeled LIDC datasets and tested on the 72 strongly-labeled LIDC and 73 LUNGx datasets; the previous best model for LUNGx had AUC = 0.68. The number-of-spiculations feature was found to be highly correlated (Spearman’s rank correlation coefficient ) with the radiologists’ spiculation score. We developed a reproducible and interpretable, parameter-free technique for quantifying spiculations on nodules. The spiculation quantification measures was then applied to the radiomics framework for pathological malignancy prediction with reproducible semi-automatic segmentation of nodule. Using our interpretable features (size, attachment, spiculation, lobulation), we were able to achieve higher performance than previous models. In the future, we will exhaustively test our model for lung cancer screening in the clinic.

Assessing the Dosimetric Links between Organ-At-Risk Delineation Variability and Treatment Planning Variability

The 2020 Joint AAPM | COMP Virtual Meeting