Skip to content

reascr/Enhancing-Image-Segmentation-with-Eye-tracking

Repository files navigation

Enhancing-Image-Segmentation-with-Eye-tracking

This repository contains code and data used for the final project in the class "Computational Cognitive Science III" at the University of Copenhagen.

Analysis-of-Fixation-Patterns

  • Determining-the-dominant-eye.ipynb
    Code to determine the dominant eye of a viewer.
  • analysis-of-fixation-patterns.ipynb
    Code to analyse which fixation point (longest duration, first, last, etc.) usually lies on the object.
  • one-fixation-point-for-sam.ipynb
    Code to explore whether the fixation point with the longest duration yields better segmentation masks.
  • Comparison-Accuracies-attention-points.ipynb
    Code to compute attention points accuracies within segmentation masks
  • Extract_fixation_point.ipynb
    Code to extract fixation point with longest duration per viewer and image

Attention Points

This folder contains json files containing attention points for fixation data, and for the faster_rcnn and fcn_resnet model.

Datasets

This folder contains the eye tracking data (fixation points) extracted from the POET dataset.

  • eye_tracking_data.csv
  • eye_tracking_data.pkl.zip

Data-Preprocessing

  • exploring-eye-tracking-data-POET.ipynb
    Code to analyse the eye tracking data of the POET data set and store it as a .csv and .pkl file for further processing.

SAM

  • compare-segmentations-with-ground_truth.ipynb
    Code to calculate the Dice coefficient between segmentation masks generated by passing SAM human fixation points and points generated by Ablation or GradCAM. Addiotnally, performs Wilcoxon signed-rank tests to see whether differences in Dice coefficients between human and model generated points are significant.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •