8000 GitHub - aoteman233/PSI-Intention: Data preprocessing for IUPUI-CSRC Pedestrian Situated Intent (PSI) benchmark dataset.
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

aoteman233/PSI-Intention

 
 

Repository files navigation

For more situated intent data and work, please see Situated Intent!


IUPUI-CSRC Pedestrian Situated Intent (PSI) Dataset

This repository contains IUPUI-CSRC Pedestrian Situated Intent (PSI) Dataset pre-processing and baseline[to be uploaded].

Download dataset and extract

Download the dataset from link, then extract via

unzip Dataset.zip

Output:

Archive:  Dataset.zip
creating: PSI_Intention/Dataset/ 
inflating: PSI_Intention/Dataset/VideoWithIndicator.zip  
inflating: PSI_Intention/Dataset/RawVideos.zip  
inflating: PSI_Intention/Dataset/README.txt  
inflating: PSI_Intention/Dataset/IntentAnnotations.xlsx
inflating: PSI_Intention/Dataset/XmlFiles.zip 

Extract videos and spatial annotations:

unzip ./PSI_Intention/Dataset/RawVideos.zip -d ./PSI_Intention/Dataset
unzip ./PSI_Intention/Dataset/XmlFiles.zip -d ./PSI_Intention/Dataset

Video to frames

python split_clips_to_frames.py

The splited frames are organized as, e.g.,

frames{
    video_0001{
        000.jpg,
        001.jpg,
        ...
    }
}

CV_annotations and NLP_annotations re-organize

python reorganize_annotations.py

Note: video_0060 and video_0093 are removed due to the missing of spatial segmentation annotations.

Create database with frames labeled

python pedestrian_intention_database_processing.py

Output:

  • database_*.pkl: The annotaions of reasoning and intention do not exactly match, i.e., the last several frames only have intention annotations without reasoning, because the reasoning is only for the previous time period before the last annotated time point, while the intention annotation lasts till the end of the video.
  • database_*_overlap.pkl: By removing the last frames only with intention labels, the length of annotated reasoning and intention frames are Equal now.

Train/Val/Test split

  • train: [1 ~ 75]
  • val: [76 ~ 80]
  • test: [81 ~ 110]

Note: Due to the missing of spatial segmentation annotations, video_0060 and video_0093 are removed. Besides, video_0003 and video_0028 are ignored as the annotated frame sequences are too short.

In our PSI paper experiments, the observed tracks length is 15, while predicting the 16-th frame intention. The overlap rate is set as 0.8 for both train and test stages.

Citing

@article{chen2021psi,
title   = {PSI: A Pedestrian Behavior Dataset for Socially Intelligent Autonomous Car},
author  = {Chen, Tina and Tian, Renran and Chen, Yaobin and Domeyer, Joshua and Toyoda, Heishiro and Sherony, Rini and Jing, Taotao and Ding, Zhengming},
journal = {arXiv preprint arXiv:2112.02604},
year    = {2021} }

About

Data preprocessing for IUPUI-CSRC Pedestrian Situated Intent (PSI) benchmark dataset.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%
0