Reference Action Segmentation Evaluation Code
This repository contains the reference code for action segmentation evaluation.
If you have a bug-fix/improvement or if you want to add a new features please send a pull request or open an issue.
Example Usage
All the metrics have the same api.
from actseg.eval import MoFAccuracy, Edit
pred1 = [0, 0, 0, 1, 0, 1, 1, 1, 0]
pred2 = [1, 2, 3, 0, 0, 1, 2, 3, 0, 0, 0, 1, 2, 3, 0, 0, 0, 0]
target1 = [0, 0, 1, 1, 2, 1, 1, 0, 0]
target2 = [1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 3, 3, 3, 3, 3, 3, 3]
metrics = [MoFAccuracy(), Edit()]
for p, t in zip([pred1, pred2], [target1, target2]):
for m in metrics:
m(targets=t, predictions=p)
for m in metrics:
print(m)
# MoF: 0.3333333333333333
# Edit: 52.5
Metrics
Frame-wise Metrics
- MoF (Accuracy)
- F1Score
- IoD
- IoU
Segment-wise Metrics
- Edit (Edit distance or matching score)
Specifying Ignore Class
For some Metrics it is possible to specify the indices of classes to ignore (e.g. Background) by passing ignore_ids
parameter to the constructor.
Authors
Yasser Souri: @yassersouri
Zijia Lu: @ZijiaLewisLu
Acknowledgement
Please see actseg/external
for external sources used in this project.