Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Replicating implementation doubts. Wilderness Impact metrics, training protocol... #18

Open
aitor-martinez-seras opened this issue Mar 22, 2024 · 0 comments

Comments

@aitor-martinez-seras
Copy link

aitor-martinez-seras commented Mar 22, 2024

Hello. First of all, thank you for your work and for sharing it!

I am implementing the evaluation process of OWOD (defined in PascalVOCDetectionEvaluator.evaluate). I have encountered several problems to correctly apply the benchmark to my own work. Here are them:

  • For evaluating every task and every metric (even WI), I assume you are always performing evaluation against all_task_test.txt. Is that right?
  • For the Wilderness Impact, they get reported the WI values for different recalls. Which one do you chose?
  • Are instances categorized as "difficult" used in for the metrics?
  • The reported number of test instances of T1 in the below image of your paper refer only to the KNOWN classes or are accounting for both KNOWN and UNKNOWN objects?
    image
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant