-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
APT results #11
base: main
Are you sure you want to change the base?
APT results #11
Conversation
Hi @mkabra, thanks for flagging! I will look into the proposed updates for the docs, and will confirm here again if your submission is working. |
Hi @mkabra , thanks for adding the most recent comments. Looking into this in the next days and will try to get back by early next week. Thanks again for the contribution! |
bump @stes |
@mkabra sorry for the really slow response time - there are still a few issues in your code that I've fixed on my end, and I'll push those changes. Meanwhile, do you have scores available for your predictions? I've noticed our docs were incorrect and mentioned that the results should be given in the format return {
"path/to/image.png" : (
# animal 1
{
"snout" : (0, 1),
"leftear" : (2, 3),
...
},
# animal 2
{
"snout" : (0, 1),
"leftear" : (2, 3),
...
},
),
...
} when they should be given in the format return {
"path/to/image.png" : (
# animal 1
{
"pose": {
"snout" : (12, 17),
"leftear" : (15, 13),
...
},
"score": 0.9172,
},
...
),
...
} To compute evaluation metrics, the model confidence is very important. I've been able to evaluate your model with random scores, but to get the true performance I would also need the score for each individual. |
@mkabra I just pushed the updated code (with 413b8e5) which changes the The scores for each prediction is generated with the first prediction for each being given the highest score, the 2nd the 2nd highest, etc. (updating the json files to individual scores will mean this is no longer needed - and produce the correct evaluation results). |
Hi,
We have benchmarked three multi-animal networks from APT on Deeplabcut Datasets. We tested our submission by running "python -m benchmark". We get the following output which we believe suggests that the submission is working:
The documentation for creating the submission though is slightly out of date. We had to make the following changes to get the test to work
__init__.py
to benchmark/submissionsinstead of
Hope everything else is as expected.
Best,
Mayank