Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question in decoding performance. #15

Open
jbkoh opened this issue Oct 2, 2017 · 0 comments
Open

Question in decoding performance. #15

jbkoh opened this issue Oct 2, 2017 · 0 comments

Comments

@jbkoh
Copy link

jbkoh commented Oct 2, 2017

Hi there,

Thanks for sharing a great project. Not sure how much this repository is active, but I have a question.

I am running precompiled version of CRFSharp v1.2.0.0 on Ubuntu 16.04 with Mono. 5.2.0.224.

I love the performance increase in encoding (or learning), but I see that CPU resources are underexploited when I decode files. I run it with 30 cores and usage rate of most of the cores are below 10 % for half of the time and 30-50 % for the rest of the time. (In encoding stage, I see 100% usages all over the cores.) I would like to know if this is inherent in CRFSharp's algorithm or this is my configuration issue.

One of the observations I had was that there is not much difference in delays of decoding 10 samples to 1000 samples like from 4 mins for 5 sentences (each sentence consists of 150 tags), 5 mins for 100 sentences, and 10 mins for 1000 sentences. It looks like there is a large delay in loading the model. I currently use nbest=10, but I previously saw that there is not much difference in nbest=1 and nbest=10.

Let me know if there is anything that I can learn for this issue.

Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant