You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is it possible to have some sort of interface that's not lm-format-enforcer wrapping vLLM, rather, users can directly install both library and configure it using SamplingParams?
the logits processor needs the tokenizer to prepare itself. To make an API like this possible, the logitsprocessor interface would have to be more than a Callable, but a class with some init(tokenizer) functionality. This is inconsistent with how other inference libraries open logits processing APIs, which is why I chose this approach with vLLM.
Its technically possible, but would require changes to vLLM's SamplingParams, and I actually think the current interface is the right one.
Is it possible to have some sort of interface that's not
lm-format-enforcer
wrappingvLLM
, rather, users can directly install both library and configure it usingSamplingParams
?The pseudocode in my mind:
Here's a similar proposal to Outlines: dottxt-ai/outlines#163 (comment)
The text was updated successfully, but these errors were encountered: