You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have looked through all the examples we currently have for kinference, but nothing is doing tokenisation yet. You would sort of expect an example like POSTagger to be doing tokenisation, but it seems to skip the hard part and load the end result directly in as the input.(Unless I'm misreading the code?)
How do I go from a string prompt, into an ONNXData object that would be accepted by this model?
The text was updated successfully, but these errors were encountered:
You are correct, KInference expects you to do all the input data preprocessing yourself (e.g. tokenization), as it is an inference-only library. So in order to get ONNXData you have to implement your own tokenizer that converts input string to NDArray and then use .asTensor(name) on it.
Unfortunately, we don't have any plans to add built-in tokenization yet.
Tokenisation is something I can do, my biggest stumbling block is not knowing the structure of the data I have to provide. If I provide nothing then it throws an error saying that "input" is missing, so that's currently the best hint I have to work with. It would be super nice if it threw some detailed error about the shape of the data it expected to be fed in.
I have been studying the Python demo code for llama.onnx, found here:
https://github.com/tpoisonooo/llama.onnx/blob/main/demo_llama.py#L184
I have looked through all the examples we currently have for kinference, but nothing is doing tokenisation yet. You would sort of expect an example like
POSTagger
to be doing tokenisation, but it seems to skip the hard part and load the end result directly in as the input.(Unless I'm misreading the code?)How do I go from a string prompt, into an
ONNXData
object that would be accepted by this model?The text was updated successfully, but these errors were encountered: