You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
I am trying to run the llama2 demo. Through export.py, I get several .bin files of HTP. Can you provide the test input files of this model together so that I can run this model on my device using qnn-net-run. Test locally. I think there should be such files on the cloud device. Can I download them to the local computer? Thank you.
Describe the solution you'd like
export.py will export the compiled test input file
The text was updated successfully, but these errors were encountered:
Hi @shifeiwen that's a great suggestion.
We currently store user provided data (inference job's input dataset) and simply serialize it to numpy tensor to use along with qnn-net-run.
You can serialize input as follow
For each value in input data serialize as numpy file
Create input_list.txt file with key and local serialized data relative file path
to get an understanding on how to convert input_prompt into tensors for input of a first model.
By running each model, you can quickly create inputs for sub-sequent model parts.
Let us know if this unblocks you running these models via qnn-net-run or have any follow up questions
mestrona-3
added
the
question
Please ask any questions on Slack. This issue will be closed once responded to.
label
Aug 8, 2024
Is your feature request related to a problem? Please describe.
I am trying to run the llama2 demo. Through export.py, I get several .bin files of HTP. Can you provide the test input files of this model together so that I can run this model on my device using qnn-net-run. Test locally. I think there should be such files on the cloud device. Can I download them to the local computer? Thank you.
Describe the solution you'd like
export.py will export the compiled test input file
The text was updated successfully, but these errors were encountered: