Forum Discussion
Hi Ruben,
Unfortunately I do not have the setup to this up and I am working with engineering to see how we can implement the flow that you requested.
I am sorry if I am not able expediate this support and I am trying my best to help you resolved the issue.
Hello @JohnT_Intel,
Sorry. As you suggested reusing the inference request instead of creating a new one for each inference, I thought the solution was trivial and that the problem was in my implementation or concept.
I look forward to a solution.
I believe the concept of using the DLA is correct in a real application: deploy the accelerator and configure it with the graph, then keep it configured and continuously feed it with new data for inference. Isn't that right? Of course, new inferences must wait for the previous one to finish. Is this correct, or have I misunderstood something about the working principle of the accelerator?