fix inf,-inf with 2 concurrency

This commit is contained in:
yoni13
2025-01-10 14:04:18 +00:00
parent 9bc3e5b2e2
commit 4d704e9f73
2 changed files with 74 additions and 27 deletions
+21 -27
View File
@@ -6,7 +6,7 @@ from typing import Any, List
import numpy as np
import onnxruntime as ort
from numpy.typing import NDArray
from rknnlite.api import RKNNLite
from rknn.rknnpool import rknnPoolExecutor
from app.models.constants import SUPPORTED_PROVIDERS
@@ -14,38 +14,40 @@ from app.schemas import SessionNode
from ..config import log, settings
def runInfrence(rknn_lite, input):
outputs = rknn_lite.inference(inputs=[input], data_format='nchw')
return outputs
class RknnSession:
def __init__(self, model_path: Path | str):
self.model_path = Path(model_path)
self.ort_model_path = str(self.model_path).replace(".rknn", ".onnx")
log.info(f"Loading RKNN model from {self.model_path} with {1 if 'textual' in str(self.model_path) else 2} threads.")
self.rknnpool = rknnPoolExecutor(
rknnModel=self.model_path.as_posix(),
TPEs= 1 if 'textual' in str(self.model_path) else 2,
func=runInfrence)
self.rknn = RKNNLite()
log.info(f"Loading RKNN model from {self.model_path}")
self.ort_session = ort.InferenceSession(
self.ort_model_path,
)
self.inputs = self.ort_session.get_inputs()
self.outputs = self.ort_session.get_outputs()
# ret = self.rknn.load_onnx(self.model_path)
print('--> Load RKNN model')
ret = self.rknn.load_rknn(self.model_path.as_posix())
if ret != 0:
raise RuntimeError("Failed to load RKNN model")
del self.ort_session
ret = self.rknn.init_runtime()
if ret != 0:
raise RuntimeError("Failed to initialize RKNN runtime")
def __del__(self):
self.rknnpool.release()
def get_inputs(self) -> list[SessionNode]:
inputs: list[SessionNode] = self.ort_session.get_inputs()
return inputs
return self.inputs
def get_outputs(self) -> list[SessionNode]:
outputs: list[SessionNode] = self.ort_session.get_outputs()
return outputs
return self.outputs
def run(
self,
@@ -55,15 +57,7 @@ class RknnSession:
):
input_data = [np.ascontiguousarray(v) for v in input_feed.values()][0]
# log.info(f"Running inference on RKNN model")
outputs = self.rknn.inference(inputs=[input_data], data_format='nchw')
# log.info("inputs:")
# log.info(input_data)
# log.info("outputs:")
# log.info(outputs)
# log.info("RKNN END")
self.rknnpool.put(input_data)
outputs = self.rknnpool.get()[0]
return outputs