GPU环境使用ocrv5的server模型,异常RUNTIME_EXCEPTION:Failed to allocate memory for requested buffer of size 8797618176 #549
Unanswered
loneWolf1127
asked this question in
Q&A
Replies: 3 comments 2 replies
-
|
GPU环境下不建议使用onnxruntime作为推理引擎。因此我这里并没有测过。 |
Beta Was this translation helpful? Give feedback.
1 reply
-
|
paddle或者torch
…---- 回复的原邮件 ----
| 发件人 | ***@***.***> |
| 发送日期 | 2025年09月12日 11:49 |
| 收件人 | RapidAI/RapidOCR ***@***.***> |
| 抄送人 | SWHL ***@***.***>,
Comment ***@***.***> |
| 主题 | Re: [RapidAI/RapidOCR] GPU环境使用ocrv5的server模型,异常RUNTIME_EXCEPTION:Failed to allocate memory for requested buffer of size 8797618176 (Discussion #549) |
GPU环境下建议使用哪个推理引擎?cpu下没问题
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you commented.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
0 replies
-
|
I have similar issue, but I cannot change to different inference engine as using it under immich docker app, I noticed that when forced to run the rapidocr inference sequentially, avoiding any form of concurrency, the memory allocation issue is happening less. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
问题描述 / Problem Description
GPU环境使用ocrv5的server模型,异常RUNTIME_EXCEPTION:
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/app/llmdoc/code/gpu_service/onnx_test.py", line 34, in
result = ocr_engine(img_url)
^^^^^^^^^^^^^^^^^^^
File "/opt/conda/lib/python3.11/site-packages/rapidocr/main.py", line 116, in call
img, det_res = self.get_det_res(img, op_record)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/lib/python3.11/site-packages/rapidocr/main.py", line 222, in get_det_res
det_res = self.text_det(img)
^^^^^^^^^^^^^^^^^^
File "/opt/conda/lib/python3.11/site-packages/rapidocr/ch_ppocr_det/main.py", line 59, in call
preds = self.session(prepro_img)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/lib/python3.11/site-packages/rapidocr/inference_engine/onnxruntime/main.py", line 93, in call
raise ONNXRuntimeError(error_info) from e
rapidocr.inference_engine.onnxruntime.main.ONNXRuntimeError: Traceback (most recent call last):
File "/opt/conda/lib/python3.11/site-packages/rapidocr/inference_engine/onnxruntime/main.py", line 90, in call
return self.session.run(self.get_output_names(), input_dict)[0]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/conda/lib/python3.11/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 270, in run
return self._sess.run(output_names, input_feed, run_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
onnxruntime.capi.onnxruntime_pybind11_state.RuntimeException: [ONNXRuntimeError] : 6 : RUNTIME_EXCEPTION : Non-zero status code returned while running Conv node. Name:'Conv.86' Status Message: /onnxruntime_src/onnxruntime/core/framework/bfc_arena.cc:376 void* onnxruntime::BFCArena::AllocateRawInternal(size_t, bool, onnxruntime::Stream*, bool, onnxruntime::WaitNotificationFn) Failed to allocate memory for requested buffer of size 8797618176
运行环境 / Runtime Environment
Docker version 26.1.0
cuda 12.4
复现代码 / Reproduction Code
可能解决方案 / Possible solutions
Beta Was this translation helpful? Give feedback.
All reactions