lowering to Ethos-U85 with input of HW1 #18047
Unanswered
shmulik-akerman
asked this question in
Q&A
Replies: 1 comment
-
|
hey shmulik, with Ethos-U85, the memory blowup during transpose is a known issue. you might want to look into optimizing the model with TensorFlow Lite's model optimization toolkit or try reducing the input size, if possible. not entirely sure about other workarounds for this specific case tbh. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello all,
I'm trying to lower a tiny model so it can run in a very memory-tight environment. input image is [300,400,1] of size. when looking in the tensor allocation table vela reports, I see the following lines right the the table top:
0 - 1: 0x1d4c00 - 0x1f20c0: 120000: 2040512: NHWC : quantized_decomposed_quantize_per_tensor_default
0 - 3: 0x0 - 0x1d4c00: 1920000: 2400944: NHCWB16 : tosa_transpose_default
as 300x400x16 = 1920000, I'm obviously not handling the NCHW->NHWC correctly, and the explicit transpose creates that huge (in my world of scarce memory) tensor.
the peak memory for that model, if not that ill behavior, is lower than 600K.
I'll be very grateful if someone who already solved this will share.
Shmulik
Beta Was this translation helpful? Give feedback.
All reactions