You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
I'm attempting to eliminate most if not all CPU fallback code paths in my GGML backend. One of which is reshaping a single vector into a 3D tensor - I know this is not efficient on TT hardware and I should eliminate it completely by graph rewrite in the future - This operation fails when performed purely on device.
To Reproduce
The following is the minimal reproducible example. Compile and run it:
Metal | INFO | Initializing device 0. Program cache is NOT enabled
Metal | INFO | AI CLK for device 0 is: 1000 MHz
Metal | INFO | MMIO Device 0 : Tunnel 0 : Device 0
Metal | INFO | MMIO Device 0 : Tunnel 0 : Device 4
Original shape: ttnn.Shape([1, 1, 1[32], 65536])
Always | FATAL | Statically allocated circular buffers on core range [(x=0,y=0) - (x=0,y=0)] grow to 8487424 B which is beyond max L1 size of 1499136 B
terminate called after throwing an instance of 'std::runtime_error'
what(): TT_THROW @ /home/marty/Documents/tt/tt-metal/tt_metal/impl/program/program.cpp:761: tt::exception
info:
Statically allocated circular buffers on core range [(x=0,y=0) - (x=0,y=0)] grow to 8487424 B which is beyond max L1 size of 1499136 B
Expected behavior
Reshape from 1D vector should work no matter the vector size. Or at least the API/document should provide some way to indicate the algorithm cannot run if the input exceeds a certain size.
Screenshots
If applicable, add screenshots to help explain your problem.
Please complete the following environment information:
@dmakoviichuk-ttThe problem is not tilize_with_zero_padding. It's with ttnn::reshape. The line Original shape: ttnn.Shape([1, 1, 1[32], 65536]) is printed. I assume this means tilize_with_zero_padding passes?
You are right. The problem is indeed in tilize_with_zero_padding.
Update: No,with multicore=true i get the same error.
Describe the bug
I'm attempting to eliminate most if not all CPU fallback code paths in my GGML backend. One of which is reshaping a single vector into a 3D tensor - I know this is not efficient on TT hardware and I should eliminate it completely by graph rewrite in the future - This operation fails when performed purely on device.
To Reproduce
The following is the minimal reproducible example. Compile and run it:
Observe the error:
Expected behavior
Reshape from 1D vector should work no matter the vector size. Or at least the API/document should provide some way to indicate the algorithm cannot run if the input exceeds a certain size.
Screenshots
If applicable, add screenshots to help explain your problem.
Please complete the following environment information:
Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered: