-
Notifications
You must be signed in to change notification settings - Fork 106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Blackhole] Llama3-8B end-to-end testing #18135
Comments
For the MLP sharding, there's a good proxy unit test that's currently failing on BH:
|
Initial testing shows ops issue in Matrix Multiply with DRAM sharding. @mtairum to open issue and assign @bbradelTT. |
The error is in i2s, not matmul. I re-assigned the issue to @llongTT |
Describe the bug
Now that all llama3 ops are supported in BH (see #16013), test the full Llama3 model on Blackhole.
Mainly testing Llama3-8B but also running 1B and 3B since they should just work as well.
Plan: Make sure the individual modules work before running the full model.
Reproduce tests
The code below should download the HF weights if not existent, and run the test.
Issues
Current Llama3 sharding specs do not work on BH.
Modules status
test_llama_mlp.py -k 32-1
Bounds-Error -- Logical_core=(x=8,y=0) is outside of logical_grid_size=(x=8,y=1)
tt-metal/ttnn/cpp/ttnn/operations/data_movement/sharded/interleaved_to_sharded/device/interleaved_to_sharded_op.cpp:21: this->output_mem_config.buffer_type == BufferType::L1
interleaved_to_sharded_op.cpp:21: this->output_mem_config.buffer_type == BufferType::L1
test_llama_rms_norm.py
test_llama_decoder.py
test_lm_head.py
Bounds-Error -- Logical_core=(x=8,y=0) is outside of logical_grid_size=(x=8,y=1)
The text was updated successfully, but these errors were encountered: