Skip to content

Instantly share code, notes, and snippets.

@AmosLewis
Created June 27, 2025 03:34
Show Gist options
  • Save AmosLewis/c8ff7300d4bcf93e665a1f7d635a0d4d to your computer and use it in GitHub Desktop.
Save AmosLewis/c8ff7300d4bcf93e665a1f7d635a0d4d to your computer and use it in GitHub Desktop.
python3 -m sharktank.examples.sharding.shard_llm_dataset \
--irpa-file native_fp8_e4m3fnuz_llama3_405b.irpa \
--output-irpa native_fp8_e4m3fnuz_llama3_405b_tp8.irpa \
--tensor-parallelism-size 8
/home/chi/src/shark-ai/.venv/lib/python3.12/site-packages/iree/turbine/aot/params.py:163: UserWarning: The given NumPy array is not writable, and PyTorch does not support non-writable tensors. This means writing to this tensor will result in undefined behavior. You may want to copy the array to protect its data or make it writable before converting it to a tensor. This type of warning will be suppressed for the rest of this program. (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:206.)
return torch.from_numpy(wrapper)
Traceback (most recent call last):
File "<frozen runpy>", line 198, in _run_module_as_main
File "<frozen runpy>", line 88, in _run_code
File "/home/chi/src/shark-ai/sharktank/sharktank/examples/sharding/shard_llm_dataset.py", line 51, in <module>
main()
File "/home/chi/src/shark-ai/sharktank/sharktank/examples/sharding/shard_llm_dataset.py", line 43, in main
sharded_theta = shard_theta(dataset.root_theta, llama_config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/types/sharding.py", line 473, in shard_theta
return ops.reshard(
^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1178, in _reshard_trampoline
result = override(input, spec)
^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 92, in func_wrapper
res = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1169, in reshard_theta_layer_sharding
return reshard(input, spec.theta_sharding())
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1178, in _reshard_trampoline
result = override(input, spec)
^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 92, in func_wrapper
res = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1158, in reshard_theta_sharding
k: make_value(input(k), spec[k])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1146, in make_value
result = reshard(input, spec)
^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1178, in _reshard_trampoline
result = override(input, spec)
^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 92, in func_wrapper
res = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1158, in reshard_theta_sharding
k: make_value(input(k), spec[k])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1146, in make_value
result = reshard(input, spec)
^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1178, in _reshard_trampoline
result = override(input, spec)
^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 92, in func_wrapper
res = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1158, in reshard_theta_sharding
k: make_value(input(k), spec[k])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1146, in make_value
result = reshard(input, spec)
^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1178, in _reshard_trampoline
result = override(input, spec)
^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 92, in func_wrapper
res = f(*args, **kwargs)
^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1158, in reshard_theta_sharding
k: make_value(input(k), spec[k])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/sharded_impls.py", line 1146, in make_value
result = reshard(input, spec)
^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 222, in __call__
selected_override, *results = trampoline(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/signatures.py", line 1182, in _reshard_trampoline
d.fail(dispatch_args)
File "/home/chi/src/shark-ai/sharktank/sharktank/ops/_registry.py", line 271, in fail
raise NotImplementedError(
NotImplementedError: Overridable operator sharktank.ops.signatures.reshard does not have an implementation for argument types: [<class 'sharktank.types.tensors.PlanarQuantizedTensor'>, <class 'sharktank.types.sharding.Split'>]
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment