You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Update base for Update on "[executorch] Propagate device metadata from partitioner result onto TensorSpecs"
Add end-to-end device type annotation support from export to runtime. Currently we only support one device per graph
The overall pipeline is:
a. Partitioner use `compile_spec` to determine which device the partitoned blob is runing on
b. after lowered partitioned graph to backend, the new-introed propagate_device_pass will annotate the input and output tensors of delegate blob as target device.
Differential Revision: [D95842511](https://our.internmc.facebook.com/intern/diff/D95842511/)
[ghstack-poisoned]
RUNNER_ARGS="$RUNNER_ARGS --tokenizer_path ${MODEL_DIR}/$TOKENIZER_FILE --prompt 'What is the capital of France?' --max_new_tokens 128 --temperature 0"
0 commit comments