TorchAO has [implemented AWQ](https://github.com/pytorch/ao/tree/main/torchao/prototype/awq), and we'd like to extend the implementation to cover dynamic activation quantization to support lowering to ExecuTorch. 1. We should modify the existing AWQ algorithm to support QDQLayout [here](https://github.com/pytorch/ao/blob/main/torchao/prototype/awq/core.py#L118C22-L118C33) to support ExecuTorch 2. The scales should have 8-bit dynamic activation quantization applied before computing the AWQ scaling. With the above changes, we should be able to quantize a model with AWQ and lower it to ExecuTorch following instructions similar to [here](https://huggingface.co/pytorch/Qwen3-4B-8da4w).