Skip to content

[Tracker] Outstanding Issues and WIP Features for version 0.1 #20

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
13 of 26 tasks
HDCharles opened this issue Dec 5, 2023 · 2 comments
Closed
13 of 26 tasks

[Tracker] Outstanding Issues and WIP Features for version 0.1 #20

HDCharles opened this issue Dec 5, 2023 · 2 comments

Comments

@HDCharles
Copy link
Contributor

HDCharles commented Dec 5, 2023

This issue tracks outstanding issues for a torchao 0.1 release

  • New Functionality

    • Test compatibility with PyTorch 2.2 and 2.3rc1 (@cpuhrsch)
    • Fix tests marked as flaky (@cpuhrsch)
    • int4, int8 weight only quantization support (only need one of the paths to work)
      • path 1: int4, int8 weight quantization subclass API works with TorchTune (@jerryzh168), blocked by tensor subclass save load
      • path 2: int4, int8 weight quantization module swap API works with TorchTune (@jerryzh168), WIP
    • Add GPTQuantizer workflow for 4-bit weight quantization (W4A16) for GPU that works for gpt-fast (and executorch) (@jerryzh168, @HDCharles)
    • Add workflow for 4-bit weight, 8-bit activation quantization (W4A8) with/without GPTQ for executorch (@jerryzh168)
      • without GPTQ path is working, still verifying the GPTQ path
    • NF4 Dtype that works for QLoRA in TorchTune (@cpuhrsch)
    • Fix API so it works with LoRACompatibleLinear
    • Allow apply_quant_api()
      • it currently looks for the children of the module and so doesn't do anything
  • Tutorials/BE

    • Using/Writing a quantization technique using torchao (@jerryzh168)
    • Using kernels written in torchao with PyTorch
    • Replace Int8WeightOnlyQuantizedLinearWeight and Int8DynamicallyQuantizedLinearWeight with a single class
    • Reconsider using class method for Int8DynamicallyQuantizedLinearWeight.from_float
    • Remove / guard catch all forward args, kwargs for module swap API
    • Land Tutorial Adding tutorial for gpu quantization using torchao tutorials#2730
  • If time permits (or v0.2)

    • Enable test_8da4w_quantize for 2.4 @jerryzh168
    • 4-bit quantization CPU perf numbers
    • Feature parity between module swap api and subclass api
    • Align smoothquant api with others
      • Add high level auto quant API for int8 dynamic and weight-only quantization with benchmarks (@HDCharles)
@HDCharles HDCharles changed the title [Tracker] [Tracker] Outstanding Issues and WIP Features Dec 5, 2023
@HDCharles HDCharles changed the title [Tracker] Outstanding Issues and WIP Features [Tracker] Outstanding Issues and WIP Features for version 0.1 Dec 5, 2023
@cpuhrsch
Copy link
Contributor

We still require users to set a private API to get good performance: torch._inductor.config.force_fuse_int_mm_with_mul = True . When do we think we can solve that by? cc @eellison

@supriyar
Copy link
Contributor

tracking spillover and new features for 0.2 in #132

jerryzh168 pushed a commit that referenced this issue Sep 18, 2024
* Bring `torch.compile` to `quant_block_v2_`. (#18)

Signed-off-by: yiliu30 <[email protected]>

* Add `AO_USE_DETERMINISTIC_ALGORITHMS` for reproducing results (#19)

Signed-off-by: yiliu30 <[email protected]>

* Add `gradient_accumulate_steps` and update results (#20)

Signed-off-by: yiliu30 <[email protected]>

* update the readme

Signed-off-by: yiliu30 <[email protected]>

* udpate

Signed-off-by: yiliu30 <[email protected]>

* update the desc

Signed-off-by: yiliu30 <[email protected]>

* rename `train_bs` to `batch_size`

Signed-off-by: yiliu30 <[email protected]>

* update the eval

Signed-off-by: yiliu30 <[email protected]>

* update

Signed-off-by: yiliu30 <[email protected]>

---------

Signed-off-by: yiliu30 <[email protected]>
jainapurva pushed a commit that referenced this issue Sep 22, 2024
* Bring `torch.compile` to `quant_block_v2_`. (#18)

Signed-off-by: yiliu30 <[email protected]>

* Add `AO_USE_DETERMINISTIC_ALGORITHMS` for reproducing results (#19)

Signed-off-by: yiliu30 <[email protected]>

* Add `gradient_accumulate_steps` and update results (#20)

Signed-off-by: yiliu30 <[email protected]>

* update the readme

Signed-off-by: yiliu30 <[email protected]>

* udpate

Signed-off-by: yiliu30 <[email protected]>

* update the desc

Signed-off-by: yiliu30 <[email protected]>

* rename `train_bs` to `batch_size`

Signed-off-by: yiliu30 <[email protected]>

* update the eval

Signed-off-by: yiliu30 <[email protected]>

* update

Signed-off-by: yiliu30 <[email protected]>

---------

Signed-off-by: yiliu30 <[email protected]>
jainapurva pushed a commit that referenced this issue Sep 23, 2024
* Bring `torch.compile` to `quant_block_v2_`. (#18)

Signed-off-by: yiliu30 <[email protected]>

* Add `AO_USE_DETERMINISTIC_ALGORITHMS` for reproducing results (#19)

Signed-off-by: yiliu30 <[email protected]>

* Add `gradient_accumulate_steps` and update results (#20)

Signed-off-by: yiliu30 <[email protected]>

* update the readme

Signed-off-by: yiliu30 <[email protected]>

* udpate

Signed-off-by: yiliu30 <[email protected]>

* update the desc

Signed-off-by: yiliu30 <[email protected]>

* rename `train_bs` to `batch_size`

Signed-off-by: yiliu30 <[email protected]>

* update the eval

Signed-off-by: yiliu30 <[email protected]>

* update

Signed-off-by: yiliu30 <[email protected]>

---------

Signed-off-by: yiliu30 <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants