Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

QAT, latency stays the same #230

Open
maiiabocharova opened this issue May 2, 2022 · 1 comment
Open

QAT, latency stays the same #230

maiiabocharova opened this issue May 2, 2022 · 1 comment

Comments

@maiiabocharova
Copy link

Trained a model according to documentation.
Added those configs lines for QAT

    cfg.QUANTIZATION.QAT.BATCH_SIZE_FACTOR = 1.0
    cfg.QUANTIZATION.BACKEND = "fbgemm"
    cfg.QUANTIZATION.QAT.FAKE_QUANT_METHOD = "default"
    cfg.QUANTIZATION.QAT.START_ITER = 1200
    cfg.QUANTIZATION.QAT.ENABLE_OBSERVER_ITER = 1200
    cfg.QUANTIZATION.QAT.ENABLE_LEARNABLE_OBSERVER_ITER = 1300
    cfg.QUANTIZATION.QAT.DISABLE_OBSERVER_ITER = 1200 + 300
    cfg.QUANTIZATION.QAT.FREEZE_BN_ITER = 12000 + 200

Trained for 1500 iterations.

The model is Ok and predictions are correct. But the model size and time of execution stayed exactly the same. Can you please help and advice on how to fix it?

@wat3rBro
Copy link
Contributor

wat3rBro commented Jun 3, 2022

Are you running the exported torchscript moddel? the speed up and model size reduction is for running torchscript model on device.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants