Skip to content

Commit 4c15f16

Browse files
authored
Update README.md
1 parent eff2102 commit 4c15f16

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ GPTQ is SOTA one-shot weight quantization method
77

88
## New Features
99
**Changed to use only pytorch instead of the current cuda kernel.
10-
It has no impact on memory usage. There is a slowdown below 128 length(If you use Transformers' use_cache, seq_len is effectively close to 1.), but much faster at 128 and above.**
10+
It has no impact on memory usage. There is a slowdown below 128 length(If you use Transformers' use_cache, length is effectively close to 1.), but much faster at 128 and above.**
1111

1212
Changed to support new features proposed by [GPTQ](https://github.com/IST-DASLab/gptq#new-features).
1313

0 commit comments

Comments
 (0)