File tree Expand file tree Collapse file tree 1 file changed +0
-55
lines changed
tests/llmcompressor/modifiers/awq Expand file tree Collapse file tree 1 file changed +0
-55
lines changed Original file line number Diff line number Diff line change @@ -119,61 +119,6 @@ def test_set_resolved_mappings():
119119
120120@pytest .mark .unit
121121def test_validate ():
122- AWQModifier (scheme = "W8A8" )
123-
124- AWQModifier (
125- config_groups = {
126- "group_0" : QuantizationScheme (
127- targets = ["Linear" ],
128- weights = QuantizationArgs (
129- num_bits = 4 ,
130- group_size = 64 ,
131- ),
132- ),
133- "group_1" : QuantizationScheme (
134- targets = ["Linear" ],
135- weights = QuantizationArgs (
136- num_bits = 4 ,
137- group_size = 128 ,
138- ),
139- ),
140- }
141- )
142-
143- with pytest .raises (ValidationError ):
144- AWQModifier (
145- config_groups = {
146- "group_0" : QuantizationScheme (
147- targets = ["Linear" ],
148- weights = QuantizationArgs (
149- num_bits = 4 ,
150- group_size = 128 ,
151- ),
152- ),
153- "group_1" : QuantizationScheme (
154- targets = ["Linear" ],
155- weights = QuantizationArgs (
156- num_bits = 8 ,
157- group_size = 128 ,
158- ),
159- ),
160- }
161- )
162-
163- # valid configuration
164- AWQModifier (
165- config_groups = {
166- "group_0" : QuantizationScheme (
167- targets = ["Linear" ],
168- weights = QuantizationArgs (num_bits = 4 , group_size = 128 , symmetric = False ),
169- ),
170- "group_1" : QuantizationScheme (
171- targets = ["Linear" ],
172- weights = QuantizationArgs (num_bits = 4 , group_size = 128 , symmetric = False ),
173- ),
174- }
175- )
176-
177122 AWQModifier (scheme = "W4A16" , duo_scaling = "both" )
178123 with pytest .raises (ValidationError ):
179124 AWQModifier (scheme = "W4A16" , duo_scaling = "Both" )
You can’t perform that action at this time.
0 commit comments