15
15
import unittest
16
16
17
17
import numpy as np
18
- from test_weight_only_linear import convert_uint16_to_float , get_cuda_version
18
+ from test_weight_only_linear import convert_uint16_to_float
19
19
20
20
import paddle
21
21
import paddle .nn .quant as Q
22
22
from paddle import base
23
- from paddle .base import core
24
23
from paddle .base .framework import default_main_program
25
24
from paddle .framework import set_default_dtype
26
25
from paddle .pir_utils import test_with_pir_api
30
29
default_main_program ().random_seed = 42
31
30
32
31
33
- @unittest .skipIf (
34
- not core .is_compiled_with_cuda ()
35
- or get_cuda_version () < 11020
36
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
37
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
38
- )
32
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
39
33
class LLMInt8LinearTestCase (unittest .TestCase ):
40
34
def config (self ):
41
35
self .dtype = 'float16'
@@ -149,25 +143,15 @@ def test_llm_int8_linear(self):
149
143
)
150
144
151
145
152
- @unittest .skipIf (
153
- not core .is_compiled_with_cuda ()
154
- or get_cuda_version () < 11020
155
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
156
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
157
- )
146
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
158
147
class LLMInt8LinearTestCase1 (LLMInt8LinearTestCase ):
159
148
def config (self ):
160
149
super ().config ()
161
150
self .dtype = 'float16'
162
151
self .weight_dtype = "int8"
163
152
164
153
165
- @unittest .skipIf (
166
- not core .is_compiled_with_cuda ()
167
- or get_cuda_version () < 11020
168
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
169
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
170
- )
154
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
171
155
class LLMInt8LinearTestCase2 (LLMInt8LinearTestCase ):
172
156
def config (self ):
173
157
super ().config ()
@@ -176,39 +160,23 @@ def config(self):
176
160
self .weight_dtype = "int8"
177
161
178
162
179
- @unittest .skipIf (
180
- not core .is_compiled_with_cuda ()
181
- or get_cuda_version () < 11020
182
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
183
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
184
- )
163
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
185
164
class LLMInt8LinearTestCase3 (LLMInt8LinearTestCase ):
186
165
def config (self ):
187
166
super ().config ()
188
167
self .dtype = 'bfloat16'
189
168
self .weight_dtype = "int8"
190
169
191
170
192
- @unittest .skipIf (
193
- not core .is_compiled_with_cuda ()
194
- or get_cuda_version () < 11020
195
- or paddle .device .cuda .get_device_capability ()[0 ] < 8
196
- or not core .is_bfloat16_supported (core .CUDAPlace (0 )),
197
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8 or core is not support bfloat16" ,
198
- )
171
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
199
172
class LLMInt8LinearTestCase4 (LLMInt8LinearTestCase ):
200
173
def config (self ):
201
174
super ().config ()
202
175
self .dtype = 'float16'
203
176
self .weight_dtype = "int4"
204
177
205
178
206
- @unittest .skipIf (
207
- not core .is_compiled_with_cuda ()
208
- or get_cuda_version () < 11020
209
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
210
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
211
- )
179
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
212
180
class LLMInt8LinearTestCase5 (LLMInt8LinearTestCase ):
213
181
def config (self ):
214
182
super ().config ()
@@ -217,26 +185,15 @@ def config(self):
217
185
self .weight_dtype = "int4"
218
186
219
187
220
- @unittest .skipIf (
221
- not core .is_compiled_with_cuda ()
222
- or get_cuda_version () < 11020
223
- or paddle .device .cuda .get_device_capability ()[0 ] < 8
224
- or not core .is_bfloat16_supported (core .CUDAPlace (0 )),
225
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8 or core is not support bfloat16" ,
226
- )
188
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
227
189
class LLMInt8LinearTestCase6 (LLMInt8LinearTestCase ):
228
190
def config (self ):
229
191
super ().config ()
230
192
self .dtype = 'bfloat16'
231
193
self .weight_dtype = "int4"
232
194
233
195
234
- @unittest .skipIf (
235
- not core .is_compiled_with_cuda ()
236
- or get_cuda_version () < 11020
237
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
238
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
239
- )
196
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
240
197
class LLMInt8LinearTestCase7 (LLMInt8LinearTestCase ):
241
198
def config (self ):
242
199
super ().config ()
@@ -246,12 +203,7 @@ def config(self):
246
203
self .token = 1
247
204
248
205
249
- @unittest .skipIf (
250
- not core .is_compiled_with_cuda ()
251
- or get_cuda_version () < 11020
252
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
253
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
254
- )
206
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
255
207
class LLMInt8LinearTestCase8 (LLMInt8LinearTestCase ):
256
208
def config (self ):
257
209
super ().config ()
@@ -262,12 +214,7 @@ def config(self):
262
214
self .token = 1
263
215
264
216
265
- @unittest .skipIf (
266
- not core .is_compiled_with_cuda ()
267
- or get_cuda_version () < 11020
268
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
269
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
270
- )
217
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
271
218
class LLMInt8LinearTestCase9 (LLMInt8LinearTestCase ):
272
219
def config (self ):
273
220
super ().config ()
@@ -277,12 +224,7 @@ def config(self):
277
224
self .token = 1
278
225
279
226
280
- @unittest .skipIf (
281
- not core .is_compiled_with_cuda ()
282
- or get_cuda_version () < 11020
283
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
284
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
285
- )
227
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
286
228
class LLMInt8LinearTestCase10 (LLMInt8LinearTestCase ):
287
229
def config (self ):
288
230
super ().config ()
@@ -293,13 +235,7 @@ def config(self):
293
235
self .token = 1
294
236
295
237
296
- @unittest .skipIf (
297
- not core .is_compiled_with_cuda ()
298
- or not core .is_compiled_with_cuda ()
299
- or get_cuda_version () < 11020
300
- or paddle .device .cuda .get_device_capability ()[0 ] < 8 ,
301
- "quantized_matmul requires CUDA >= 11.2 and CUDA_ARCH >= 8" ,
302
- )
238
+ @unittest .skipIf (True , "Disable this unit test in release/2.6" )
303
239
class LLMInt8LinearTestCaseStatic (LLMInt8LinearTestCase ):
304
240
def config (self ):
305
241
super ().config ()
0 commit comments