forked from OSchip/llvm-project
234 lines
10 KiB
MLIR
234 lines
10 KiB
MLIR
// RUN: mlir-opt %s -split-input-file -quant-convert-simulated-quantization | FileCheck %s
|
|
|
|
// -----
|
|
// Verifies a quint8 single point.
|
|
// CHECK-LABEL: fakeQuantArgs_Quint8_0
|
|
func @fakeQuantArgs_Quint8_0(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 1.000000e+00>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<u8:f32, 1.000000e+00>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 0.0 : f32, num_bits = 8
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a quint8 single point (with narrow_range = true).
|
|
// CHECK-LABEL: fakeQuantArgs_Quint8_0_NarrowRange
|
|
func @fakeQuantArgs_Quint8_0_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 1.000000e+00:1>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 1.000000e+00:1>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, narrow_range = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a quint8 asymmetric 0..1 range.
|
|
// CHECK-LABEL: fakeQuantArgs_Quint8_0_1
|
|
func @fakeQuantArgs_Quint8_0_1(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 0.0039215686274509803>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8:f32, 0.0039215686274509803>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 1.0 : f32, num_bits = 8
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a quint8 asymmetric 0..1 range (with narrow_range = true).
|
|
// CHECK-LABEL: fakeQuantArgs_Quint8_NarrowRange
|
|
func @fakeQuantArgs_Quint8_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 0.003937007874015748:1>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 0.003937007874015748:1>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, narrow_range = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a quint8 symmetric range of -1..127/128.
|
|
// CHECK-LABEL: fakeQuantArgs_Quint8_SymmetricRange
|
|
func @fakeQuantArgs_Quint8_SymmetricRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 7.812500e-03:128>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8:f32, 7.812500e-03:128>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = -1.0 : f32, max = 0.9921875 : f32, num_bits = 8, narrow_range = false
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 single point.
|
|
// CHECK-LABEL: fakeQuantArgs_Qint8_0
|
|
func @fakeQuantArgs_Qint8_0(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 1.000000e+00:-128>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 1.000000e+00:-128>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 single point (with narrow_range = true).
|
|
// CHECK-LABEL: fakeQuantArgs_Qint8_0_NarrowRange
|
|
func @fakeQuantArgs_Qint8_0_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 1.000000e+00:-127>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 1.000000e+00:-127>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, narrow_range = true, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 asymmetric 0..1 range.
|
|
// CHECK-LABEL: fakeQuantArgs_Qint8_0_1
|
|
func @fakeQuantArgs_Qint8_0_1(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 asymmetric 0..1 range (with narrow_range = true).
|
|
// CHECK-LABEL: fakeQuantArgs_Qint8_NarrowRange
|
|
func @fakeQuantArgs_Qint8_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 0.003937007874015748:-127>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 0.003937007874015748:-127>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, narrow_range = true, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 symmetric range of -1..127/128.
|
|
// CHECK-LABEL: fakeQuantArgs_Qint8_SymmetricRange
|
|
func @fakeQuantArgs_Qint8_SymmetricRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 7.812500e-03>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8:f32, 7.812500e-03>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = -1.0 : f32, max = 0.9921875 : f32, num_bits = 8, narrow_range = false, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a commonly used -1..1 symmetric 16bit range with a zero point of
|
|
// 0 and range -1.0 .. 32767/32768.
|
|
// CHECK-LABEL: fakeQuantArgs_Qint16_Symmetric
|
|
func @fakeQuantArgs_Qint16_Symmetric(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i16:f32, 3.0517578125E-5>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i16:f32, 3.0517578125E-5>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = -1.0 : f32, max = 0.999969482 : f32, num_bits = 16, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verify that lowering to barriers of unranked tensors functions.
|
|
// CHECK-LABEL: fakeQuantArgs_UnrankedTensor
|
|
func @fakeQuantArgs_UnrankedTensor(tensor<f32>) -> tensor<f32> {
|
|
^bb0(%arg0: tensor<f32>):
|
|
// CHECK: %0 = "quant.qcast"(%arg0) : (tensor<f32>)
|
|
// CHECK-SAME: -> tensor<!quant.uniform<u8:f32, 0.0039215686274509803>>
|
|
// CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<!quant.uniform<u8:f32, 0.0039215686274509803>>)
|
|
// CHECK-SAME: -> tensor<f32>
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.0 : f32, max = 1.0 : f32, num_bits = 8
|
|
} : (tensor<f32>) -> tensor<f32>
|
|
return %0 : tensor<f32>
|
|
}
|
|
|
|
// -----
|
|
// CHECK-LABEL: fakeQuantArgs_all_positive
|
|
func @fakeQuantArgs_all_positive(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = 0.5 : f32, max = 1.5 : f32, num_bits = 8, narrow_range = false, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// CHECK-LABEL: fakeQuantArgs_all_negative
|
|
func @fakeQuantArgs_all_negative(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
|
|
// CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:127>>
|
|
// CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:127>>)
|
|
// CHECK-SAME: -> tensor<8x4x3xf32>
|
|
|
|
%0 = "quant.const_fake_quant"(%arg0) {
|
|
min = -1.5 : f32, max = -0.5 : f32, num_bits = 8, narrow_range = false, is_signed = true
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|
|
|
|
// -----
|
|
// Verifies a qint8 per axis
|
|
// CHECK-LABEL: fakeQuantPerAxis
|
|
func @fakeQuantPerAxis(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> {
|
|
^bb0(%arg0: tensor<8x4x3xf32>):
|
|
|
|
// CHECK: %[[q:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>)
|
|
// CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32:2, {7.812500e-03,1.000000e+00:-128,0.0039215686274509803:-128}>>
|
|
// CHECK: %[[d:.*]] = "quant.dcast"(%[[q]])
|
|
// CHECK-SAME: (tensor<8x4x3x!quant.uniform<i8:f32:2, {7.812500e-03,1.000000e+00:-128,0.0039215686274509803:-128}>>)
|
|
|
|
%0 = "quant.const_fake_quant_per_axis"(%arg0) {
|
|
min = [-1.0 : f32, 0.0 : f32, 0.0 : f32],
|
|
max = [0.9921875 : f32, 0.0: f32, 1.0 : f32],
|
|
num_bits = 8, narrow_range = false, is_signed = true, axis = 2
|
|
} : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32>
|
|
return %0 : tensor<8x4x3xf32>
|
|
}
|