@@ -2640,7 +2640,14 @@ def DilationHFactor(self):
2640
2640
return self._tab.Get(flatbuffers.number_types.Int32Flags, o + self._tab.Pos)
2641
2641
return 1
2642
2642
2643
- def Conv2DOptionsStart(builder): builder.StartObject(6)
2643
+ # Conv2DOptions
2644
+ def QuantizedBiasType(self):
2645
+ o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(16))
2646
+ if o != 0:
2647
+ return self._tab.Get(flatbuffers.number_types.Int8Flags, o + self._tab.Pos)
2648
+ return 0
2649
+
2650
+ def Conv2DOptionsStart(builder): builder.StartObject(7)
2644
2651
def Start(builder):
2645
2652
return Conv2DOptionsStart(builder)
2646
2653
def Conv2DOptionsAddPadding(builder, padding): builder.PrependInt8Slot(0, padding, 0)
@@ -2661,6 +2668,9 @@ def AddDilationWFactor(builder, dilationWFactor):
2661
2668
def Conv2DOptionsAddDilationHFactor(builder, dilationHFactor): builder.PrependInt32Slot(5, dilationHFactor, 1)
2662
2669
def AddDilationHFactor(builder, dilationHFactor):
2663
2670
return Conv2DOptionsAddDilationHFactor(builder, dilationHFactor)
2671
+ def Conv2DOptionsAddQuantizedBiasType(builder, quantizedBiasType): builder.PrependInt8Slot(6, quantizedBiasType, 0)
2672
+ def AddQuantizedBiasType(builder, quantizedBiasType):
2673
+ return Conv2DOptionsAddQuantizedBiasType(builder, quantizedBiasType)
2664
2674
def Conv2DOptionsEnd(builder): return builder.EndObject()
2665
2675
def End(builder):
2666
2676
return Conv2DOptionsEnd(builder)
@@ -2675,6 +2685,7 @@ def __init__(self):
2675
2685
self.fusedActivationFunction = 0 # type: int
2676
2686
self.dilationWFactor = 1 # type: int
2677
2687
self.dilationHFactor = 1 # type: int
2688
+ self.quantizedBiasType = 0 # type: int
2678
2689
2679
2690
@classmethod
2680
2691
def InitFromBuf(cls, buf, pos):
@@ -2698,6 +2709,7 @@ def _UnPack(self, conv2doptions):
2698
2709
self.fusedActivationFunction = conv2doptions.FusedActivationFunction()
2699
2710
self.dilationWFactor = conv2doptions.DilationWFactor()
2700
2711
self.dilationHFactor = conv2doptions.DilationHFactor()
2712
+ self.quantizedBiasType = conv2doptions.QuantizedBiasType()
2701
2713
2702
2714
# Conv2DOptionsT
2703
2715
def Pack(self, builder):
@@ -2708,6 +2720,7 @@ def Pack(self, builder):
2708
2720
Conv2DOptionsAddFusedActivationFunction(builder, self.fusedActivationFunction)
2709
2721
Conv2DOptionsAddDilationWFactor(builder, self.dilationWFactor)
2710
2722
Conv2DOptionsAddDilationHFactor(builder, self.dilationHFactor)
2723
+ Conv2DOptionsAddQuantizedBiasType(builder, self.quantizedBiasType)
2711
2724
conv2doptions = Conv2DOptionsEnd(builder)
2712
2725
return conv2doptions
2713
2726
# automatically generated by the FlatBuffers compiler, do not modify
@@ -4512,7 +4525,14 @@ def AsymmetricQuantizeInputs(self):
4512
4525
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
4513
4526
return False
4514
4527
4515
- def FullyConnectedOptionsStart(builder): builder.StartObject(4)
4528
+ # FullyConnectedOptions
4529
+ def QuantizedBiasType(self):
4530
+ o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(12))
4531
+ if o != 0:
4532
+ return self._tab.Get(flatbuffers.number_types.Int8Flags, o + self._tab.Pos)
4533
+ return 0
4534
+
4535
+ def FullyConnectedOptionsStart(builder): builder.StartObject(5)
4516
4536
def Start(builder):
4517
4537
return FullyConnectedOptionsStart(builder)
4518
4538
def FullyConnectedOptionsAddFusedActivationFunction(builder, fusedActivationFunction): builder.PrependInt8Slot(0, fusedActivationFunction, 0)
@@ -4527,6 +4547,9 @@ def AddKeepNumDims(builder, keepNumDims):
4527
4547
def FullyConnectedOptionsAddAsymmetricQuantizeInputs(builder, asymmetricQuantizeInputs): builder.PrependBoolSlot(3, asymmetricQuantizeInputs, 0)
4528
4548
def AddAsymmetricQuantizeInputs(builder, asymmetricQuantizeInputs):
4529
4549
return FullyConnectedOptionsAddAsymmetricQuantizeInputs(builder, asymmetricQuantizeInputs)
4550
+ def FullyConnectedOptionsAddQuantizedBiasType(builder, quantizedBiasType): builder.PrependInt8Slot(4, quantizedBiasType, 0)
4551
+ def AddQuantizedBiasType(builder, quantizedBiasType):
4552
+ return FullyConnectedOptionsAddQuantizedBiasType(builder, quantizedBiasType)
4530
4553
def FullyConnectedOptionsEnd(builder): return builder.EndObject()
4531
4554
def End(builder):
4532
4555
return FullyConnectedOptionsEnd(builder)
@@ -4539,6 +4562,7 @@ def __init__(self):
4539
4562
self.weightsFormat = 0 # type: int
4540
4563
self.keepNumDims = False # type: bool
4541
4564
self.asymmetricQuantizeInputs = False # type: bool
4565
+ self.quantizedBiasType = 0 # type: int
4542
4566
4543
4567
@classmethod
4544
4568
def InitFromBuf(cls, buf, pos):
@@ -4560,6 +4584,7 @@ def _UnPack(self, fullyConnectedOptions):
4560
4584
self.weightsFormat = fullyConnectedOptions.WeightsFormat()
4561
4585
self.keepNumDims = fullyConnectedOptions.KeepNumDims()
4562
4586
self.asymmetricQuantizeInputs = fullyConnectedOptions.AsymmetricQuantizeInputs()
4587
+ self.quantizedBiasType = fullyConnectedOptions.QuantizedBiasType()
4563
4588
4564
4589
# FullyConnectedOptionsT
4565
4590
def Pack(self, builder):
@@ -4568,6 +4593,7 @@ def Pack(self, builder):
4568
4593
FullyConnectedOptionsAddWeightsFormat(builder, self.weightsFormat)
4569
4594
FullyConnectedOptionsAddKeepNumDims(builder, self.keepNumDims)
4570
4595
FullyConnectedOptionsAddAsymmetricQuantizeInputs(builder, self.asymmetricQuantizeInputs)
4596
+ FullyConnectedOptionsAddQuantizedBiasType(builder, self.quantizedBiasType)
4571
4597
fullyConnectedOptions = FullyConnectedOptionsEnd(builder)
4572
4598
return fullyConnectedOptions
4573
4599
# automatically generated by the FlatBuffers compiler, do not modify
@@ -16436,7 +16462,14 @@ def FusedActivationFunction(self):
16436
16462
return self._tab.Get(flatbuffers.number_types.Int8Flags, o + self._tab.Pos)
16437
16463
return 0
16438
16464
16439
- def TransposeConvOptionsStart(builder): builder.StartObject(4)
16465
+ # TransposeConvOptions
16466
+ def QuantizedBiasType(self):
16467
+ o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(12))
16468
+ if o != 0:
16469
+ return self._tab.Get(flatbuffers.number_types.Int8Flags, o + self._tab.Pos)
16470
+ return 0
16471
+
16472
+ def TransposeConvOptionsStart(builder): builder.StartObject(5)
16440
16473
def Start(builder):
16441
16474
return TransposeConvOptionsStart(builder)
16442
16475
def TransposeConvOptionsAddPadding(builder, padding): builder.PrependInt8Slot(0, padding, 0)
@@ -16451,6 +16484,9 @@ def AddStrideH(builder, strideH):
16451
16484
def TransposeConvOptionsAddFusedActivationFunction(builder, fusedActivationFunction): builder.PrependInt8Slot(3, fusedActivationFunction, 0)
16452
16485
def AddFusedActivationFunction(builder, fusedActivationFunction):
16453
16486
return TransposeConvOptionsAddFusedActivationFunction(builder, fusedActivationFunction)
16487
+ def TransposeConvOptionsAddQuantizedBiasType(builder, quantizedBiasType): builder.PrependInt8Slot(4, quantizedBiasType, 0)
16488
+ def AddQuantizedBiasType(builder, quantizedBiasType):
16489
+ return TransposeConvOptionsAddQuantizedBiasType(builder, quantizedBiasType)
16454
16490
def TransposeConvOptionsEnd(builder): return builder.EndObject()
16455
16491
def End(builder):
16456
16492
return TransposeConvOptionsEnd(builder)
@@ -16463,6 +16499,7 @@ def __init__(self):
16463
16499
self.strideW = 0 # type: int
16464
16500
self.strideH = 0 # type: int
16465
16501
self.fusedActivationFunction = 0 # type: int
16502
+ self.quantizedBiasType = 0 # type: int
16466
16503
16467
16504
@classmethod
16468
16505
def InitFromBuf(cls, buf, pos):
@@ -16484,6 +16521,7 @@ def _UnPack(self, transposeConvOptions):
16484
16521
self.strideW = transposeConvOptions.StrideW()
16485
16522
self.strideH = transposeConvOptions.StrideH()
16486
16523
self.fusedActivationFunction = transposeConvOptions.FusedActivationFunction()
16524
+ self.quantizedBiasType = transposeConvOptions.QuantizedBiasType()
16487
16525
16488
16526
# TransposeConvOptionsT
16489
16527
def Pack(self, builder):
@@ -16492,6 +16530,7 @@ def Pack(self, builder):
16492
16530
TransposeConvOptionsAddStrideW(builder, self.strideW)
16493
16531
TransposeConvOptionsAddStrideH(builder, self.strideH)
16494
16532
TransposeConvOptionsAddFusedActivationFunction(builder, self.fusedActivationFunction)
16533
+ TransposeConvOptionsAddQuantizedBiasType(builder, self.quantizedBiasType)
16495
16534
transposeConvOptions = TransposeConvOptionsEnd(builder)
16496
16535
return transposeConvOptions
16497
16536
# automatically generated by the FlatBuffers compiler, do not modify
0 commit comments