GlobalAveragePoolingOperatorTester Class — pytorch Architecture
Architecture documentation for the GlobalAveragePoolingOperatorTester class in global-average-pooling-operator-tester.h from the pytorch codebase.
Entity Profile
Relationship Graph
Source Code
aten/src/ATen/native/quantized/cpu/qnnpack/test/global-average-pooling-operator-tester.h lines 21–253
class GlobalAveragePoolingOperatorTester {
public:
inline GlobalAveragePoolingOperatorTester& channels(size_t channels) {
assert(channels != 0);
this->channels_ = channels;
return *this;
}
inline size_t channels() const {
return this->channels_;
}
inline GlobalAveragePoolingOperatorTester& width(size_t width) {
assert(width != 0);
this->width_ = width;
return *this;
}
inline size_t width() const {
return this->width_;
}
inline GlobalAveragePoolingOperatorTester& inputStride(size_t inputStride) {
assert(inputStride != 0);
this->inputStride_ = inputStride;
return *this;
}
inline size_t inputStride() const {
if (this->inputStride_ == 0) {
return channels();
} else {
assert(this->inputStride_ >= channels());
return this->inputStride_;
}
}
inline GlobalAveragePoolingOperatorTester& outputStride(size_t outputStride) {
assert(outputStride != 0);
this->outputStride_ = outputStride;
return *this;
}
inline size_t outputStride() const {
if (this->outputStride_ == 0) {
return channels();
} else {
assert(this->outputStride_ >= channels());
return this->outputStride_;
}
}
inline GlobalAveragePoolingOperatorTester& batchSize(size_t batchSize) {
this->batchSize_ = batchSize;
return *this;
}
inline size_t batchSize() const {
return this->batchSize_;
}
inline GlobalAveragePoolingOperatorTester& inputScale(float inputScale) {
assert(inputScale > 0.0f);
assert(std::isnormal(inputScale));
this->inputScale_ = inputScale;
return *this;
}
inline float inputScale() const {
return this->inputScale_;
}
inline GlobalAveragePoolingOperatorTester& inputZeroPoint(
uint8_t inputZeroPoint) {
this->inputZeroPoint_ = inputZeroPoint;
return *this;
}
inline uint8_t inputZeroPoint() const {
return this->inputZeroPoint_;
}
inline GlobalAveragePoolingOperatorTester& outputScale(float outputScale) {
assert(outputScale > 0.0f);
assert(std::isnormal(outputScale));
this->outputScale_ = outputScale;
return *this;
}
inline float outputScale() const {
return this->outputScale_;
}
inline GlobalAveragePoolingOperatorTester& outputZeroPoint(
uint8_t outputZeroPoint) {
this->outputZeroPoint_ = outputZeroPoint;
return *this;
}
inline uint8_t outputZeroPoint() const {
return this->outputZeroPoint_;
}
inline GlobalAveragePoolingOperatorTester& outputMin(uint8_t outputMin) {
this->outputMin_ = outputMin;
return *this;
}
inline uint8_t outputMin() const {
return this->outputMin_;
}
inline GlobalAveragePoolingOperatorTester& outputMax(uint8_t outputMax) {
this->outputMax_ = outputMax;
return *this;
}
inline uint8_t outputMax() const {
return this->outputMax_;
}
inline GlobalAveragePoolingOperatorTester& iterations(size_t iterations) {
this->iterations_ = iterations;
return *this;
}
inline size_t iterations() const {
return this->iterations_;
}
void testQ8() const {
std::random_device randomDevice;
auto rng = std::mt19937(randomDevice());
auto u8rng = std::bind(std::uniform_int_distribution<uint8_t>(), rng);
std::vector<uint8_t> input(
(batchSize() * width() - 1) * inputStride() + channels());
std::vector<uint8_t> output(batchSize() * outputStride());
std::vector<float> outputRef(batchSize() * channels());
for (size_t iteration = 0; iteration < iterations(); iteration++) {
std::generate(input.begin(), input.end(), std::ref(u8rng));
std::fill(output.begin(), output.end(), 0xA5);
/* Compute reference results */
const double scale =
double(inputScale()) / (double(width()) * double(outputScale()));
for (size_t i = 0; i < batchSize(); i++) {
for (size_t j = 0; j < channels(); j++) {
double acc = 0.0f;
for (size_t k = 0; k < width(); k++) {
acc += double(
int32_t(input[(i * width() + k) * inputStride() + j]) -
int32_t(inputZeroPoint()));
}
outputRef[i * channels() + j] =
float(acc * scale + double(outputZeroPoint()));
outputRef[i * channels() + j] = std::min<float>(
outputRef[i * channels() + j], float(outputMax()));
outputRef[i * channels() + j] = std::max<float>(
outputRef[i * channels() + j], float(outputMin()));
}
}
/* Create, setup, run, and destroy Add operator */
ASSERT_EQ(pytorch_qnnp_status_success, pytorch_qnnp_initialize());
pytorch_qnnp_operator_t globalAveragePoolingOp = nullptr;
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_create_global_average_pooling_nwc_q8(
channels(),
inputZeroPoint(),
inputScale(),
outputZeroPoint(),
outputScale(),
outputMin(),
outputMax(),
0,
&globalAveragePoolingOp));
ASSERT_NE(nullptr, globalAveragePoolingOp);
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_setup_global_average_pooling_nwc_q8(
globalAveragePoolingOp,
batchSize(),
width(),
input.data(),
inputStride(),
output.data(),
outputStride()));
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_run_operator(
globalAveragePoolingOp, nullptr /* thread pool */));
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_delete_operator(globalAveragePoolingOp));
globalAveragePoolingOp = nullptr;
/* Verify results */
for (size_t i = 0; i < batchSize(); i++) {
for (size_t c = 0; c < channels(); c++) {
ASSERT_LE(
uint32_t(output[i * outputStride() + c]), uint32_t(outputMax()));
ASSERT_GE(
uint32_t(output[i * outputStride() + c]), uint32_t(outputMin()));
ASSERT_NEAR(
float(int32_t(output[i * outputStride() + c])),
outputRef[i * channels() + c],
0.80f)
<< "in batch index " << i << ", channel " << c;
}
}
}
}
private:
size_t batchSize_{1};
size_t width_{1};
size_t channels_{1};
size_t inputStride_{0};
size_t outputStride_{0};
float inputScale_{1.0f};
float outputScale_{1.0f};
uint8_t inputZeroPoint_{121};
uint8_t outputZeroPoint_{133};
uint8_t outputMin_{0};
uint8_t outputMax_{255};
size_t iterations_{1};
};
Domain
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free