LeakyReLUOperatorTester Class — pytorch Architecture
Architecture documentation for the LeakyReLUOperatorTester class in leaky-relu-operator-tester.h from the pytorch codebase.
Entity Profile
Relationship Graph
Source Code
aten/src/ATen/native/quantized/cpu/qnnpack/test/leaky-relu-operator-tester.h lines 22–240
class LeakyReLUOperatorTester {
public:
inline LeakyReLUOperatorTester& channels(size_t channels) {
assert(channels != 0);
this->channels_ = channels;
return *this;
}
inline size_t channels() const {
return this->channels_;
}
inline LeakyReLUOperatorTester& inputStride(size_t inputStride) {
assert(inputStride != 0);
this->inputStride_ = inputStride;
return *this;
}
inline size_t inputStride() const {
if (this->inputStride_ == 0) {
return this->channels_;
} else {
assert(this->inputStride_ >= this->channels_);
return this->inputStride_;
}
}
inline LeakyReLUOperatorTester& outputStride(size_t outputStride) {
assert(outputStride != 0);
this->outputStride_ = outputStride;
return *this;
}
inline size_t outputStride() const {
if (this->outputStride_ == 0) {
return this->channels_;
} else {
assert(this->outputStride_ >= this->channels_);
return this->outputStride_;
}
}
inline LeakyReLUOperatorTester& batchSize(size_t batchSize) {
this->batchSize_ = batchSize;
return *this;
}
inline size_t batchSize() const {
return this->batchSize_;
}
inline LeakyReLUOperatorTester& negativeSlope(float negativeSlope) {
assert(negativeSlope > 0.0f);
assert(negativeSlope < 1.0f);
this->negativeSlope_ = negativeSlope;
return *this;
}
inline float negativeSlope() const {
return this->negativeSlope_;
}
inline LeakyReLUOperatorTester& inputScale(float inputScale) {
assert(inputScale > 0.0f);
assert(std::isnormal(inputScale));
this->inputScale_ = inputScale;
return *this;
}
inline float inputScale() const {
return this->inputScale_;
}
inline LeakyReLUOperatorTester& inputZeroPoint(uint8_t inputZeroPoint) {
this->inputZeroPoint_ = inputZeroPoint;
return *this;
}
inline uint8_t inputZeroPoint() const {
return this->inputZeroPoint_;
}
inline LeakyReLUOperatorTester& outputScale(float outputScale) {
assert(outputScale > 0.0f);
assert(std::isnormal(outputScale));
this->outputScale_ = outputScale;
return *this;
}
inline float outputScale() const {
return this->outputScale_;
}
inline LeakyReLUOperatorTester& outputZeroPoint(uint8_t outputZeroPoint) {
this->outputZeroPoint_ = outputZeroPoint;
return *this;
}
inline uint8_t outputZeroPoint() const {
return this->outputZeroPoint_;
}
inline LeakyReLUOperatorTester& qmin(uint8_t qmin) {
this->qmin_ = qmin;
return *this;
}
inline uint8_t qmin() const {
return this->qmin_;
}
inline LeakyReLUOperatorTester& qmax(uint8_t qmax) {
this->qmax_ = qmax;
return *this;
}
inline uint8_t qmax() const {
return this->qmax_;
}
inline LeakyReLUOperatorTester& iterations(size_t iterations) {
this->iterations_ = iterations;
return *this;
}
inline size_t iterations() const {
return this->iterations_;
}
void testQ8() const {
std::random_device randomDevice;
auto rng = std::mt19937(randomDevice());
auto u8rng = std::bind(std::uniform_int_distribution<uint8_t>(), rng);
std::vector<uint8_t> input((batchSize() - 1) * inputStride() + channels());
std::vector<uint8_t> output(
(batchSize() - 1) * outputStride() + channels());
std::vector<float> outputRef(batchSize() * channels());
for (size_t iteration = 0; iteration < iterations(); iteration++) {
std::generate(input.begin(), input.end(), std::ref(u8rng));
std::fill(output.begin(), output.end(), 0xA5);
/* Compute reference results */
for (size_t i = 0; i < batchSize(); i++) {
for (size_t c = 0; c < channels(); c++) {
const float x = inputScale() *
(int32_t(input[i * inputStride() + c]) -
int32_t(inputZeroPoint()));
float y = (x < 0.0f ? x * negativeSlope() : x) / outputScale();
y = std::min<float>(y, int32_t(qmax()) - int32_t(outputZeroPoint()));
y = std::max<float>(y, int32_t(qmin()) - int32_t(outputZeroPoint()));
outputRef[i * channels() + c] = y + float(int32_t(outputZeroPoint()));
}
}
/* Create, setup, run, and destroy LeakyReLU operator */
ASSERT_EQ(pytorch_qnnp_status_success, pytorch_qnnp_initialize());
pytorch_qnnp_operator_t leakyReLUOp = nullptr;
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_create_leaky_relu_nc_q8(
channels(),
negativeSlope(),
inputZeroPoint(),
inputScale(),
outputZeroPoint(),
outputScale(),
qmin(),
qmax(),
0,
&leakyReLUOp));
ASSERT_NE(nullptr, leakyReLUOp);
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_setup_leaky_relu_nc_q8(
leakyReLUOp,
batchSize(),
input.data(),
inputStride(),
output.data(),
outputStride()));
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_run_operator(leakyReLUOp, nullptr /* thread pool */));
ASSERT_EQ(
pytorch_qnnp_status_success,
pytorch_qnnp_delete_operator(leakyReLUOp));
leakyReLUOp = nullptr;
/* Verify results */
for (size_t i = 0; i < batchSize(); i++) {
for (size_t c = 0; c < channels(); c++) {
ASSERT_NEAR(
float(int32_t(output[i * outputStride() + c])),
outputRef[i * channels() + c],
0.6f);
}
}
}
}
private:
size_t batchSize_{1};
size_t channels_{1};
size_t inputStride_{0};
size_t outputStride_{0};
float negativeSlope_{0.5f};
float outputScale_{0.75f};
uint8_t outputZeroPoint_{133};
float inputScale_{1.25f};
uint8_t inputZeroPoint_{121};
uint8_t qmin_{0};
uint8_t qmax_{255};
size_t iterations_{15};
};
Domain
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free