copy_value_with_pad Class — pytorch Architecture
Architecture documentation for the copy_value_with_pad class in FlashAttentionKernel.cpp from the pytorch codebase.
Entity Profile
Source Code
aten/src/ATen/native/cpu/FlashAttentionKernel.cpp lines 216–266
template <typename scalar_t>
inline void copy_value_with_pad(
const scalar_t* value_ptr,
scalar_t* dst_ptr,
int64_t rows,
int64_t cols,
int64_t prows,
int64_t pcols,
int64_t ldi) {
auto vec_size = at::vec::Vectorized<scalar_t>::size();
int64_t i = 0;
for (; i < rows; i++) {
int64_t j = 0;
for (; j < cols - (cols % vec_size); j += vec_size) {
auto vec_v =
at::vec::Vectorized<scalar_t>::loadu(value_ptr + i * ldi + j);
vec_v.store(dst_ptr + i * pcols + j);
}
if (j < cols) {
auto vec_v = at::vec::Vectorized<scalar_t>::loadu(
value_ptr + i * ldi + j, cols - j);
vec_v.store(dst_ptr + i * pcols + j, cols - j);
}
// col padding
auto psize = pcols - cols;
if (psize > 0) {
auto zero_vec = at::vec::Vectorized<scalar_t>(0);
int64_t pj = 0;
for (; pj < psize - (psize % vec_size); pj += vec_size) {
zero_vec.store(dst_ptr + i * pcols + cols + pj);
}
if (pj < psize) {
zero_vec.store(dst_ptr + i * pcols + cols + pj, psize - pj);
}
}
}
// row padding
for (; i < prows; i++) {
auto zero_vec = at::vec::Vectorized<scalar_t>(0);
int64_t j = 0;
for (; j < pcols - (pcols % vec_size); j += vec_size) {
zero_vec.store(dst_ptr + i * pcols + j);
}
if (j < pcols) {
zero_vec.store(dst_ptr + i * pcols + j, pcols - j);
}
}
}
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free