| // Auto-generated file. Do not edit! |
| // Template: src/f32-ppmm/scalar.c.in |
| // Generator: tools/xngen |
| // |
| // Copyright 2019 Google LLC |
| // |
| // This source code is licensed under the BSD-style license found in the |
| // LICENSE file in the root directory of this source tree. |
| |
| #include <assert.h> |
| |
| #include <xnnpack/math.h> |
| #include <xnnpack/ppmm.h> |
| |
| |
| void xnn_f32_ppmm_ukernel_4x2__scalar( |
| size_t mr, |
| size_t nc, |
| size_t kc, |
| const float*restrict a, |
| const float*restrict w, |
| float*restrict c, |
| size_t cm_stride, |
| size_t cn_stride, |
| const union xnn_f32_minmax_params params[restrict static 1]) |
| { |
| assert(mr != 0); |
| assert(mr <= 4); |
| assert(nc != 0); |
| assert(kc != 0); |
| assert(kc % sizeof(float) == 0); |
| |
| float* c0 = c; |
| float* c1 = (float*) ((uintptr_t) c0 + cm_stride); |
| if XNN_UNPREDICTABLE(mr < 2) { |
| c1 = c0; |
| } |
| float* c2 = (float*) ((uintptr_t) c1 + cm_stride); |
| if XNN_UNPREDICTABLE(mr <= 2) { |
| c2 = c1; |
| } |
| float* c3 = (float*) ((uintptr_t) c2 + cm_stride); |
| if XNN_UNPREDICTABLE(mr != 4) { |
| c3 = c2; |
| } |
| |
| do { |
| float vacc0x0 = w[0]; |
| float vacc0x1 = w[1]; |
| float vacc1x0 = vacc0x0; |
| float vacc1x1 = vacc0x1; |
| float vacc2x0 = vacc0x0; |
| float vacc2x1 = vacc0x1; |
| float vacc3x0 = vacc0x0; |
| float vacc3x1 = vacc0x1; |
| w += 2; |
| |
| size_t k = kc; |
| do { |
| const float va0 = a[0]; |
| const float va1 = a[1]; |
| const float va2 = a[2]; |
| const float va3 = a[3]; |
| a += 4; |
| |
| const float vb0 = w[0]; |
| const float vb1 = w[1]; |
| w += 2; |
| |
| vacc0x0 += va0 * vb0; |
| vacc1x0 += va1 * vb0; |
| vacc2x0 += va2 * vb0; |
| vacc3x0 += va3 * vb0; |
| vacc0x1 += va0 * vb1; |
| vacc1x1 += va1 * vb1; |
| vacc2x1 += va2 * vb1; |
| vacc3x1 += va3 * vb1; |
| |
| k -= sizeof(float); |
| } while (k != 0); |
| |
| const float vmax = params->scalar.max; |
| vacc0x0 = math_min_f32(vacc0x0, vmax); |
| vacc1x0 = math_min_f32(vacc1x0, vmax); |
| vacc2x0 = math_min_f32(vacc2x0, vmax); |
| vacc3x0 = math_min_f32(vacc3x0, vmax); |
| vacc0x1 = math_min_f32(vacc0x1, vmax); |
| vacc1x1 = math_min_f32(vacc1x1, vmax); |
| vacc2x1 = math_min_f32(vacc2x1, vmax); |
| vacc3x1 = math_min_f32(vacc3x1, vmax); |
| |
| const float vmin = params->scalar.min; |
| vacc0x0 = math_max_f32(vacc0x0, vmin); |
| vacc1x0 = math_max_f32(vacc1x0, vmin); |
| vacc2x0 = math_max_f32(vacc2x0, vmin); |
| vacc3x0 = math_max_f32(vacc3x0, vmin); |
| vacc0x1 = math_max_f32(vacc0x1, vmin); |
| vacc1x1 = math_max_f32(vacc1x1, vmin); |
| vacc2x1 = math_max_f32(vacc2x1, vmin); |
| vacc3x1 = math_max_f32(vacc3x1, vmin); |
| |
| if XNN_LIKELY(nc >= 2) { |
| c3[0] = vacc3x0; |
| c3[1] = vacc3x1; |
| c2[0] = vacc2x0; |
| c2[1] = vacc2x1; |
| c1[0] = vacc1x0; |
| c1[1] = vacc1x1; |
| c0[0] = vacc0x0; |
| c0[1] = vacc0x1; |
| |
| a = (const float*) ((uintptr_t) a - kc * 4); |
| |
| c3 = (float*) ((uintptr_t) c3 + cn_stride); |
| c2 = (float*) ((uintptr_t) c2 + cn_stride); |
| c1 = (float*) ((uintptr_t) c1 + cn_stride); |
| c0 = (float*) ((uintptr_t) c0 + cn_stride); |
| |
| nc -= 2; |
| } else { |
| if (nc & 1) { |
| *c3 = vacc3x0; |
| *c2 = vacc2x0; |
| *c1 = vacc1x0; |
| *c0 = vacc0x0; |
| } |
| |
| nc = 0; |
| } |
| } while (nc != 0); |
| } |