diff options
author | Tim Rowley <[email protected]> | 2017-03-29 12:58:18 -0500 |
---|---|---|
committer | Tim Rowley <[email protected]> | 2017-04-05 18:20:45 -0500 |
commit | d5157ddca4072856e0afce3d7af8929a7d387044 (patch) | |
tree | fd993fd46af84bef289447493fcb38fcf1ae23a2 | |
parent | b8515d5c0f9c97ede62082e47855e885463f4b76 (diff) |
swr: [rasterizer core] SIMD16 Frontend WIP
Implement widened binner for SIMD16
Reviewed-by: Bruce Cherniak <[email protected]>
-rw-r--r-- | src/gallium/drivers/swr/rasterizer/common/simd16intrin.h | 44 | ||||
-rw-r--r-- | src/gallium/drivers/swr/rasterizer/core/binner.cpp | 1222 | ||||
-rw-r--r-- | src/gallium/drivers/swr/rasterizer/core/frontend.h | 98 | ||||
-rw-r--r-- | src/gallium/drivers/swr/rasterizer/core/utils.h | 10 |
4 files changed, 1299 insertions, 75 deletions
diff --git a/src/gallium/drivers/swr/rasterizer/common/simd16intrin.h b/src/gallium/drivers/swr/rasterizer/common/simd16intrin.h index 3b43d510e68..e5c34c2c876 100644 --- a/src/gallium/drivers/swr/rasterizer/common/simd16intrin.h +++ b/src/gallium/drivers/swr/rasterizer/common/simd16intrin.h @@ -46,10 +46,6 @@ struct simd16scalari }; typedef uint16_t simd16mask; -#define _simd16_masklo(mask) ((mask) & 0xFF) -#define _simd16_maskhi(mask) (((mask) >> 8)) -#define _simd16_setmask(hi, lo) (((hi) << 8) | (lo)) - #else typedef __m512 simd16scalar; typedef __m512d simd16scalard; @@ -60,6 +56,10 @@ typedef __mmask16 simd16mask; #error Unsupported vector width #endif//KNOB_SIMD16_WIDTH == 16 +#define _simd16_masklo(mask) ((mask) & 0xFF) +#define _simd16_maskhi(mask) (((mask) >> 8) & 0xFF) +#define _simd16_setmask(hi, lo) (((hi) << 8) | (lo)) + OSALIGN(union, KNOB_SIMD16_BYTES) simd16vector { simd16scalar v[4]; @@ -383,32 +383,26 @@ SIMD16_EMU_AVX512_2(simd16scalar, _simd16_max_ps, _mm256_max_ps) INLINE simd16mask _simd16_movemask_ps(simd16scalar a) { - simd16mask mask; + simdmask mask_lo = _mm256_movemask_ps(a.lo); + simdmask mask_hi = _mm256_movemask_ps(a.hi); - reinterpret_cast<uint8_t *>(&mask)[0] = _mm256_movemask_ps(a.lo); - reinterpret_cast<uint8_t *>(&mask)[1] = _mm256_movemask_ps(a.hi); - - return mask; + return static_cast<simd16mask>(mask_lo) | (static_cast<simd16mask>(mask_hi) << 8); } INLINE simd16mask _simd16_movemask_pd(simd16scalard a) { - simd16mask mask; + simdmask mask_lo = _mm256_movemask_pd(a.lo); + simdmask mask_hi = _mm256_movemask_pd(a.hi); - reinterpret_cast<uint8_t *>(&mask)[0] = _mm256_movemask_pd(a.lo); - reinterpret_cast<uint8_t *>(&mask)[1] = _mm256_movemask_pd(a.hi); - - return mask; + return static_cast<simd16mask>(mask_lo) | (static_cast<simd16mask>(mask_hi) << 4); } -INLINE simd16mask _simd16_movemask_epi8(simd16scalari a) +INLINE uint64_t _simd16_movemask_epi8(simd16scalari a) { - simd16mask mask; + uint32_t mask_lo = _mm256_movemask_epi8(a.lo); + uint32_t mask_hi = _mm256_movemask_epi8(a.hi); - reinterpret_cast<uint8_t *>(&mask)[0] = _mm256_movemask_epi8(a.lo); - reinterpret_cast<uint8_t *>(&mask)[1] = _mm256_movemask_epi8(a.hi); - - return mask; + return static_cast<uint64_t>(mask_lo) | (static_cast<uint64_t>(mask_hi) << 32); } INLINE simd16scalari _simd16_cvtps_epi32(simd16scalar a) @@ -809,12 +803,10 @@ INLINE simd16mask _simd16_scalari2mask(simd16scalari mask) return _mm512_cmpneq_epu32_mask(mask, _mm512_setzero_epi32()); } -#if 0 INLINE simd16mask _simd16_scalard2mask(simd16scalard mask) { - return _mm512_cmpneq_epu64_mask(mask, _mm512_setzero_epi64()); + return _mm512_cmpneq_epu64_mask(_mm512_castpd_si512(mask), _mm512_setzero_si512()); } -#endif #define _simd16_setzero_ps _mm512_setzero_ps #define _simd16_setzero_si _mm512_setzero_si512 @@ -889,6 +881,7 @@ INLINE simd16scalari _simd16_blendv_epi32(simd16scalari a, simd16scalari b, cons } #define _simd16_mul_ps _mm512_mul_ps +#define _simd16_div_ps _mm512_div_ps #define _simd16_add_ps _mm512_add_ps #define _simd16_sub_ps _mm512_sub_ps #define _simd16_rsqrt_ps _mm512_rsqrt14_ps @@ -900,12 +893,10 @@ INLINE simd16mask _simd16_movemask_ps(simd16scalar a) return _simd16_scalari2mask(_mm512_castps_si512(a)); } -#if 0 INLINE simd16mask _simd16_movemask_pd(simd16scalard a) { - return _simd16_scalard2mask(_mm512i_castpd_si512(a)); + return _simd16_scalard2mask(a); } -#endif #if 0 INLINE int _simd16_movemask_epi8(simd16scalari a) @@ -1040,7 +1031,6 @@ INLINE simd16scalar _simd16_mask_i32gather_ps_temp(simd16scalar a, const float * #define _simd16_mask_i32gather_ps(a, m, index, mask, scale) _simd16_mask_i32gather_ps_temp<scale>(a, m, index, mask) #define _simd16_abs_epi32 _mm512_abs_epi32 -#define _simd16_cmpeq_epi64 _mm512_abs_epi32 INLINE simd16scalari _simd16_cmpeq_epi64(simd16scalari a, simd16scalari b) { diff --git a/src/gallium/drivers/swr/rasterizer/core/binner.cpp b/src/gallium/drivers/swr/rasterizer/core/binner.cpp index f00701f8192..9ec5bea76ee 100644 --- a/src/gallium/drivers/swr/rasterizer/core/binner.cpp +++ b/src/gallium/drivers/swr/rasterizer/core/binner.cpp @@ -37,15 +37,27 @@ // Function Prototype void BinPostSetupLines(DRAW_CONTEXT *pDC, PA_STATE& pa, uint32_t workerId, simdvector prims[3], simdscalar vRecipW[2], uint32_t primMask, simdscalari primID, simdscalari viewportIdx); +#if USE_SIMD16_FRONTEND +void BinPostSetupLines_simd16(DRAW_CONTEXT *pDC, PA_STATE& pa, uint32_t workerId, simd16vector prims[3], simd16scalar vRecipW[2], uint32_t primMask, simd16scalari primID, simd16scalari viewportIdx); +#endif + ////////////////////////////////////////////////////////////////////////// /// @brief Offsets added to post-viewport vertex positions based on /// raster state. static const simdscalar g_pixelOffsets[SWR_PIXEL_LOCATION_UL + 1] = { - _simd_set1_ps(0.0f), // SWR_PIXEL_LOCATION_CENTER - _simd_set1_ps(0.5f), // SWR_PIXEL_LOCATION_UL + _simd_set1_ps(0.0f), // SWR_PIXEL_LOCATION_CENTER + _simd_set1_ps(0.5f), // SWR_PIXEL_LOCATION_UL +}; + +#if USE_SIMD16_FRONTEND +static const simd16scalar g_pixelOffsets_simd16[SWR_PIXEL_LOCATION_UL + 1] = +{ + _simd16_set1_ps(0.0f), // SWR_PIXEL_LOCATION_CENTER + _simd16_set1_ps(0.5f), // SWR_PIXEL_LOCATION_UL }; +#endif ////////////////////////////////////////////////////////////////////////// /// @brief Convert the X,Y coords of a triangle to the requested Fixed /// Point precision from FP32. @@ -56,6 +68,15 @@ INLINE simdscalari fpToFixedPointVertical(const simdscalar vIn) return _simd_cvtps_epi32(vFixed); } +#if USE_SIMD16_FRONTEND +template <typename PT = FixedPointTraits<Fixed_16_8>> +INLINE simd16scalari fpToFixedPointVertical(const simd16scalar vIn) +{ + simd16scalar vFixed = _simd16_mul_ps(vIn, _simd16_set1_ps(PT::ScaleT::value)); + return _simd16_cvtps_epi32(vFixed); +} + +#endif ////////////////////////////////////////////////////////////////////////// /// @brief Helper function to set the X,Y coords of a triangle to the /// requested Fixed Point precision from FP32. @@ -72,6 +93,18 @@ INLINE static void FPToFixedPoint(const simdvector * const tri, simdscalari(&vXi vYi[2] = fpToFixedPointVertical(tri[2].y); } +#if USE_SIMD16_FRONTEND +INLINE static void FPToFixedPoint(const simd16vector * const tri, simd16scalari(&vXi)[3], simd16scalari(&vYi)[3]) +{ + vXi[0] = fpToFixedPointVertical(tri[0].x); + vYi[0] = fpToFixedPointVertical(tri[0].y); + vXi[1] = fpToFixedPointVertical(tri[1].x); + vYi[1] = fpToFixedPointVertical(tri[1].y); + vXi[2] = fpToFixedPointVertical(tri[2].x); + vYi[2] = fpToFixedPointVertical(tri[2].y); +} + +#endif ////////////////////////////////////////////////////////////////////////// /// @brief Calculate bounding box for current triangle /// @tparam CT: ConservativeRastFETraits type @@ -105,6 +138,37 @@ INLINE void calcBoundingBoxIntVertical(const simdvector * const tri, simdscalari bbox.ymax = vMaxY; } +#if USE_SIMD16_FRONTEND +template <typename CT> +INLINE void calcBoundingBoxIntVertical(const simd16vector * const tri, simd16scalari(&vX)[3], simd16scalari(&vY)[3], simd16BBox &bbox) +{ + simd16scalari vMinX = vX[0]; + + vMinX = _simd16_min_epi32(vMinX, vX[1]); + vMinX = _simd16_min_epi32(vMinX, vX[2]); + + simd16scalari vMaxX = vX[0]; + + vMaxX = _simd16_max_epi32(vMaxX, vX[1]); + vMaxX = _simd16_max_epi32(vMaxX, vX[2]); + + simd16scalari vMinY = vY[0]; + + vMinY = _simd16_min_epi32(vMinY, vY[1]); + vMinY = _simd16_min_epi32(vMinY, vY[2]); + + simd16scalari vMaxY = vY[0]; + + vMaxY = _simd16_max_epi32(vMaxY, vY[1]); + vMaxY = _simd16_max_epi32(vMaxY, vY[2]); + + bbox.xmin = vMinX; + bbox.xmax = vMaxX; + bbox.ymin = vMinY; + bbox.ymax = vMaxY; +} + +#endif ////////////////////////////////////////////////////////////////////////// /// @brief FEConservativeRastT specialization of calcBoundingBoxIntVertical /// Offsets BBox for conservative rast @@ -356,6 +420,96 @@ struct GatherScissors<8> } }; +#if USE_SIMD16_FRONTEND +template<size_t SimdWidth> +struct GatherScissors_simd16 +{ + static void Gather(const SWR_RECT* pScissorsInFixedPoint, const uint32_t* pViewportIndex, + simd16scalari &scisXmin, simd16scalari &scisYmin, + simd16scalari &scisXmax, simd16scalari &scisYmax) + { + SWR_INVALID("Unhandled Simd Width in Scissor Rect Gather"); + } +}; + +template<> +struct GatherScissors_simd16<16> +{ + static void Gather(const SWR_RECT* pScissorsInFixedPoint, const uint32_t* pViewportIndex, + simd16scalari &scisXmin, simd16scalari &scisYmin, + simd16scalari &scisXmax, simd16scalari &scisYmax) + { + scisXmin = _simd16_set_epi32(pScissorsInFixedPoint[pViewportIndex[0]].xmin, + pScissorsInFixedPoint[pViewportIndex[1]].xmin, + pScissorsInFixedPoint[pViewportIndex[2]].xmin, + pScissorsInFixedPoint[pViewportIndex[3]].xmin, + pScissorsInFixedPoint[pViewportIndex[4]].xmin, + pScissorsInFixedPoint[pViewportIndex[5]].xmin, + pScissorsInFixedPoint[pViewportIndex[6]].xmin, + pScissorsInFixedPoint[pViewportIndex[7]].xmin, + pScissorsInFixedPoint[pViewportIndex[8]].xmin, + pScissorsInFixedPoint[pViewportIndex[9]].xmin, + pScissorsInFixedPoint[pViewportIndex[10]].xmin, + pScissorsInFixedPoint[pViewportIndex[11]].xmin, + pScissorsInFixedPoint[pViewportIndex[12]].xmin, + pScissorsInFixedPoint[pViewportIndex[13]].xmin, + pScissorsInFixedPoint[pViewportIndex[14]].xmin, + pScissorsInFixedPoint[pViewportIndex[15]].xmin); + + scisYmin = _simd16_set_epi32(pScissorsInFixedPoint[pViewportIndex[0]].ymin, + pScissorsInFixedPoint[pViewportIndex[1]].ymin, + pScissorsInFixedPoint[pViewportIndex[2]].ymin, + pScissorsInFixedPoint[pViewportIndex[3]].ymin, + pScissorsInFixedPoint[pViewportIndex[4]].ymin, + pScissorsInFixedPoint[pViewportIndex[5]].ymin, + pScissorsInFixedPoint[pViewportIndex[6]].ymin, + pScissorsInFixedPoint[pViewportIndex[7]].ymin, + pScissorsInFixedPoint[pViewportIndex[8]].ymin, + pScissorsInFixedPoint[pViewportIndex[9]].ymin, + pScissorsInFixedPoint[pViewportIndex[10]].ymin, + pScissorsInFixedPoint[pViewportIndex[11]].ymin, + pScissorsInFixedPoint[pViewportIndex[12]].ymin, + pScissorsInFixedPoint[pViewportIndex[13]].ymin, + pScissorsInFixedPoint[pViewportIndex[14]].ymin, + pScissorsInFixedPoint[pViewportIndex[15]].ymin); + + scisXmax = _simd16_set_epi32(pScissorsInFixedPoint[pViewportIndex[0]].xmax, + pScissorsInFixedPoint[pViewportIndex[1]].xmax, + pScissorsInFixedPoint[pViewportIndex[2]].xmax, + pScissorsInFixedPoint[pViewportIndex[3]].xmax, + pScissorsInFixedPoint[pViewportIndex[4]].xmax, + pScissorsInFixedPoint[pViewportIndex[5]].xmax, + pScissorsInFixedPoint[pViewportIndex[6]].xmax, + pScissorsInFixedPoint[pViewportIndex[7]].xmax, + pScissorsInFixedPoint[pViewportIndex[8]].xmax, + pScissorsInFixedPoint[pViewportIndex[9]].xmax, + pScissorsInFixedPoint[pViewportIndex[10]].xmax, + pScissorsInFixedPoint[pViewportIndex[11]].xmax, + pScissorsInFixedPoint[pViewportIndex[12]].xmax, + pScissorsInFixedPoint[pViewportIndex[13]].xmax, + pScissorsInFixedPoint[pViewportIndex[14]].xmax, + pScissorsInFixedPoint[pViewportIndex[15]].xmax); + + scisYmax = _simd16_set_epi32(pScissorsInFixedPoint[pViewportIndex[0]].ymax, + pScissorsInFixedPoint[pViewportIndex[1]].ymax, + pScissorsInFixedPoint[pViewportIndex[2]].ymax, + pScissorsInFixedPoint[pViewportIndex[3]].ymax, + pScissorsInFixedPoint[pViewportIndex[4]].ymax, + pScissorsInFixedPoint[pViewportIndex[5]].ymax, + pScissorsInFixedPoint[pViewportIndex[6]].ymax, + pScissorsInFixedPoint[pViewportIndex[7]].ymax, + pScissorsInFixedPoint[pViewportIndex[8]].ymax, + pScissorsInFixedPoint[pViewportIndex[9]].ymax, + pScissorsInFixedPoint[pViewportIndex[10]].ymax, + pScissorsInFixedPoint[pViewportIndex[11]].ymax, + pScissorsInFixedPoint[pViewportIndex[12]].ymax, + pScissorsInFixedPoint[pViewportIndex[13]].ymax, + pScissorsInFixedPoint[pViewportIndex[14]].ymax, + pScissorsInFixedPoint[pViewportIndex[15]].ymax); + } +}; + +#endif typedef void(*PFN_PROCESS_ATTRIBUTES)(DRAW_CONTEXT*, PA_STATE&, uint32_t, uint32_t, float*); struct ProcessAttributesChooser @@ -873,34 +1027,444 @@ void BinTriangles_simd16( simd16scalari primID, simd16scalari viewportIdx) { - enum { VERTS_PER_PRIM = 3 }; + SWR_CONTEXT *pContext = pDC->pContext; - simdvector verts[VERTS_PER_PRIM]; + AR_BEGIN(FEBinTriangles, pDC->drawId); - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) + const API_STATE& state = GetApiState(pDC); + const SWR_RASTSTATE& rastState = state.rastState; + const SWR_FRONTEND_STATE& feState = state.frontendState; + const SWR_GS_STATE& gsState = state.gsState; + + MacroTileMgr *pTileMgr = pDC->pTileMgr; + + simd16scalar vRecipW0 = _simd16_set1_ps(1.0f); + simd16scalar vRecipW1 = _simd16_set1_ps(1.0f); + simd16scalar vRecipW2 = _simd16_set1_ps(1.0f); + + if (feState.vpTransformDisable) + { + // RHW is passed in directly when VP transform is disabled + vRecipW0 = tri[0].v[3]; + vRecipW1 = tri[1].v[3]; + vRecipW2 = tri[2].v[3]; + } + else { - for (uint32_t j = 0; j < 4; j += 1) + // Perspective divide + vRecipW0 = _simd16_div_ps(_simd16_set1_ps(1.0f), tri[0].w); + vRecipW1 = _simd16_div_ps(_simd16_set1_ps(1.0f), tri[1].w); + vRecipW2 = _simd16_div_ps(_simd16_set1_ps(1.0f), tri[2].w); + + tri[0].v[0] = _simd16_mul_ps(tri[0].v[0], vRecipW0); + tri[1].v[0] = _simd16_mul_ps(tri[1].v[0], vRecipW1); + tri[2].v[0] = _simd16_mul_ps(tri[2].v[0], vRecipW2); + + tri[0].v[1] = _simd16_mul_ps(tri[0].v[1], vRecipW0); + tri[1].v[1] = _simd16_mul_ps(tri[1].v[1], vRecipW1); + tri[2].v[1] = _simd16_mul_ps(tri[2].v[1], vRecipW2); + + tri[0].v[2] = _simd16_mul_ps(tri[0].v[2], vRecipW0); + tri[1].v[2] = _simd16_mul_ps(tri[1].v[2], vRecipW1); + tri[2].v[2] = _simd16_mul_ps(tri[2].v[2], vRecipW2); + + // Viewport transform to screen space coords + if (state.gsState.emitsViewportArrayIndex) { - verts[i][j] = _simd16_extract_ps(tri[i][j], 0); + viewportTransform<3>(tri, state.vpMatrices, viewportIdx); + } + else + { + viewportTransform<3>(tri, state.vpMatrices); } } - pa.useAlternateOffset = false; - BinTriangles<CT>(pDC, pa, workerId, verts, GetPrimMaskLo(triMask), _simd16_extract_si(primID, 0), _simd16_extract_si(viewportIdx, 0)); + // Adjust for pixel center location + const simd16scalar offset = g_pixelOffsets_simd16[rastState.pixelLocation]; + + tri[0].x = _simd16_add_ps(tri[0].x, offset); + tri[0].y = _simd16_add_ps(tri[0].y, offset); + + tri[1].x = _simd16_add_ps(tri[1].x, offset); + tri[1].y = _simd16_add_ps(tri[1].y, offset); + + tri[2].x = _simd16_add_ps(tri[2].x, offset); + tri[2].y = _simd16_add_ps(tri[2].y, offset); + + simd16scalari vXi[3], vYi[3]; + + // Set vXi, vYi to required fixed point precision + FPToFixedPoint(tri, vXi, vYi); + + // triangle setup + simd16scalari vAi[3], vBi[3]; + triangleSetupABIntVertical(vXi, vYi, vAi, vBi); + + // determinant + simd16scalari vDet[2]; + calcDeterminantIntVertical(vAi, vBi, vDet); + + // cull zero area + simd16mask maskLo = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpeq_epi64(vDet[0], _simd16_setzero_si()))); + simd16mask maskHi = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpeq_epi64(vDet[1], _simd16_setzero_si()))); + + uint32_t cullZeroAreaMask = maskLo | (maskHi << (KNOB_SIMD16_WIDTH / 2)); + + // don't cull degenerate triangles if we're conservatively rasterizing + uint32_t origTriMask = triMask; + if (rastState.fillMode == SWR_FILLMODE_SOLID && !CT::IsConservativeT::value) + { + triMask &= ~cullZeroAreaMask; + } + + // determine front winding tris + // CW +det + // CCW det < 0; + // 0 area triangles are marked as backfacing regardless of winding order, + // which is required behavior for conservative rast and wireframe rendering + uint32_t frontWindingTris; + if (rastState.frontWinding == SWR_FRONTWINDING_CW) + { + maskLo = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpgt_epi64(vDet[0], _simd16_setzero_si()))); + maskHi = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpgt_epi64(vDet[1], _simd16_setzero_si()))); + } + else + { + maskLo = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpgt_epi64(_simd16_setzero_si(), vDet[0]))); + maskHi = _simd16_movemask_pd(_simd16_castsi_pd(_simd16_cmpgt_epi64(_simd16_setzero_si(), vDet[1]))); + } + + frontWindingTris = maskLo | (maskHi << (KNOB_SIMD16_WIDTH / 2)); + + // cull + uint32_t cullTris; + switch ((SWR_CULLMODE)rastState.cullMode) + { + case SWR_CULLMODE_BOTH: cullTris = 0xffffffff; break; + case SWR_CULLMODE_NONE: cullTris = 0x0; break; + case SWR_CULLMODE_FRONT: cullTris = frontWindingTris; break; + // 0 area triangles are marked as backfacing, which is required behavior for conservative rast + case SWR_CULLMODE_BACK: cullTris = ~frontWindingTris; break; + default: SWR_INVALID("Invalid cull mode: %d", rastState.cullMode); cullTris = 0x0; break; + } + + triMask &= ~cullTris; + + if (origTriMask ^ triMask) + { + RDTSC_EVENT(FECullZeroAreaAndBackface, _mm_popcnt_u32(origTriMask ^ triMask), 0); + } - if (GetPrimMaskHi(triMask)) + // Simple non-conformant wireframe mode, useful for debugging + if (rastState.fillMode == SWR_FILLMODE_WIREFRAME) + { + // construct 3 SIMD lines out of the triangle and call the line binner for each SIMD + simd16vector line[2]; + simd16scalar recipW[2]; + line[0] = tri[0]; + line[1] = tri[1]; + recipW[0] = vRecipW0; + recipW[1] = vRecipW1; + BinPostSetupLines_simd16(pDC, pa, workerId, line, recipW, triMask, primID, viewportIdx); + + line[0] = tri[1]; + line[1] = tri[2]; + recipW[0] = vRecipW1; + recipW[1] = vRecipW2; + BinPostSetupLines_simd16(pDC, pa, workerId, line, recipW, triMask, primID, viewportIdx); + + line[0] = tri[2]; + line[1] = tri[0]; + recipW[0] = vRecipW2; + recipW[1] = vRecipW0; + BinPostSetupLines_simd16(pDC, pa, workerId, line, recipW, triMask, primID, viewportIdx); + + AR_END(FEBinTriangles, 1); + return; + } + + /// Note: these variable initializations must stay above any 'goto endBenTriangles' + // compute per tri backface + uint32_t frontFaceMask = frontWindingTris; + uint32_t *pPrimID = (uint32_t *)&primID; + const uint32_t *pViewportIndex = (uint32_t *)&viewportIdx; + DWORD triIndex = 0; + + uint32_t edgeEnable; + PFN_WORK_FUNC pfnWork; + if (CT::IsConservativeT::value) { - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) + // determine which edges of the degenerate tri, if any, are valid to rasterize. + // used to call the appropriate templated rasterizer function + if (cullZeroAreaMask > 0) + { + // e0 = v1-v0 + const simd16scalari x0x1Mask = _simd16_cmpeq_epi32(vXi[0], vXi[1]); + const simd16scalari y0y1Mask = _simd16_cmpeq_epi32(vYi[0], vYi[1]); + + uint32_t e0Mask = _simd16_movemask_ps(_simd16_castsi_ps(_simd16_and_si(x0x1Mask, y0y1Mask))); + + // e1 = v2-v1 + const simd16scalari x1x2Mask = _simd16_cmpeq_epi32(vXi[1], vXi[2]); + const simd16scalari y1y2Mask = _simd16_cmpeq_epi32(vYi[1], vYi[2]); + + uint32_t e1Mask = _simd16_movemask_ps(_simd16_castsi_ps(_simd16_and_si(x1x2Mask, y1y2Mask))); + + // e2 = v0-v2 + // if v0 == v1 & v1 == v2, v0 == v2 + uint32_t e2Mask = e0Mask & e1Mask; + SWR_ASSERT(KNOB_SIMD_WIDTH == 8, "Need to update degenerate mask code for avx512"); + + // edge order: e0 = v0v1, e1 = v1v2, e2 = v0v2 + // 32 bit binary: 0000 0000 0010 0100 1001 0010 0100 1001 + e0Mask = pdep_u32(e0Mask, 0x00249249); + + // 32 bit binary: 0000 0000 0100 1001 0010 0100 1001 0010 + e1Mask = pdep_u32(e1Mask, 0x00492492); + + // 32 bit binary: 0000 0000 1001 0010 0100 1001 0010 0100 + e2Mask = pdep_u32(e2Mask, 0x00924924); + + edgeEnable = (0x00FFFFFF & (~(e0Mask | e1Mask | e2Mask))); + } + else + { + edgeEnable = 0x00FFFFFF; + } + } + else + { + // degenerate triangles won't be sent to rasterizer; just enable all edges + pfnWork = GetRasterizerFunc(rastState.sampleCount, rastState.bIsCenterPattern, (rastState.conservativeRast > 0), + (SWR_INPUT_COVERAGE)pDC->pState->state.psState.inputCoverage, ALL_EDGES_VALID, (state.scissorsTileAligned == false)); + } + + if (!triMask) + { + goto endBinTriangles; + } + + // Calc bounding box of triangles + simd16BBox bbox; + calcBoundingBoxIntVertical<CT>(tri, vXi, vYi, bbox); + + // determine if triangle falls between pixel centers and discard + // only discard for non-MSAA case and when conservative rast is disabled + // (xmin + 127) & ~255 + // (xmax + 128) & ~255 + if ((rastState.sampleCount == SWR_MULTISAMPLE_1X || rastState.bIsCenterPattern) && + (!CT::IsConservativeT::value)) + { + origTriMask = triMask; + + int cullCenterMask; + + { + simd16scalari xmin = _simd16_add_epi32(bbox.xmin, _simd16_set1_epi32(127)); + xmin = _simd16_and_si(xmin, _simd16_set1_epi32(~255)); + simd16scalari xmax = _simd16_add_epi32(bbox.xmax, _simd16_set1_epi32(128)); + xmax = _simd16_and_si(xmax, _simd16_set1_epi32(~255)); + + simd16scalari vMaskH = _simd16_cmpeq_epi32(xmin, xmax); + + simd16scalari ymin = _simd16_add_epi32(bbox.ymin, _simd16_set1_epi32(127)); + ymin = _simd16_and_si(ymin, _simd16_set1_epi32(~255)); + simd16scalari ymax = _simd16_add_epi32(bbox.ymax, _simd16_set1_epi32(128)); + ymax = _simd16_and_si(ymax, _simd16_set1_epi32(~255)); + + simd16scalari vMaskV = _simd16_cmpeq_epi32(ymin, ymax); + + vMaskV = _simd16_or_si(vMaskH, vMaskV); + cullCenterMask = _simd16_movemask_ps(_simd16_castsi_ps(vMaskV)); + } + + triMask &= ~cullCenterMask; + + if (origTriMask ^ triMask) { - for (uint32_t j = 0; j < 4; j += 1) + RDTSC_EVENT(FECullBetweenCenters, _mm_popcnt_u32(origTriMask ^ triMask), 0); + } + } + + // Intersect with scissor/viewport. Subtract 1 ULP in x.8 fixed point since xmax/ymax edge is exclusive. + // Gather the AOS effective scissor rects based on the per-prim VP index. + /// @todo: Look at speeding this up -- weigh against corresponding costs in rasterizer. + simd16scalari scisXmin, scisYmin, scisXmax, scisYmax; + + if (state.gsState.emitsViewportArrayIndex) + { + GatherScissors_simd16<KNOB_SIMD16_WIDTH>::Gather(&state.scissorsInFixedPoint[0], pViewportIndex, + scisXmin, scisYmin, scisXmax, scisYmax); + } + else // broadcast fast path for non-VPAI case. + { + scisXmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmin); + scisYmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymin); + scisXmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmax); + scisYmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymax); + } + + bbox.xmin = _simd16_max_epi32(bbox.xmin, scisXmin); + bbox.ymin = _simd16_max_epi32(bbox.ymin, scisYmin); + bbox.xmax = _simd16_min_epi32(_simd16_sub_epi32(bbox.xmax, _simd16_set1_epi32(1)), scisXmax); + bbox.ymax = _simd16_min_epi32(_simd16_sub_epi32(bbox.ymax, _simd16_set1_epi32(1)), scisYmax); + + if (CT::IsConservativeT::value) + { + // in the case where a degenerate triangle is on a scissor edge, we need to make sure the primitive bbox has + // some area. Bump the xmax/ymax edges out + simd16scalari topEqualsBottom = _simd16_cmpeq_epi32(bbox.ymin, bbox.ymax); + bbox.ymax = _simd16_blendv_epi32(bbox.ymax, _simd16_add_epi32(bbox.ymax, _simd16_set1_epi32(1)), topEqualsBottom); + simd16scalari leftEqualsRight = _simd16_cmpeq_epi32(bbox.xmin, bbox.xmax); + bbox.xmax = _simd16_blendv_epi32(bbox.xmax, _simd16_add_epi32(bbox.xmax, _simd16_set1_epi32(1)), leftEqualsRight); + } + + // Cull tris completely outside scissor + { + simd16scalari maskOutsideScissorX = _simd16_cmpgt_epi32(bbox.xmin, bbox.xmax); + simd16scalari maskOutsideScissorY = _simd16_cmpgt_epi32(bbox.ymin, bbox.ymax); + simd16scalari maskOutsideScissorXY = _simd16_or_si(maskOutsideScissorX, maskOutsideScissorY); + uint32_t maskOutsideScissor = _simd16_movemask_ps(_simd16_castsi_ps(maskOutsideScissorXY)); + triMask = triMask & ~maskOutsideScissor; + } + + if (!triMask) + { + goto endBinTriangles; + } + + // Convert triangle bbox to macrotile units. + bbox.xmin = _simd16_srai_epi32(bbox.xmin, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymin = _simd16_srai_epi32(bbox.ymin, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + bbox.xmax = _simd16_srai_epi32(bbox.xmax, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymax = _simd16_srai_epi32(bbox.ymax, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + + OSALIGNSIMD16(uint32_t) aMTLeft[KNOB_SIMD16_WIDTH], aMTRight[KNOB_SIMD16_WIDTH], aMTTop[KNOB_SIMD16_WIDTH], aMTBottom[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTLeft), bbox.xmin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTRight), bbox.xmax); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTTop), bbox.ymin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTBottom), bbox.ymax); + + // transpose verts needed for backend + /// @todo modify BE to take non-transformed verts + __m128 vHorizX[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizY[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizZ[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizW[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + + vTranspose3x8(vHorizX[0], _simd16_extract_ps(tri[0].x, 0), _simd16_extract_ps(tri[1].x, 0), _simd16_extract_ps(tri[2].x, 0)); + vTranspose3x8(vHorizY[0], _simd16_extract_ps(tri[0].y, 0), _simd16_extract_ps(tri[1].y, 0), _simd16_extract_ps(tri[2].y, 0)); + vTranspose3x8(vHorizZ[0], _simd16_extract_ps(tri[0].z, 0), _simd16_extract_ps(tri[1].z, 0), _simd16_extract_ps(tri[2].z, 0)); + vTranspose3x8(vHorizW[0], _simd16_extract_ps(vRecipW0, 0), _simd16_extract_ps(vRecipW1, 0), _simd16_extract_ps(vRecipW2, 0)); + + vTranspose3x8(vHorizX[1], _simd16_extract_ps(tri[0].x, 1), _simd16_extract_ps(tri[1].x, 1), _simd16_extract_ps(tri[2].x, 1)); + vTranspose3x8(vHorizY[1], _simd16_extract_ps(tri[0].y, 1), _simd16_extract_ps(tri[1].y, 1), _simd16_extract_ps(tri[2].y, 1)); + vTranspose3x8(vHorizZ[1], _simd16_extract_ps(tri[0].z, 1), _simd16_extract_ps(tri[1].z, 1), _simd16_extract_ps(tri[2].z, 1)); + vTranspose3x8(vHorizW[1], _simd16_extract_ps(vRecipW0, 1), _simd16_extract_ps(vRecipW1, 1), _simd16_extract_ps(vRecipW2, 1)); + + // store render target array index + OSALIGNSIMD16(uint32_t) aRTAI[KNOB_SIMD16_WIDTH]; + if (gsState.gsEnable && gsState.emitsRenderTargetArrayIndex) + { + simd16vector vRtai[3]; + pa.Assemble_simd16(VERTEX_RTAI_SLOT, vRtai); + simd16scalari vRtaii; + vRtaii = _simd16_castps_si(vRtai[0].x); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), vRtaii); + } + else + { + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), _simd16_setzero_si()); + } + +endBinTriangles: + + + // scan remaining valid triangles and bin each separately + while (_BitScanForward(&triIndex, triMask)) + { + uint32_t linkageCount = state.backendState.numAttributes; + uint32_t numScalarAttribs = linkageCount * 4; + + BE_WORK work; + work.type = DRAW; + + bool isDegenerate; + if (CT::IsConservativeT::value) + { + // only rasterize valid edges if we have a degenerate primitive + int32_t triEdgeEnable = (edgeEnable >> (triIndex * 3)) & ALL_EDGES_VALID; + work.pfnWork = GetRasterizerFunc(rastState.sampleCount, rastState.bIsCenterPattern, (rastState.conservativeRast > 0), + (SWR_INPUT_COVERAGE)pDC->pState->state.psState.inputCoverage, triEdgeEnable, (state.scissorsTileAligned == false)); + + // Degenerate triangles are required to be constant interpolated + isDegenerate = (triEdgeEnable != ALL_EDGES_VALID) ? true : false; + } + else + { + isDegenerate = false; + work.pfnWork = pfnWork; + } + + // Select attribute processor + PFN_PROCESS_ATTRIBUTES pfnProcessAttribs = GetProcessAttributesFunc(3, + state.backendState.swizzleEnable, state.backendState.constantInterpolationMask, isDegenerate); + + TRIANGLE_WORK_DESC &desc = work.desc.tri; + + desc.triFlags.frontFacing = state.forceFront ? 1 : ((frontFaceMask >> triIndex) & 1); + desc.triFlags.primID = pPrimID[triIndex]; + desc.triFlags.renderTargetArrayIndex = aRTAI[triIndex]; + desc.triFlags.viewportIndex = pViewportIndex[triIndex]; + + auto pArena = pDC->pArena; + SWR_ASSERT(pArena != nullptr); + + // store active attribs + float *pAttribs = (float*)pArena->AllocAligned(numScalarAttribs * 3 * sizeof(float), 16); + desc.pAttribs = pAttribs; + desc.numAttribs = linkageCount; + pfnProcessAttribs(pDC, pa, triIndex, pPrimID[triIndex], desc.pAttribs); + + // store triangle vertex data + desc.pTriBuffer = (float*)pArena->AllocAligned(4 * 4 * sizeof(float), 16); + + { + const uint32_t i = triIndex >> 3; // triIndex / KNOB_SIMD_WIDTH + const uint32_t j = triIndex & 7; // triIndex % KNOB_SIMD_WIDTH + + _mm_store_ps(&desc.pTriBuffer[ 0], vHorizX[i][j]); + _mm_store_ps(&desc.pTriBuffer[ 4], vHorizY[i][j]); + _mm_store_ps(&desc.pTriBuffer[ 8], vHorizZ[i][j]); + _mm_store_ps(&desc.pTriBuffer[12], vHorizW[i][j]); + } + + // store user clip distances + if (rastState.clipDistanceMask) + { + uint32_t numClipDist = _mm_popcnt_u32(rastState.clipDistanceMask); + desc.pUserClipBuffer = (float*)pArena->Alloc(numClipDist * 3 * sizeof(float)); + ProcessUserClipDist<3>(pa, triIndex, rastState.clipDistanceMask, &desc.pTriBuffer[12], desc.pUserClipBuffer); + } + + for (uint32_t y = aMTTop[triIndex]; y <= aMTBottom[triIndex]; ++y) + { + for (uint32_t x = aMTLeft[triIndex]; x <= aMTRight[triIndex]; ++x) { - verts[i][j] = _simd16_extract_ps(tri[i][j], 1); +#if KNOB_ENABLE_TOSS_POINTS + if (!KNOB_TOSS_SETUP_TRIS) +#endif + { + pTileMgr->enqueue(x, y, &work); + } } } - pa.useAlternateOffset = true; - BinTriangles<CT>(pDC, pa, workerId, verts, GetPrimMaskHi(triMask), _simd16_extract_si(primID, 1), _simd16_extract_si(viewportIdx, 1)); + triMask &= ~(1 << triIndex); } + + AR_END(FEBinTriangles, 1); } #endif @@ -1294,34 +1858,341 @@ void BinPoints_simd16( simd16scalari primID, simd16scalari viewportIdx) { - enum { VERTS_PER_PRIM = 1 }; + SWR_CONTEXT *pContext = pDC->pContext; - simdvector verts[VERTS_PER_PRIM]; + AR_BEGIN(FEBinPoints, pDC->drawId); + + simd16vector& primVerts = prim[0]; + + const API_STATE& state = GetApiState(pDC); + const SWR_FRONTEND_STATE& feState = state.frontendState; + const SWR_GS_STATE& gsState = state.gsState; + const SWR_RASTSTATE& rastState = state.rastState; + const uint32_t *pViewportIndex = (uint32_t *)&viewportIdx; + + // Select attribute processor + PFN_PROCESS_ATTRIBUTES pfnProcessAttribs = GetProcessAttributesFunc(1, + state.backendState.swizzleEnable, state.backendState.constantInterpolationMask); - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) + if (!feState.vpTransformDisable) { - for (uint32_t j = 0; j < 4; j += 1) + // perspective divide + simd16scalar vRecipW0 = _simd16_div_ps(_simd16_set1_ps(1.0f), primVerts.w); + + primVerts.x = _simd16_mul_ps(primVerts.x, vRecipW0); + primVerts.y = _simd16_mul_ps(primVerts.y, vRecipW0); + primVerts.z = _simd16_mul_ps(primVerts.z, vRecipW0); + + // viewport transform to screen coords + if (state.gsState.emitsViewportArrayIndex) { - verts[i][j] = _simd16_extract_ps(prim[i][j], 0); + viewportTransform<1>(&primVerts, state.vpMatrices, viewportIdx); + } + else + { + viewportTransform<1>(&primVerts, state.vpMatrices); } } - pa.useAlternateOffset = false; - BinPoints(pDC, pa, workerId, verts, GetPrimMaskLo(primMask), _simd16_extract_si(primID, 0), _simd16_extract_si(viewportIdx, 0)); + const simd16scalar offset = g_pixelOffsets_simd16[rastState.pixelLocation]; + + primVerts.x = _simd16_add_ps(primVerts.x, offset); + primVerts.y = _simd16_add_ps(primVerts.y, offset); + + // convert to fixed point + simd16scalari vXi, vYi; + + vXi = fpToFixedPointVertical(primVerts.x); + vYi = fpToFixedPointVertical(primVerts.y); - if (GetPrimMaskHi(primMask)) + if (CanUseSimplePoints(pDC)) { - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) + // adjust for ymin-xmin rule + vXi = _simd16_sub_epi32(vXi, _simd16_set1_epi32(1)); + vYi = _simd16_sub_epi32(vYi, _simd16_set1_epi32(1)); + + // cull points off the ymin-xmin edge of the viewport + primMask &= ~_simd16_movemask_ps(_simd16_castsi_ps(vXi)); + primMask &= ~_simd16_movemask_ps(_simd16_castsi_ps(vYi)); + + // compute macro tile coordinates + simd16scalari macroX = _simd16_srai_epi32(vXi, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + simd16scalari macroY = _simd16_srai_epi32(vYi, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + + OSALIGNSIMD16(uint32_t) aMacroX[KNOB_SIMD16_WIDTH], aMacroY[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMacroX), macroX); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMacroY), macroY); + + // compute raster tile coordinates + simd16scalari rasterX = _simd16_srai_epi32(vXi, KNOB_TILE_X_DIM_SHIFT + FIXED_POINT_SHIFT); + simd16scalari rasterY = _simd16_srai_epi32(vYi, KNOB_TILE_Y_DIM_SHIFT + FIXED_POINT_SHIFT); + + // compute raster tile relative x,y for coverage mask + simd16scalari tileAlignedX = _simd16_slli_epi32(rasterX, KNOB_TILE_X_DIM_SHIFT); + simd16scalari tileAlignedY = _simd16_slli_epi32(rasterY, KNOB_TILE_Y_DIM_SHIFT); + + simd16scalari tileRelativeX = _simd16_sub_epi32(_simd16_srai_epi32(vXi, FIXED_POINT_SHIFT), tileAlignedX); + simd16scalari tileRelativeY = _simd16_sub_epi32(_simd16_srai_epi32(vYi, FIXED_POINT_SHIFT), tileAlignedY); + + OSALIGNSIMD16(uint32_t) aTileRelativeX[KNOB_SIMD16_WIDTH]; + OSALIGNSIMD16(uint32_t) aTileRelativeY[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aTileRelativeX), tileRelativeX); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aTileRelativeY), tileRelativeY); + + OSALIGNSIMD16(uint32_t) aTileAlignedX[KNOB_SIMD16_WIDTH]; + OSALIGNSIMD16(uint32_t) aTileAlignedY[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aTileAlignedX), tileAlignedX); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aTileAlignedY), tileAlignedY); + + OSALIGNSIMD16(float) aZ[KNOB_SIMD16_WIDTH]; + _simd16_store_ps(reinterpret_cast<float *>(aZ), primVerts.z); + + // store render target array index + OSALIGNSIMD16(uint32_t) aRTAI[KNOB_SIMD16_WIDTH]; + if (gsState.gsEnable && gsState.emitsRenderTargetArrayIndex) + { + simd16vector vRtai; + pa.Assemble_simd16(VERTEX_RTAI_SLOT, &vRtai); + simd16scalari vRtaii = _simd16_castps_si(vRtai.x); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), vRtaii); + } + else + { + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), _simd16_setzero_si()); + } + + uint32_t *pPrimID = (uint32_t *)&primID; + DWORD primIndex = 0; + + const SWR_BACKEND_STATE& backendState = pDC->pState->state.backendState; + + // scan remaining valid triangles and bin each separately + while (_BitScanForward(&primIndex, primMask)) { - for (uint32_t j = 0; j < 4; j += 1) + uint32_t linkageCount = backendState.numAttributes; + uint32_t numScalarAttribs = linkageCount * 4; + + BE_WORK work; + work.type = DRAW; + + TRIANGLE_WORK_DESC &desc = work.desc.tri; + + // points are always front facing + desc.triFlags.frontFacing = 1; + desc.triFlags.primID = pPrimID[primIndex]; + desc.triFlags.renderTargetArrayIndex = aRTAI[primIndex]; + desc.triFlags.viewportIndex = pViewportIndex[primIndex]; + + work.pfnWork = RasterizeSimplePoint; + + auto pArena = pDC->pArena; + SWR_ASSERT(pArena != nullptr); + + // store attributes + float *pAttribs = (float*)pArena->AllocAligned(3 * numScalarAttribs * sizeof(float), 16); + desc.pAttribs = pAttribs; + desc.numAttribs = linkageCount; + + pfnProcessAttribs(pDC, pa, primIndex, pPrimID[primIndex], pAttribs); + + // store raster tile aligned x, y, perspective correct z + float *pTriBuffer = (float*)pArena->AllocAligned(4 * sizeof(float), 16); + desc.pTriBuffer = pTriBuffer; + *(uint32_t*)pTriBuffer++ = aTileAlignedX[primIndex]; + *(uint32_t*)pTriBuffer++ = aTileAlignedY[primIndex]; + *pTriBuffer = aZ[primIndex]; + + uint32_t tX = aTileRelativeX[primIndex]; + uint32_t tY = aTileRelativeY[primIndex]; + + // pack the relative x,y into the coverageMask, the rasterizer will + // generate the true coverage mask from it + work.desc.tri.triFlags.coverageMask = tX | (tY << 4); + + // bin it + MacroTileMgr *pTileMgr = pDC->pTileMgr; +#if KNOB_ENABLE_TOSS_POINTS + if (!KNOB_TOSS_SETUP_TRIS) +#endif { - verts[i][j] = _simd16_extract_ps(prim[i][j], 1); + pTileMgr->enqueue(aMacroX[primIndex], aMacroY[primIndex], &work); } + + primMask &= ~(1 << primIndex); } + } + else + { + // non simple points need to be potentially binned to multiple macro tiles + simd16scalar vPointSize; + + if (rastState.pointParam) + { + simd16vector size[3]; + pa.Assemble_simd16(VERTEX_POINT_SIZE_SLOT, size); + vPointSize = size[0].x; + } + else + { + vPointSize = _simd16_set1_ps(rastState.pointSize); + } + + // bloat point to bbox + simd16BBox bbox; + + bbox.xmin = bbox.xmax = vXi; + bbox.ymin = bbox.ymax = vYi; + + simd16scalar vHalfWidth = _simd16_mul_ps(vPointSize, _simd16_set1_ps(0.5f)); + simd16scalari vHalfWidthi = fpToFixedPointVertical(vHalfWidth); + + bbox.xmin = _simd16_sub_epi32(bbox.xmin, vHalfWidthi); + bbox.xmax = _simd16_add_epi32(bbox.xmax, vHalfWidthi); + bbox.ymin = _simd16_sub_epi32(bbox.ymin, vHalfWidthi); + bbox.ymax = _simd16_add_epi32(bbox.ymax, vHalfWidthi); + + // Intersect with scissor/viewport. Subtract 1 ULP in x.8 fixed point since xmax/ymax edge is exclusive. + // Gather the AOS effective scissor rects based on the per-prim VP index. + /// @todo: Look at speeding this up -- weigh against corresponding costs in rasterizer. + simd16scalari scisXmin, scisYmin, scisXmax, scisYmax; + if (state.gsState.emitsViewportArrayIndex) + { + GatherScissors_simd16<KNOB_SIMD16_WIDTH>::Gather(&state.scissorsInFixedPoint[0], pViewportIndex, + scisXmin, scisYmin, scisXmax, scisYmax); + } + else // broadcast fast path for non-VPAI case. + { + scisXmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmin); + scisYmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymin); + scisXmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmax); + scisYmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymax); + } + + bbox.xmin = _simd16_max_epi32(bbox.xmin, scisXmin); + bbox.ymin = _simd16_max_epi32(bbox.ymin, scisYmin); + bbox.xmax = _simd16_min_epi32(_simd16_sub_epi32(bbox.xmax, _simd16_set1_epi32(1)), scisXmax); + bbox.ymax = _simd16_min_epi32(_simd16_sub_epi32(bbox.ymax, _simd16_set1_epi32(1)), scisYmax); + + // Cull bloated points completely outside scissor + simd16scalari maskOutsideScissorX = _simd16_cmpgt_epi32(bbox.xmin, bbox.xmax); + simd16scalari maskOutsideScissorY = _simd16_cmpgt_epi32(bbox.ymin, bbox.ymax); + simd16scalari maskOutsideScissorXY = _simd16_or_si(maskOutsideScissorX, maskOutsideScissorY); + uint32_t maskOutsideScissor = _simd16_movemask_ps(_simd16_castsi_ps(maskOutsideScissorXY)); + primMask = primMask & ~maskOutsideScissor; + + // Convert bbox to macrotile units. + bbox.xmin = _simd16_srai_epi32(bbox.xmin, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymin = _simd16_srai_epi32(bbox.ymin, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + bbox.xmax = _simd16_srai_epi32(bbox.xmax, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymax = _simd16_srai_epi32(bbox.ymax, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + + OSALIGNSIMD16(uint32_t) aMTLeft[KNOB_SIMD16_WIDTH], aMTRight[KNOB_SIMD16_WIDTH], aMTTop[KNOB_SIMD16_WIDTH], aMTBottom[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTLeft), bbox.xmin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTRight), bbox.xmax); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTTop), bbox.ymin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTBottom), bbox.ymax); + + // store render target array index + OSALIGNSIMD16(uint32_t) aRTAI[KNOB_SIMD16_WIDTH]; + if (gsState.gsEnable && gsState.emitsRenderTargetArrayIndex) + { + simd16vector vRtai[2]; + pa.Assemble_simd16(VERTEX_RTAI_SLOT, vRtai); + simd16scalari vRtaii = _simd16_castps_si(vRtai[0].x); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), vRtaii); + } + else + { + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), _simd16_setzero_si()); + } + + OSALIGNSIMD16(float) aPointSize[KNOB_SIMD16_WIDTH]; + _simd16_store_ps(reinterpret_cast<float *>(aPointSize), vPointSize); + + uint32_t *pPrimID = (uint32_t *)&primID; + + OSALIGNSIMD16(float) aPrimVertsX[KNOB_SIMD16_WIDTH]; + OSALIGNSIMD16(float) aPrimVertsY[KNOB_SIMD16_WIDTH]; + OSALIGNSIMD16(float) aPrimVertsZ[KNOB_SIMD16_WIDTH]; + + _simd16_store_ps(reinterpret_cast<float *>(aPrimVertsX), primVerts.x); + _simd16_store_ps(reinterpret_cast<float *>(aPrimVertsY), primVerts.y); + _simd16_store_ps(reinterpret_cast<float *>(aPrimVertsZ), primVerts.z); + + // scan remaining valid prims and bin each separately + const SWR_BACKEND_STATE& backendState = state.backendState; + DWORD primIndex; + while (_BitScanForward(&primIndex, primMask)) + { + uint32_t linkageCount = backendState.numAttributes; + uint32_t numScalarAttribs = linkageCount * 4; + + BE_WORK work; + work.type = DRAW; + + TRIANGLE_WORK_DESC &desc = work.desc.tri; + + desc.triFlags.frontFacing = 1; + desc.triFlags.primID = pPrimID[primIndex]; + desc.triFlags.pointSize = aPointSize[primIndex]; + desc.triFlags.renderTargetArrayIndex = aRTAI[primIndex]; + desc.triFlags.viewportIndex = pViewportIndex[primIndex]; + + work.pfnWork = RasterizeTriPoint; + + auto pArena = pDC->pArena; + SWR_ASSERT(pArena != nullptr); + + // store active attribs + desc.pAttribs = (float*)pArena->AllocAligned(numScalarAttribs * 3 * sizeof(float), 16); + desc.numAttribs = linkageCount; + pfnProcessAttribs(pDC, pa, primIndex, pPrimID[primIndex], desc.pAttribs); + + // store point vertex data + float *pTriBuffer = (float*)pArena->AllocAligned(4 * sizeof(float), 16); + desc.pTriBuffer = pTriBuffer; + *pTriBuffer++ = aPrimVertsX[primIndex]; + *pTriBuffer++ = aPrimVertsY[primIndex]; + *pTriBuffer = aPrimVertsZ[primIndex]; + + // store user clip distances + if (rastState.clipDistanceMask) + { + uint32_t numClipDist = _mm_popcnt_u32(rastState.clipDistanceMask); + desc.pUserClipBuffer = (float*)pArena->Alloc(numClipDist * 3 * sizeof(float)); + float dists[8]; + float one = 1.0f; + ProcessUserClipDist<1>(pa, primIndex, rastState.clipDistanceMask, &one, dists); + for (uint32_t i = 0; i < numClipDist; i++) { + desc.pUserClipBuffer[3 * i + 0] = 0.0f; + desc.pUserClipBuffer[3 * i + 1] = 0.0f; + desc.pUserClipBuffer[3 * i + 2] = dists[i]; + } + } + + MacroTileMgr *pTileMgr = pDC->pTileMgr; + for (uint32_t y = aMTTop[primIndex]; y <= aMTBottom[primIndex]; ++y) + { + for (uint32_t x = aMTLeft[primIndex]; x <= aMTRight[primIndex]; ++x) + { +#if KNOB_ENABLE_TOSS_POINTS + if (!KNOB_TOSS_SETUP_TRIS) +#endif + { + pTileMgr->enqueue(x, y, &work); + } + } + } - pa.useAlternateOffset = true; - BinPoints(pDC, pa, workerId, verts, GetPrimMaskHi(primMask), _simd16_extract_si(primID, 1), _simd16_extract_si(viewportIdx, 1)); + primMask &= ~(1 << primIndex); + } } + + AR_END(FEBinPoints, 1); } #endif @@ -1537,6 +2408,232 @@ endBinLines: AR_END(FEBinLines, 1); } +#if USE_SIMD16_FRONTEND +void BinPostSetupLines_simd16( + DRAW_CONTEXT *pDC, + PA_STATE& pa, + uint32_t workerId, + simd16vector prim[], + simd16scalar recipW[], + uint32_t primMask, + simd16scalari primID, + simd16scalari viewportIdx) +{ + SWR_CONTEXT *pContext = pDC->pContext; + + AR_BEGIN(FEBinLines, pDC->drawId); + + const API_STATE& state = GetApiState(pDC); + const SWR_RASTSTATE& rastState = state.rastState; + const SWR_FRONTEND_STATE& feState = state.frontendState; + const SWR_GS_STATE& gsState = state.gsState; + + // Select attribute processor + PFN_PROCESS_ATTRIBUTES pfnProcessAttribs = GetProcessAttributesFunc(2, + state.backendState.swizzleEnable, state.backendState.constantInterpolationMask); + + simd16scalar& vRecipW0 = recipW[0]; + simd16scalar& vRecipW1 = recipW[1]; + + // convert to fixed point + simd16scalari vXi[2], vYi[2]; + + vXi[0] = fpToFixedPointVertical(prim[0].x); + vYi[0] = fpToFixedPointVertical(prim[0].y); + vXi[1] = fpToFixedPointVertical(prim[1].x); + vYi[1] = fpToFixedPointVertical(prim[1].y); + + // compute x-major vs y-major mask + simd16scalari xLength = _simd16_abs_epi32(_simd16_sub_epi32(vXi[0], vXi[1])); + simd16scalari yLength = _simd16_abs_epi32(_simd16_sub_epi32(vYi[0], vYi[1])); + simd16scalar vYmajorMask = _simd16_castsi_ps(_simd16_cmpgt_epi32(yLength, xLength)); + uint32_t yMajorMask = _simd16_movemask_ps(vYmajorMask); + + // cull zero-length lines + simd16scalari vZeroLengthMask = _simd16_cmpeq_epi32(xLength, _simd16_setzero_si()); + vZeroLengthMask = _simd16_and_si(vZeroLengthMask, _simd16_cmpeq_epi32(yLength, _simd16_setzero_si())); + + primMask &= ~_simd16_movemask_ps(_simd16_castsi_ps(vZeroLengthMask)); + + uint32_t *pPrimID = (uint32_t *)&primID; + const uint32_t *pViewportIndex = (uint32_t *)&viewportIdx; + + // Calc bounding box of lines + simd16BBox bbox; + bbox.xmin = _simd16_min_epi32(vXi[0], vXi[1]); + bbox.xmax = _simd16_max_epi32(vXi[0], vXi[1]); + bbox.ymin = _simd16_min_epi32(vYi[0], vYi[1]); + bbox.ymax = _simd16_max_epi32(vYi[0], vYi[1]); + + // bloat bbox by line width along minor axis + simd16scalar vHalfWidth = _simd16_set1_ps(rastState.lineWidth / 2.0f); + simd16scalari vHalfWidthi = fpToFixedPointVertical(vHalfWidth); + + simd16BBox bloatBox; + + bloatBox.xmin = _simd16_sub_epi32(bbox.xmin, vHalfWidthi); + bloatBox.xmax = _simd16_add_epi32(bbox.xmax, vHalfWidthi); + bloatBox.ymin = _simd16_sub_epi32(bbox.ymin, vHalfWidthi); + bloatBox.ymax = _simd16_add_epi32(bbox.ymax, vHalfWidthi); + + bbox.xmin = _simd16_blendv_epi32(bbox.xmin, bloatBox.xmin, vYmajorMask); + bbox.xmax = _simd16_blendv_epi32(bbox.xmax, bloatBox.xmax, vYmajorMask); + bbox.ymin = _simd16_blendv_epi32(bloatBox.ymin, bbox.ymin, vYmajorMask); + bbox.ymax = _simd16_blendv_epi32(bloatBox.ymax, bbox.ymax, vYmajorMask); + + // Intersect with scissor/viewport. Subtract 1 ULP in x.8 fixed point since xmax/ymax edge is exclusive. + simd16scalari scisXmin, scisYmin, scisXmax, scisYmax; + + if (state.gsState.emitsViewportArrayIndex) + { + GatherScissors_simd16<KNOB_SIMD16_WIDTH>::Gather(&state.scissorsInFixedPoint[0], pViewportIndex, + scisXmin, scisYmin, scisXmax, scisYmax); + } + else // broadcast fast path for non-VPAI case. + { + scisXmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmin); + scisYmin = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymin); + scisXmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].xmax); + scisYmax = _simd16_set1_epi32(state.scissorsInFixedPoint[0].ymax); + } + + bbox.xmin = _simd16_max_epi32(bbox.xmin, scisXmin); + bbox.ymin = _simd16_max_epi32(bbox.ymin, scisYmin); + bbox.xmax = _simd16_min_epi32(_simd16_sub_epi32(bbox.xmax, _simd16_set1_epi32(1)), scisXmax); + bbox.ymax = _simd16_min_epi32(_simd16_sub_epi32(bbox.ymax, _simd16_set1_epi32(1)), scisYmax); + + // Cull prims completely outside scissor + { + simd16scalari maskOutsideScissorX = _simd16_cmpgt_epi32(bbox.xmin, bbox.xmax); + simd16scalari maskOutsideScissorY = _simd16_cmpgt_epi32(bbox.ymin, bbox.ymax); + simd16scalari maskOutsideScissorXY = _simd16_or_si(maskOutsideScissorX, maskOutsideScissorY); + uint32_t maskOutsideScissor = _simd16_movemask_ps(_simd16_castsi_ps(maskOutsideScissorXY)); + primMask = primMask & ~maskOutsideScissor; + } + + if (!primMask) + { + goto endBinLines; + } + + // Convert triangle bbox to macrotile units. + bbox.xmin = _simd16_srai_epi32(bbox.xmin, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymin = _simd16_srai_epi32(bbox.ymin, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + bbox.xmax = _simd16_srai_epi32(bbox.xmax, KNOB_MACROTILE_X_DIM_FIXED_SHIFT); + bbox.ymax = _simd16_srai_epi32(bbox.ymax, KNOB_MACROTILE_Y_DIM_FIXED_SHIFT); + + OSALIGNSIMD16(uint32_t) aMTLeft[KNOB_SIMD16_WIDTH], aMTRight[KNOB_SIMD16_WIDTH], aMTTop[KNOB_SIMD16_WIDTH], aMTBottom[KNOB_SIMD16_WIDTH]; + + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTLeft), bbox.xmin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTRight), bbox.xmax); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTTop), bbox.ymin); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aMTBottom), bbox.ymax); + + // transpose verts needed for backend + /// @todo modify BE to take non-transformed verts + __m128 vHorizX[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizY[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizZ[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + __m128 vHorizW[2][KNOB_SIMD_WIDTH]; // KNOB_SIMD16_WIDTH + + const simdscalar unused = _simd_setzero_ps(); + + vTranspose3x8(vHorizX[0], _simd16_extract_ps(prim[0].x, 0), _simd16_extract_ps(prim[1].x, 0), unused); + vTranspose3x8(vHorizY[0], _simd16_extract_ps(prim[0].y, 0), _simd16_extract_ps(prim[1].y, 0), unused); + vTranspose3x8(vHorizZ[0], _simd16_extract_ps(prim[0].z, 0), _simd16_extract_ps(prim[1].z, 0), unused); + vTranspose3x8(vHorizW[0], _simd16_extract_ps(vRecipW0, 0), _simd16_extract_ps(vRecipW1, 0), unused); + + vTranspose3x8(vHorizX[1], _simd16_extract_ps(prim[0].x, 1), _simd16_extract_ps(prim[1].x, 1), unused); + vTranspose3x8(vHorizY[1], _simd16_extract_ps(prim[0].y, 1), _simd16_extract_ps(prim[1].y, 1), unused); + vTranspose3x8(vHorizZ[1], _simd16_extract_ps(prim[0].z, 1), _simd16_extract_ps(prim[1].z, 1), unused); + vTranspose3x8(vHorizW[1], _simd16_extract_ps(vRecipW0, 1), _simd16_extract_ps(vRecipW1, 1), unused); + + // store render target array index + OSALIGNSIMD16(uint32_t) aRTAI[KNOB_SIMD16_WIDTH]; + if (gsState.gsEnable && gsState.emitsRenderTargetArrayIndex) + { + simd16vector vRtai[2]; + pa.Assemble_simd16(VERTEX_RTAI_SLOT, vRtai); + simd16scalari vRtaii = _simd16_castps_si(vRtai[0].x); + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), vRtaii); + } + else + { + _simd16_store_si(reinterpret_cast<simd16scalari *>(aRTAI), _simd16_setzero_si()); + } + + // scan remaining valid prims and bin each separately + DWORD primIndex; + while (_BitScanForward(&primIndex, primMask)) + { + uint32_t linkageCount = state.backendState.numAttributes; + uint32_t numScalarAttribs = linkageCount * 4; + + BE_WORK work; + work.type = DRAW; + + TRIANGLE_WORK_DESC &desc = work.desc.tri; + + desc.triFlags.frontFacing = 1; + desc.triFlags.primID = pPrimID[primIndex]; + desc.triFlags.yMajor = (yMajorMask >> primIndex) & 1; + desc.triFlags.renderTargetArrayIndex = aRTAI[primIndex]; + desc.triFlags.viewportIndex = pViewportIndex[primIndex]; + + work.pfnWork = RasterizeLine; + + auto pArena = pDC->pArena; + SWR_ASSERT(pArena != nullptr); + + // store active attribs + desc.pAttribs = (float*)pArena->AllocAligned(numScalarAttribs * 3 * sizeof(float), 16); + desc.numAttribs = linkageCount; + pfnProcessAttribs(pDC, pa, primIndex, pPrimID[primIndex], desc.pAttribs); + + // store line vertex data + desc.pTriBuffer = (float*)pArena->AllocAligned(4 * 4 * sizeof(float), 16); + + { + const uint32_t i = primIndex >> 3; // triIndex / KNOB_SIMD_WIDTH + const uint32_t j = primIndex & 7; // triIndex % KNOB_SIMD_WIDTH + + _mm_store_ps(&desc.pTriBuffer[ 0], vHorizX[i][j]); + _mm_store_ps(&desc.pTriBuffer[ 4], vHorizY[i][j]); + _mm_store_ps(&desc.pTriBuffer[ 8], vHorizZ[i][j]); + _mm_store_ps(&desc.pTriBuffer[12], vHorizW[i][j]); + } + + // store user clip distances + if (rastState.clipDistanceMask) + { + uint32_t numClipDist = _mm_popcnt_u32(rastState.clipDistanceMask); + desc.pUserClipBuffer = (float*)pArena->Alloc(numClipDist * 2 * sizeof(float)); + ProcessUserClipDist<2>(pa, primIndex, rastState.clipDistanceMask, &desc.pTriBuffer[12], desc.pUserClipBuffer); + } + + MacroTileMgr *pTileMgr = pDC->pTileMgr; + for (uint32_t y = aMTTop[primIndex]; y <= aMTBottom[primIndex]; ++y) + { + for (uint32_t x = aMTLeft[primIndex]; x <= aMTRight[primIndex]; ++x) + { +#if KNOB_ENABLE_TOSS_POINTS + if (!KNOB_TOSS_SETUP_TRIS) +#endif + { + pTileMgr->enqueue(x, y, &work); + } + } + } + + primMask &= ~(1 << primIndex); + } + +endBinLines: + + AR_END(FEBinLines, 1); +} + +#endif ////////////////////////////////////////////////////////////////////////// /// @brief Bin SIMD lines to the backend. /// @param pDC - pointer to draw context. @@ -1622,34 +2719,63 @@ void BinLines_simd16( simd16scalari primID, simd16scalari viewportIdx) { - enum { VERTS_PER_PRIM = 2 }; + SWR_CONTEXT *pContext = pDC->pContext; - simdvector verts[VERTS_PER_PRIM]; + const API_STATE& state = GetApiState(pDC); + const SWR_RASTSTATE& rastState = state.rastState; + const SWR_FRONTEND_STATE& feState = state.frontendState; + const SWR_GS_STATE& gsState = state.gsState; - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) - { - for (uint32_t j = 0; j < 4; j += 1) - { - verts[i][j] = _simd16_extract_ps(prim[i][j], 0); - } - } + // Select attribute processor + PFN_PROCESS_ATTRIBUTES pfnProcessAttribs = GetProcessAttributesFunc(2, + state.backendState.swizzleEnable, state.backendState.constantInterpolationMask); - pa.useAlternateOffset = false; - BinLines(pDC, pa, workerId, verts, GetPrimMaskLo(primMask), _simd16_extract_si(primID, 0), _simd16_extract_si(viewportIdx, 0)); + simd16scalar vRecipW[2] = { _simd16_set1_ps(1.0f), _simd16_set1_ps(1.0f) }; - if (GetPrimMaskHi(primMask)) + if (!feState.vpTransformDisable) { - for (uint32_t i = 0; i < VERTS_PER_PRIM; i += 1) + // perspective divide + vRecipW[0] = _simd16_div_ps(_simd16_set1_ps(1.0f), prim[0].w); + vRecipW[1] = _simd16_div_ps(_simd16_set1_ps(1.0f), prim[1].w); + + prim[0].v[0] = _simd16_mul_ps(prim[0].v[0], vRecipW[0]); + prim[1].v[0] = _simd16_mul_ps(prim[1].v[0], vRecipW[1]); + + prim[0].v[1] = _simd16_mul_ps(prim[0].v[1], vRecipW[0]); + prim[1].v[1] = _simd16_mul_ps(prim[1].v[1], vRecipW[1]); + + prim[0].v[2] = _simd16_mul_ps(prim[0].v[2], vRecipW[0]); + prim[1].v[2] = _simd16_mul_ps(prim[1].v[2], vRecipW[1]); + + // viewport transform to screen coords + if (state.gsState.emitsViewportArrayIndex) { - for (uint32_t j = 0; j < 4; j += 1) - { - verts[i][j] = _simd16_extract_ps(prim[i][j], 1); - } + viewportTransform<2>(prim, state.vpMatrices, viewportIdx); } + else + { + viewportTransform<2>(prim, state.vpMatrices); + } +} - pa.useAlternateOffset = true; - BinLines(pDC, pa, workerId, verts, GetPrimMaskHi(primMask), _simd16_extract_si(primID, 1), _simd16_extract_si(viewportIdx, 1)); - } + // adjust for pixel center location + simd16scalar offset = g_pixelOffsets_simd16[rastState.pixelLocation]; + + prim[0].x = _simd16_add_ps(prim[0].x, offset); + prim[0].y = _simd16_add_ps(prim[0].y, offset); + + prim[1].x = _simd16_add_ps(prim[1].x, offset); + prim[1].y = _simd16_add_ps(prim[1].y, offset); + + BinPostSetupLines_simd16( + pDC, + pa, + workerId, + prim, + vRecipW, + primMask, + primID, + viewportIdx); } #endif diff --git a/src/gallium/drivers/swr/rasterizer/core/frontend.h b/src/gallium/drivers/swr/rasterizer/core/frontend.h index 37b7215c516..e880ead71dc 100644 --- a/src/gallium/drivers/swr/rasterizer/core/frontend.h +++ b/src/gallium/drivers/swr/rasterizer/core/frontend.h @@ -112,6 +112,23 @@ void triangleSetupABIntVertical(const simdscalari vX[3], const simdscalari vY[3] vB[1] = _simd_sub_epi32(vX[2], vX[1]); vB[2] = _simd_sub_epi32(vX[0], vX[2]); } + +#if ENABLE_AVX512_SIMD16 +INLINE +void triangleSetupABIntVertical(const simd16scalari vX[3], const simd16scalari vY[3], simd16scalari(&vA)[3], simd16scalari(&vB)[3]) +{ + // A = y0 - y1 + // B = x1 - x0 + vA[0] = _simd16_sub_epi32(vY[0], vY[1]); + vA[1] = _simd16_sub_epi32(vY[1], vY[2]); + vA[2] = _simd16_sub_epi32(vY[2], vY[0]); + + vB[0] = _simd16_sub_epi32(vX[1], vX[0]); + vB[1] = _simd16_sub_epi32(vX[2], vX[1]); + vB[2] = _simd16_sub_epi32(vX[0], vX[2]); +} + +#endif // Calculate the determinant of the triangle // 2 vectors between the 3 points: P, Q // Px = x0-x2, Py = y0-y2 @@ -185,6 +202,44 @@ void calcDeterminantIntVertical(const simdscalari vA[3], const simdscalari vB[3] pvDet[1] = vResultHi; } +#if ENABLE_AVX512_SIMD16 +INLINE +void calcDeterminantIntVertical(const simd16scalari vA[3], const simd16scalari vB[3], simd16scalari *pvDet) +{ + // refer to calcDeterminantInt comment for calculation explanation + // A1*B2 + simd16scalari vA1Lo = _simd16_unpacklo_epi32(vA[1], vA[1]); // 0 0 1 1 4 4 5 5 + simd16scalari vA1Hi = _simd16_unpackhi_epi32(vA[1], vA[1]); // 2 2 3 3 6 6 7 7 + + simd16scalari vB2Lo = _simd16_unpacklo_epi32(vB[2], vB[2]); + simd16scalari vB2Hi = _simd16_unpackhi_epi32(vB[2], vB[2]); + + simd16scalari vA1B2Lo = _simd16_mul_epi32(vA1Lo, vB2Lo); // 0 1 4 5 + simd16scalari vA1B2Hi = _simd16_mul_epi32(vA1Hi, vB2Hi); // 2 3 6 7 + + // B1*A2 + simd16scalari vA2Lo = _simd16_unpacklo_epi32(vA[2], vA[2]); + simd16scalari vA2Hi = _simd16_unpackhi_epi32(vA[2], vA[2]); + + simd16scalari vB1Lo = _simd16_unpacklo_epi32(vB[1], vB[1]); + simd16scalari vB1Hi = _simd16_unpackhi_epi32(vB[1], vB[1]); + + simd16scalari vA2B1Lo = _simd16_mul_epi32(vA2Lo, vB1Lo); + simd16scalari vA2B1Hi = _simd16_mul_epi32(vA2Hi, vB1Hi); + + // A1*B2 - A2*B1 + simd16scalari detLo = _simd16_sub_epi64(vA1B2Lo, vA2B1Lo); + simd16scalari detHi = _simd16_sub_epi64(vA1B2Hi, vA2B1Hi); + + // shuffle 0 1 4 5 -> 0 1 2 3 + simd16scalari vResultLo = _simd16_permute2f128_si(detLo, detHi, 0x20); + simd16scalari vResultHi = _simd16_permute2f128_si(detLo, detHi, 0x31); + + pvDet[0] = vResultLo; + pvDet[1] = vResultHi; +} + +#endif INLINE void triangleSetupC(const __m128 vX, const __m128 vY, const __m128 vA, const __m128 &vB, __m128 &vC) { @@ -227,6 +282,27 @@ void viewportTransform(simdvector *v, const SWR_VIEWPORT_MATRICES & vpMatrices) } } +#if USE_SIMD16_FRONTEND +template<uint32_t NumVerts> +INLINE +void viewportTransform(simd16vector *v, const SWR_VIEWPORT_MATRICES & vpMatrices) +{ + const simd16scalar m00 = _simd16_broadcast_ss(&vpMatrices.m00[0]); + const simd16scalar m30 = _simd16_broadcast_ss(&vpMatrices.m30[0]); + const simd16scalar m11 = _simd16_broadcast_ss(&vpMatrices.m11[0]); + const simd16scalar m31 = _simd16_broadcast_ss(&vpMatrices.m31[0]); + const simd16scalar m22 = _simd16_broadcast_ss(&vpMatrices.m22[0]); + const simd16scalar m32 = _simd16_broadcast_ss(&vpMatrices.m32[0]); + + for (uint32_t i = 0; i < NumVerts; ++i) + { + v[i].x = _simd16_fmadd_ps(v[i].x, m00, m30); + v[i].y = _simd16_fmadd_ps(v[i].y, m11, m31); + v[i].z = _simd16_fmadd_ps(v[i].z, m22, m32); + } +} + +#endif template<uint32_t NumVerts> INLINE void viewportTransform(simdvector *v, const SWR_VIEWPORT_MATRICES & vpMatrices, simdscalari vViewportIdx) @@ -247,6 +323,28 @@ void viewportTransform(simdvector *v, const SWR_VIEWPORT_MATRICES & vpMatrices, } } +#if USE_SIMD16_FRONTEND +template<uint32_t NumVerts> +INLINE +void viewportTransform(simd16vector *v, const SWR_VIEWPORT_MATRICES & vpMatrices, simd16scalari vViewportIdx) +{ + // perform a gather of each matrix element based on the viewport array indexes + const simd16scalar m00 = _simd16_i32gather_ps(&vpMatrices.m00[0], vViewportIdx, 4); + const simd16scalar m30 = _simd16_i32gather_ps(&vpMatrices.m30[0], vViewportIdx, 4); + const simd16scalar m11 = _simd16_i32gather_ps(&vpMatrices.m11[0], vViewportIdx, 4); + const simd16scalar m31 = _simd16_i32gather_ps(&vpMatrices.m31[0], vViewportIdx, 4); + const simd16scalar m22 = _simd16_i32gather_ps(&vpMatrices.m22[0], vViewportIdx, 4); + const simd16scalar m32 = _simd16_i32gather_ps(&vpMatrices.m32[0], vViewportIdx, 4); + + for (uint32_t i = 0; i < NumVerts; ++i) + { + v[i].x = _simd16_fmadd_ps(v[i].x, m00, m30); + v[i].y = _simd16_fmadd_ps(v[i].y, m11, m31); + v[i].z = _simd16_fmadd_ps(v[i].z, m22, m32); + } +} + +#endif INLINE void calcBoundingBoxInt(const __m128i &vX, const __m128i &vY, SWR_RECT &bbox) { diff --git a/src/gallium/drivers/swr/rasterizer/core/utils.h b/src/gallium/drivers/swr/rasterizer/core/utils.h index 9dfa16a529e..660a63fe4e3 100644 --- a/src/gallium/drivers/swr/rasterizer/core/utils.h +++ b/src/gallium/drivers/swr/rasterizer/core/utils.h @@ -83,6 +83,16 @@ struct simdBBox simdscalari xmax; }; +#if ENABLE_AVX512_SIMD16 +struct simd16BBox +{ + simd16scalari ymin; + simd16scalari ymax; + simd16scalari xmin; + simd16scalari xmax; +}; + +#endif INLINE void vTranspose(__m128 &row0, __m128 &row1, __m128 &row2, __m128 &row3) { |