// Copyright 2009-2020 Intel Corporation // SPDX-License-Identifier: Apache-2.0 #pragma once #include "node_intersector_packet_stream.h" #include "node_intersector_frustum.h" #include "bvh_traverser_stream.h" namespace embree { namespace isa { /*! BVH ray stream intersector. */ template class BVHNIntersectorStream { static const int Nxd = (Nx == N) ? N : Nx/2; /* shortcuts for frequently used types */ template using PrimitiveIntersectorK = typename PrimitiveIntersector::template Type; template using PrimitiveK = typename PrimitiveIntersectorK::PrimitiveK; typedef BVHN BVH; typedef typename BVH::NodeRef NodeRef; typedef typename BVH::BaseNode BaseNode; typedef typename BVH::AABBNode AABBNode; typedef typename BVH::AABBNodeMB AABBNodeMB; template __forceinline static size_t initPacketsAndFrustum(RayK** inputPackets, size_t numOctantRays, TravRayKStream* packets, Frustum& frustum, bool& commonOctant) { const size_t numPackets = (numOctantRays+K-1)/K; Vec3vf tmp_min_rdir(pos_inf); Vec3vf tmp_max_rdir(neg_inf); Vec3vf tmp_min_org(pos_inf); Vec3vf tmp_max_org(neg_inf); vfloat tmp_min_dist(pos_inf); vfloat tmp_max_dist(neg_inf); size_t m_active = 0; for (size_t i = 0; i < numPackets; i++) { const vfloat tnear = inputPackets[i]->tnear(); const vfloat tfar = inputPackets[i]->tfar; vbool m_valid = (tnear <= tfar) & (tnear >= 0.0f); #if defined(EMBREE_IGNORE_INVALID_RAYS) m_valid &= inputPackets[i]->valid(); #endif m_active |= (size_t)movemask(m_valid) << (i*K); vfloat packet_min_dist = max(tnear, 0.0f); vfloat packet_max_dist = select(m_valid, tfar, neg_inf); tmp_min_dist = min(tmp_min_dist, packet_min_dist); tmp_max_dist = max(tmp_max_dist, packet_max_dist); const Vec3vf& org = inputPackets[i]->org; const Vec3vf& dir = inputPackets[i]->dir; new (&packets[i]) TravRayKStream(org, dir, packet_min_dist, packet_max_dist); tmp_min_rdir = min(tmp_min_rdir, select(m_valid, packets[i].rdir, Vec3vf(pos_inf))); tmp_max_rdir = max(tmp_max_rdir, select(m_valid, packets[i].rdir, Vec3vf(neg_inf))); tmp_min_org = min(tmp_min_org , select(m_valid,org , Vec3vf(pos_inf))); tmp_max_org = max(tmp_max_org , select(m_valid,org , Vec3vf(neg_inf))); } m_active &= (numOctantRays == (8 * sizeof(size_t))) ? (size_t)-1 : (((size_t)1 << numOctantRays)-1); const Vec3fa reduced_min_rdir(reduce_min(tmp_min_rdir.x), reduce_min(tmp_min_rdir.y), reduce_min(tmp_min_rdir.z)); const Vec3fa reduced_max_rdir(reduce_max(tmp_max_rdir.x), reduce_max(tmp_max_rdir.y), reduce_max(tmp_max_rdir.z)); const Vec3fa reduced_min_origin(reduce_min(tmp_min_org.x), reduce_min(tmp_min_org.y), reduce_min(tmp_min_org.z)); const Vec3fa reduced_max_origin(reduce_max(tmp_max_org.x), reduce_max(tmp_max_org.y), reduce_max(tmp_max_org.z)); commonOctant = (reduced_max_rdir.x < 0.0f || reduced_min_rdir.x >= 0.0f) && (reduced_max_rdir.y < 0.0f || reduced_min_rdir.y >= 0.0f) && (reduced_max_rdir.z < 0.0f || reduced_min_rdir.z >= 0.0f); const float frustum_min_dist = reduce_min(tmp_min_dist); const float frustum_max_dist = reduce_max(tmp_max_dist); frustum.init(reduced_min_origin, reduced_max_origin, reduced_min_rdir, reduced_max_rdir, frustum_min_dist, frustum_max_dist, N); return m_active; } template __forceinline static size_t intersectAABBNodePacket(size_t m_active, const TravRayKStream* packets, const AABBNode* __restrict__ node, size_t boxID, const NearFarPrecalculations& nf) { assert(m_active); const size_t startPacketID = bsf(m_active) / K; const size_t endPacketID = bsr(m_active) / K; size_t m_trav_active = 0; for (size_t i = startPacketID; i <= endPacketID; i++) { const size_t m_hit = intersectNodeK(node, boxID, packets[i], nf); m_trav_active |= m_hit << (i*K); } return m_trav_active; } template __forceinline static size_t traverseCoherentStream(size_t m_active, TravRayKStream* packets, const AABBNode* __restrict__ node, const Frustum& frustum, size_t* maskK, vfloat& dist) { size_t m_node_hit = intersectNodeFrustum(node, frustum, dist); const size_t first_index = bsf(m_active); const size_t first_packetID = first_index / K; const size_t first_rayID = first_index % K; size_t m_first_hit = intersectNode1(node, packets[first_packetID], first_rayID, frustum.nf); /* this make traversal independent of the ordering of rays */ size_t m_node = m_node_hit ^ m_first_hit; while (unlikely(m_node)) { const size_t boxID = bscf(m_node); const size_t m_current = m_active & intersectAABBNodePacket(m_active, packets, node, boxID, frustum.nf); m_node_hit ^= m_current ? (size_t)0 : ((size_t)1 << boxID); maskK[boxID] = m_current; } return m_node_hit; } // TODO: explicit 16-wide path for KNL template __forceinline static vint traverseIncoherentStream(size_t m_active, TravRayKStreamFast* __restrict__ packets, const AABBNode* __restrict__ node, const NearFarPrecalculations& nf, const int shiftTable[32]) { const vfloat bminX = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearX)); const vfloat bminY = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearY)); const vfloat bminZ = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearZ)); const vfloat bmaxX = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farX)); const vfloat bmaxY = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farY)); const vfloat bmaxZ = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farZ)); assert(m_active); vint vmask(zero); do { STAT3(shadow.trav_nodes,1,1,1); const size_t rayID = bscf(m_active); assert(rayID < MAX_INTERNAL_STREAM_SIZE); TravRayKStream &p = packets[rayID / K]; const size_t i = rayID % K; const vint bitmask(shiftTable[rayID]); const vfloat tNearX = msub(bminX, p.rdir.x[i], p.org_rdir.x[i]); const vfloat tNearY = msub(bminY, p.rdir.y[i], p.org_rdir.y[i]); const vfloat tNearZ = msub(bminZ, p.rdir.z[i], p.org_rdir.z[i]); const vfloat tFarX = msub(bmaxX, p.rdir.x[i], p.org_rdir.x[i]); const vfloat tFarY = msub(bmaxY, p.rdir.y[i], p.org_rdir.y[i]); const vfloat tFarZ = msub(bmaxZ, p.rdir.z[i], p.org_rdir.z[i]); const vfloat tNear = maxi(tNearX, tNearY, tNearZ, vfloat(p.tnear[i])); const vfloat tFar = mini(tFarX , tFarY , tFarZ, vfloat(p.tfar[i])); #if defined(__AVX512ER__) const vboolx m_node((1 << N)-1); const vbool hit_mask = le(m_node, tNear, tFar); vmask = mask_or(hit_mask, vmask, vmask, bitmask); #else const vbool hit_mask = tNear <= tFar; #if defined(__AVX2__) vmask = vmask | (bitmask & vint(hit_mask)); #else vmask = select(hit_mask, vmask | bitmask, vmask); #endif #endif } while(m_active); return vmask; } template __forceinline static vint traverseIncoherentStream(size_t m_active, TravRayKStreamRobust* __restrict__ packets, const AABBNode* __restrict__ node, const NearFarPrecalculations& nf, const int shiftTable[32]) { const vfloat bminX = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearX)); const vfloat bminY = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearY)); const vfloat bminZ = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.nearZ)); const vfloat bmaxX = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farX)); const vfloat bmaxY = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farY)); const vfloat bmaxZ = vfloat(*(const vfloat*)((const char*)&node->lower_x + nf.farZ)); assert(m_active); vint vmask(zero); do { STAT3(shadow.trav_nodes,1,1,1); const size_t rayID = bscf(m_active); assert(rayID < MAX_INTERNAL_STREAM_SIZE); TravRayKStream &p = packets[rayID / K]; const size_t i = rayID % K; const vint bitmask(shiftTable[rayID]); const vfloat tNearX = (bminX - p.org.x[i]) * p.rdir.x[i]; const vfloat tNearY = (bminY - p.org.y[i]) * p.rdir.y[i]; const vfloat tNearZ = (bminZ - p.org.z[i]) * p.rdir.z[i]; const vfloat tFarX = (bmaxX - p.org.x[i]) * p.rdir.x[i]; const vfloat tFarY = (bmaxY - p.org.y[i]) * p.rdir.y[i]; const vfloat tFarZ = (bmaxZ - p.org.z[i]) * p.rdir.z[i]; const vfloat tNear = maxi(tNearX, tNearY, tNearZ, vfloat(p.tnear[i])); const vfloat tFar = mini(tFarX , tFarY , tFarZ, vfloat(p.tfar[i])); const float round_down = 1.0f-2.0f*float(ulp); const float round_up = 1.0f+2.0f*float(ulp); #if defined(__AVX512ER__) const vboolx m_node((1 << N)-1); const vbool hit_mask = le(m_node, round_down*tNear, round_up*tFar); vmask = mask_or(hit_mask, vmask, vmask, bitmask); #else const vbool hit_mask = round_down*tNear <= round_up*tFar; #if defined(__AVX2__) vmask = vmask | (bitmask & vint(hit_mask)); #else vmask = select(hit_mask, vmask | bitmask, vmask); #endif #endif } while(m_active); return vmask; } static const size_t stackSizeSingle = 1+(N-1)*BVH::maxDepth; public: static void intersect(Accel::Intersectors* This, RayHitN** inputRays, size_t numRays, IntersectContext* context); static void occluded (Accel::Intersectors* This, RayN** inputRays, size_t numRays, IntersectContext* context); private: template static void intersectCoherent(Accel::Intersectors* This, RayHitK** inputRays, size_t numRays, IntersectContext* context); template static void occludedCoherent(Accel::Intersectors* This, RayK** inputRays, size_t numRays, IntersectContext* context); template static void occludedIncoherent(Accel::Intersectors* This, RayK** inputRays, size_t numRays, IntersectContext* context); }; /*! BVH ray stream intersector with direct fallback to packets. */ template class BVHNIntersectorStreamPacketFallback { public: static void intersect(Accel::Intersectors* This, RayHitN** inputRays, size_t numRays, IntersectContext* context); static void occluded (Accel::Intersectors* This, RayN** inputRays, size_t numRays, IntersectContext* context); private: template static void intersectK(Accel::Intersectors* This, RayHitK** inputRays, size_t numRays, IntersectContext* context); template static void occludedK(Accel::Intersectors* This, RayK** inputRays, size_t numRays, IntersectContext* context); }; } }