feat(physics): wire physx sdk into build
This commit is contained in:
46
engine/third_party/physx/include/foundation/unix/PxUnixAoS.h
vendored
Normal file
46
engine/third_party/physx/include/foundation/unix/PxUnixAoS.h
vendored
Normal file
@@ -0,0 +1,46 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXAOS_H
|
||||
#define PXFOUNDATION_PXUNIXAOS_H
|
||||
|
||||
// no includes here! this file should be included from PxcVecMath.h only!!!
|
||||
|
||||
#if !COMPILE_VECTOR_INTRINSICS
|
||||
#error Vector intrinsics should not be included when using scalar implementation.
|
||||
#endif
|
||||
|
||||
#if PX_INTEL_FAMILY
|
||||
#include "foundation/unix/sse2/PxUnixSse2AoS.h"
|
||||
#elif PX_NEON
|
||||
#include "foundation/unix/neon/PxUnixNeonAoS.h"
|
||||
#else
|
||||
#error No SIMD implementation for this unix platform.
|
||||
#endif
|
||||
|
||||
#endif // PXFOUNDATION_PXUNIXAOS_H
|
||||
83
engine/third_party/physx/include/foundation/unix/PxUnixFPU.h
vendored
Normal file
83
engine/third_party/physx/include/foundation/unix/PxUnixFPU.h
vendored
Normal file
@@ -0,0 +1,83 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXFPU_H
|
||||
#define PXFOUNDATION_PXUNIXFPU_H
|
||||
|
||||
#include "foundation/PxPreprocessor.h"
|
||||
|
||||
#if PX_LINUX || PX_OSX
|
||||
|
||||
#if PX_X86 || PX_X64
|
||||
#if PX_EMSCRIPTEN
|
||||
#include <emmintrin.h>
|
||||
#endif
|
||||
#include <xmmintrin.h>
|
||||
#elif PX_NEON
|
||||
#include <arm_neon.h>
|
||||
#endif
|
||||
|
||||
PX_INLINE physx::PxSIMDGuard::PxSIMDGuard(bool enable)
|
||||
#if !PX_EMSCRIPTEN && (PX_X86 || PX_X64)
|
||||
: mEnabled(enable)
|
||||
#endif
|
||||
{
|
||||
#if !PX_EMSCRIPTEN && (PX_X86 || PX_X64)
|
||||
if(enable)
|
||||
{
|
||||
mControlWord = _mm_getcsr();
|
||||
// set default (disable exceptions: _MM_MASK_MASK) and FTZ (_MM_FLUSH_ZERO_ON), DAZ (_MM_DENORMALS_ZERO_ON: (1<<6))
|
||||
_mm_setcsr(_MM_MASK_MASK | _MM_FLUSH_ZERO_ON | (1 << 6));
|
||||
}
|
||||
else
|
||||
{
|
||||
PX_UNUSED(enable);
|
||||
PX_ASSERT(_mm_getcsr() & _MM_FLUSH_ZERO_ON);
|
||||
PX_ASSERT(_mm_getcsr() & (1 << 6));
|
||||
PX_ASSERT(_mm_getcsr() & _MM_MASK_MASK);
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
PX_INLINE physx::PxSIMDGuard::~PxSIMDGuard()
|
||||
{
|
||||
#if !PX_EMSCRIPTEN && (PX_X86 || PX_X64)
|
||||
if(mEnabled)
|
||||
{
|
||||
// restore control word and clear exception flags
|
||||
// (setting exception state flags cause exceptions on the first following fp operation)
|
||||
_mm_setcsr(mControlWord & PxU32(~_MM_EXCEPT_MASK));
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
#else
|
||||
#error No SIMD implementation for this unix platform.
|
||||
#endif // PX_LINUX || PX_OSX
|
||||
|
||||
#endif // #ifndef PXFOUNDATION_PXUNIXFPU_H
|
||||
44
engine/third_party/physx/include/foundation/unix/PxUnixInlineAoS.h
vendored
Normal file
44
engine/third_party/physx/include/foundation/unix/PxUnixInlineAoS.h
vendored
Normal file
@@ -0,0 +1,44 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXINLINEAOS_H
|
||||
#define PXFOUNDATION_PXUNIXINLINEAOS_H
|
||||
|
||||
#if !COMPILE_VECTOR_INTRINSICS
|
||||
#error Vector intrinsics should not be included when using scalar implementation.
|
||||
#endif
|
||||
|
||||
#if PX_INTEL_FAMILY
|
||||
#include "foundation/unix/sse2/PxUnixSse2InlineAoS.h"
|
||||
#elif PX_NEON
|
||||
#include "foundation/unix/neon/PxUnixNeonInlineAoS.h"
|
||||
#else
|
||||
#error No SIMD implementation for this unix platform.
|
||||
#endif
|
||||
|
||||
#endif // PXFOUNDATION_PXUNIXINLINEAOS_H
|
||||
127
engine/third_party/physx/include/foundation/unix/PxUnixIntrinsics.h
vendored
Normal file
127
engine/third_party/physx/include/foundation/unix/PxUnixIntrinsics.h
vendored
Normal file
@@ -0,0 +1,127 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PSFOUNDATION_PSUNIXINTRINSICS_H
|
||||
#define PSFOUNDATION_PSUNIXINTRINSICS_H
|
||||
|
||||
#include "foundation/PxAssert.h"
|
||||
#include <math.h>
|
||||
|
||||
// this file is for internal intrinsics - that is, intrinsics that are used in
|
||||
// cross platform code but do not appear in the API
|
||||
|
||||
#if !(PX_LINUX || PX_APPLE_FAMILY)
|
||||
#error "This file should only be included by unix builds!!"
|
||||
#endif
|
||||
|
||||
#if !PX_DOXYGEN
|
||||
namespace physx
|
||||
{
|
||||
#endif
|
||||
|
||||
PX_FORCE_INLINE void PxMemoryBarrier()
|
||||
{
|
||||
__sync_synchronize();
|
||||
}
|
||||
|
||||
/*!
|
||||
Return the index of the highest set bit. Undefined for zero arg.
|
||||
*/
|
||||
PX_INLINE uint32_t PxHighestSetBitUnsafe(uint64_t v)
|
||||
{
|
||||
return uint32_t(63 - __builtin_clzl(v));
|
||||
}
|
||||
|
||||
/*!
|
||||
Return the index of the highest set bit. Undefined for zero arg.
|
||||
*/
|
||||
PX_INLINE uint32_t PxHighestSetBitUnsafe(uint32_t v)
|
||||
{
|
||||
return uint32_t(31 - __builtin_clz(v));
|
||||
}
|
||||
|
||||
/*!
|
||||
Return the index of the lowest set bit. Undefined for zero arg.
|
||||
*/
|
||||
PX_INLINE uint32_t PxLowestSetBitUnsafe(uint64_t v)
|
||||
{
|
||||
return uint32_t(__builtin_ctzl(v));
|
||||
}
|
||||
|
||||
/*!
|
||||
Return the index of the lowest set bit. Undefined for zero arg.
|
||||
*/
|
||||
PX_INLINE uint32_t PxLowestSetBitUnsafe(uint32_t v)
|
||||
{
|
||||
return uint32_t(__builtin_ctz(v));
|
||||
}
|
||||
|
||||
/*!
|
||||
Returns the index of the highest set bit. Returns 32 for v=0.
|
||||
*/
|
||||
PX_INLINE uint32_t PxCountLeadingZeros(uint32_t v)
|
||||
{
|
||||
if(v)
|
||||
return uint32_t(__builtin_clz(v));
|
||||
else
|
||||
return 32u;
|
||||
}
|
||||
|
||||
/*!
|
||||
Prefetch aligned 64B x86, 32b ARM around \c ptr+offset.
|
||||
*/
|
||||
PX_FORCE_INLINE void PxPrefetchLine(const void* ptr, uint32_t offset = 0)
|
||||
{
|
||||
#if PX_CUDA_COMPILER
|
||||
__builtin_prefetch(reinterpret_cast<const char*>(ptr) + offset, 0, 3);
|
||||
#else
|
||||
__builtin_prefetch(reinterpret_cast<const char* PX_RESTRICT>(ptr) + offset, 0, 3);
|
||||
#endif
|
||||
}
|
||||
|
||||
/*!
|
||||
Prefetch \c count bytes starting at \c ptr.
|
||||
*/
|
||||
PX_FORCE_INLINE void PxPrefetch(const void* ptr, uint32_t count = 1)
|
||||
{
|
||||
const char* cp = reinterpret_cast<const char*>(ptr);
|
||||
uint64_t p = size_t(ptr);
|
||||
uint64_t startLine = p >> 6, endLine = (p + count - 1) >> 6;
|
||||
uint64_t lines = endLine - startLine + 1;
|
||||
do
|
||||
{
|
||||
PxPrefetchLine(cp);
|
||||
cp += 64;
|
||||
} while(--lines);
|
||||
}
|
||||
|
||||
#if !PX_DOXYGEN
|
||||
} // namespace physx
|
||||
#endif
|
||||
|
||||
#endif // #ifndef PSFOUNDATION_PSUNIXINTRINSICS_H
|
||||
179
engine/third_party/physx/include/foundation/unix/PxUnixMathIntrinsics.h
vendored
Normal file
179
engine/third_party/physx/include/foundation/unix/PxUnixMathIntrinsics.h
vendored
Normal file
@@ -0,0 +1,179 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXINTRINSICS_H
|
||||
#define PXFOUNDATION_PXUNIXINTRINSICS_H
|
||||
|
||||
#include "foundation/PxAssert.h"
|
||||
|
||||
#if !(PX_LINUX || PX_APPLE_FAMILY)
|
||||
#error "This file should only be included by Unix builds!!"
|
||||
#endif
|
||||
|
||||
#if PX_LINUX && !PX_CUDA_COMPILER && !PX_EMSCRIPTEN
|
||||
// Linux and CUDA compilation does not work with std::isfnite, as it is not marked as CUDA callable
|
||||
#include <cmath>
|
||||
#ifndef isfinite
|
||||
using std::isfinite;
|
||||
#endif
|
||||
#endif
|
||||
|
||||
#include <math.h>
|
||||
#include <float.h>
|
||||
|
||||
#if !PX_DOXYGEN
|
||||
namespace physx
|
||||
{
|
||||
#endif
|
||||
|
||||
namespace intrinsics
|
||||
{
|
||||
//! \brief platform-specific absolute value
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float abs(float a)
|
||||
{
|
||||
return ::fabsf(a);
|
||||
}
|
||||
|
||||
//! \brief platform-specific select float
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float fsel(float a, float b, float c)
|
||||
{
|
||||
return (a >= 0.0f) ? b : c;
|
||||
}
|
||||
|
||||
//! \brief platform-specific sign
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float sign(float a)
|
||||
{
|
||||
return (a >= 0.0f) ? 1.0f : -1.0f;
|
||||
}
|
||||
|
||||
//! \brief platform-specific reciprocal
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float recip(float a)
|
||||
{
|
||||
return 1.0f / a;
|
||||
}
|
||||
|
||||
//! \brief platform-specific reciprocal estimate
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float recipFast(float a)
|
||||
{
|
||||
return 1.0f / a;
|
||||
}
|
||||
|
||||
//! \brief platform-specific square root
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float sqrt(float a)
|
||||
{
|
||||
return ::sqrtf(a);
|
||||
}
|
||||
|
||||
//! \brief platform-specific reciprocal square root
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float recipSqrt(float a)
|
||||
{
|
||||
return 1.0f / ::sqrtf(a);
|
||||
}
|
||||
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float recipSqrtFast(float a)
|
||||
{
|
||||
return 1.0f / ::sqrtf(a);
|
||||
}
|
||||
|
||||
//! \brief platform-specific sine
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float sin(float a)
|
||||
{
|
||||
return ::sinf(a);
|
||||
}
|
||||
|
||||
//! \brief platform-specific cosine
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float cos(float a)
|
||||
{
|
||||
return ::cosf(a);
|
||||
}
|
||||
|
||||
//! \brief platform-specific minimum
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float selectMin(float a, float b)
|
||||
{
|
||||
return a < b ? a : b;
|
||||
}
|
||||
|
||||
//! \brief platform-specific maximum
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE float selectMax(float a, float b)
|
||||
{
|
||||
return a > b ? a : b;
|
||||
}
|
||||
|
||||
//! \brief platform-specific finiteness check (not INF or NAN)
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE bool isFinite(float a)
|
||||
{
|
||||
//std::isfinite not recommended as of Feb 2017, since it doesn't work with g++/clang's floating point optimization.
|
||||
union localU { PxU32 i; float f; } floatUnion;
|
||||
floatUnion.f = a;
|
||||
return !((floatUnion.i & 0x7fffffff) >= 0x7f800000);
|
||||
}
|
||||
|
||||
//! \brief platform-specific finiteness check (not INF or NAN)
|
||||
PX_CUDA_CALLABLE PX_FORCE_INLINE bool isFinite(double a)
|
||||
{
|
||||
return !!isfinite(a);
|
||||
}
|
||||
|
||||
/*!
|
||||
Sets \c count bytes starting at \c dst to zero.
|
||||
*/
|
||||
PX_FORCE_INLINE void* memZero(void* dest, size_t count)
|
||||
{
|
||||
return memset(dest, 0, count);
|
||||
}
|
||||
|
||||
/*!
|
||||
Sets \c count bytes starting at \c dst to \c c.
|
||||
*/
|
||||
PX_FORCE_INLINE void* memSet(void* dest, int32_t c, size_t count)
|
||||
{
|
||||
return memset(dest, c, count);
|
||||
}
|
||||
|
||||
/*!
|
||||
Copies \c count bytes from \c src to \c dst. User memMove if regions overlap.
|
||||
*/
|
||||
PX_FORCE_INLINE void* memCopy(void* dest, const void* src, size_t count)
|
||||
{
|
||||
return memcpy(dest, src, count);
|
||||
}
|
||||
|
||||
/*!
|
||||
Copies \c count bytes from \c src to \c dst. Supports overlapping regions.
|
||||
*/
|
||||
PX_FORCE_INLINE void* memMove(void* dest, const void* src, size_t count)
|
||||
{
|
||||
return memmove(dest, src, count);
|
||||
}
|
||||
|
||||
} // namespace intrinsics
|
||||
#if !PX_DOXYGEN
|
||||
} // namespace physx
|
||||
#endif
|
||||
|
||||
#endif // #ifndef PXFOUNDATION_PXUNIXINTRINSICS_H
|
||||
76
engine/third_party/physx/include/foundation/unix/PxUnixTrigConstants.h
vendored
Normal file
76
engine/third_party/physx/include/foundation/unix/PxUnixTrigConstants.h
vendored
Normal file
@@ -0,0 +1,76 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXTRIGCONSTANTS_H
|
||||
#define PXFOUNDATION_PXUNIXTRIGCONSTANTS_H
|
||||
|
||||
#include "foundation/PxMath.h"
|
||||
#include "foundation/PxPreprocessor.h"
|
||||
|
||||
namespace physx
|
||||
{
|
||||
namespace aos
|
||||
{
|
||||
|
||||
#if PX_CLANG
|
||||
#if PX_LINUX
|
||||
#pragma clang diagnostic push
|
||||
#pragma clang diagnostic ignored "-Wmissing-variable-declarations"
|
||||
#endif
|
||||
#endif
|
||||
|
||||
#define PX_GLOBALCONST extern const __attribute__((weak))
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct PX_VECTORF32
|
||||
{
|
||||
float f[4];
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_GLOBALCONST PX_VECTORF32 g_PXSinCoefficients0 = { { 1.0f, -0.166666667f, 8.333333333e-3f, -1.984126984e-4f } };
|
||||
PX_GLOBALCONST PX_VECTORF32
|
||||
g_PXSinCoefficients1 = { { 2.755731922e-6f, -2.505210839e-8f, 1.605904384e-10f, -7.647163732e-13f } };
|
||||
PX_GLOBALCONST PX_VECTORF32
|
||||
g_PXSinCoefficients2 = { { 2.811457254e-15f, -8.220635247e-18f, 1.957294106e-20f, -3.868170171e-23f } };
|
||||
PX_GLOBALCONST PX_VECTORF32 g_PXCosCoefficients0 = { { 1.0f, -0.5f, 4.166666667e-2f, -1.388888889e-3f } };
|
||||
PX_GLOBALCONST PX_VECTORF32
|
||||
g_PXCosCoefficients1 = { { 2.480158730e-5f, -2.755731922e-7f, 2.087675699e-9f, -1.147074560e-11f } };
|
||||
PX_GLOBALCONST PX_VECTORF32
|
||||
g_PXCosCoefficients2 = { { 4.779477332e-14f, -1.561920697e-16f, 4.110317623e-19f, -8.896791392e-22f } };
|
||||
PX_GLOBALCONST PX_VECTORF32 g_PXReciprocalTwoPi = { { PxInvTwoPi, PxInvTwoPi, PxInvTwoPi, PxInvTwoPi } };
|
||||
PX_GLOBALCONST PX_VECTORF32 g_PXTwoPi = { { PxTwoPi, PxTwoPi, PxTwoPi, PxTwoPi } };
|
||||
|
||||
#if PX_CLANG
|
||||
#if PX_LINUX
|
||||
#pragma clang diagnostic pop
|
||||
#endif
|
||||
#endif
|
||||
} // namespace aos
|
||||
} // namespace physx
|
||||
|
||||
#endif //PXFOUNDATION_PXUNIXTRIGCONSTANTS_H
|
||||
136
engine/third_party/physx/include/foundation/unix/neon/PxUnixNeonAoS.h
vendored
Normal file
136
engine/third_party/physx/include/foundation/unix/neon/PxUnixNeonAoS.h
vendored
Normal file
@@ -0,0 +1,136 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXNEONAOS_H
|
||||
#define PXFOUNDATION_PXUNIXNEONAOS_H
|
||||
|
||||
// no includes here! this file should be included from PxcVecMath.h only!!!
|
||||
|
||||
#if !COMPILE_VECTOR_INTRINSICS
|
||||
#error Vector intrinsics should not be included when using scalar implementation.
|
||||
#endif
|
||||
|
||||
// only ARM NEON compatible platforms should reach this
|
||||
#include <arm_neon.h>
|
||||
|
||||
namespace physx
|
||||
{
|
||||
namespace aos
|
||||
{
|
||||
|
||||
typedef float32x2_t FloatV;
|
||||
typedef float32x4_t Vec3V;
|
||||
typedef float32x4_t Vec4V;
|
||||
typedef uint32x4_t BoolV;
|
||||
typedef float32x4_t QuatV;
|
||||
|
||||
typedef uint32x4_t VecU32V;
|
||||
typedef int32x4_t VecI32V;
|
||||
typedef uint16x8_t VecU16V;
|
||||
typedef int16x8_t VecI16V;
|
||||
typedef uint8x16_t VecU8V;
|
||||
|
||||
#define FloatVArg FloatV &
|
||||
#define Vec3VArg Vec3V &
|
||||
#define Vec4VArg Vec4V &
|
||||
#define BoolVArg BoolV &
|
||||
#define VecU32VArg VecU32V &
|
||||
#define VecI32VArg VecI32V &
|
||||
#define VecU16VArg VecU16V &
|
||||
#define VecI16VArg VecI16V &
|
||||
#define VecU8VArg VecU8V &
|
||||
#define QuatVArg QuatV &
|
||||
|
||||
// KS - TODO - make an actual VecCrossV type for NEON
|
||||
#define VecCrossV Vec3V
|
||||
|
||||
typedef VecI32V VecShiftV;
|
||||
#define VecShiftVArg VecShiftV &
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat33V
|
||||
{
|
||||
Mat33V()
|
||||
{
|
||||
}
|
||||
Mat33V(const Vec3V& c0, const Vec3V& c1, const Vec3V& c2) : col0(c0), col1(c1), col2(c2)
|
||||
{
|
||||
}
|
||||
Vec3V PX_ALIGN(16, col0);
|
||||
Vec3V PX_ALIGN(16, col1);
|
||||
Vec3V PX_ALIGN(16, col2);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat34V
|
||||
{
|
||||
Mat34V()
|
||||
{
|
||||
}
|
||||
Mat34V(const Vec3V& c0, const Vec3V& c1, const Vec3V& c2, const Vec3V& c3) : col0(c0), col1(c1), col2(c2), col3(c3)
|
||||
{
|
||||
}
|
||||
Vec3V PX_ALIGN(16, col0);
|
||||
Vec3V PX_ALIGN(16, col1);
|
||||
Vec3V PX_ALIGN(16, col2);
|
||||
Vec3V PX_ALIGN(16, col3);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat43V
|
||||
{
|
||||
Mat43V()
|
||||
{
|
||||
}
|
||||
Mat43V(const Vec4V& c0, const Vec4V& c1, const Vec4V& c2) : col0(c0), col1(c1), col2(c2)
|
||||
{
|
||||
}
|
||||
Vec4V PX_ALIGN(16, col0);
|
||||
Vec4V PX_ALIGN(16, col1);
|
||||
Vec4V PX_ALIGN(16, col2);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat44V
|
||||
{
|
||||
Mat44V()
|
||||
{
|
||||
}
|
||||
Mat44V(const Vec4V& c0, const Vec4V& c1, const Vec4V& c2, const Vec4V& c3) : col0(c0), col1(c1), col2(c2), col3(c3)
|
||||
{
|
||||
}
|
||||
Vec4V PX_ALIGN(16, col0);
|
||||
Vec4V PX_ALIGN(16, col1);
|
||||
Vec4V PX_ALIGN(16, col2);
|
||||
Vec4V PX_ALIGN(16, col3);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
} // namespace aos
|
||||
} // namespace physx
|
||||
|
||||
#endif // PXFOUNDATION_PXUNIXNEONAOS_H
|
||||
3636
engine/third_party/physx/include/foundation/unix/neon/PxUnixNeonInlineAoS.h
vendored
Normal file
3636
engine/third_party/physx/include/foundation/unix/neon/PxUnixNeonInlineAoS.h
vendored
Normal file
File diff suppressed because it is too large
Load Diff
187
engine/third_party/physx/include/foundation/unix/sse2/PxUnixSse2AoS.h
vendored
Normal file
187
engine/third_party/physx/include/foundation/unix/sse2/PxUnixSse2AoS.h
vendored
Normal file
@@ -0,0 +1,187 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXSSE2AOS_H
|
||||
#define PXFOUNDATION_PXUNIXSSE2AOS_H
|
||||
|
||||
// no includes here! this file should be included from PxcVecMath.h only!!!
|
||||
|
||||
#if !COMPILE_VECTOR_INTRINSICS
|
||||
#error Vector intrinsics should not be included when using scalar implementation.
|
||||
#endif
|
||||
|
||||
namespace physx
|
||||
{
|
||||
namespace aos
|
||||
{
|
||||
|
||||
#if PX_EMSCRIPTEN
|
||||
typedef int8_t __int8_t;
|
||||
typedef int16_t __int16_t;
|
||||
typedef int32_t __int32_t;
|
||||
typedef int64_t __int64_t;
|
||||
typedef uint16_t __uint16_t;
|
||||
typedef uint32_t __uint32_t;
|
||||
typedef uint64_t __uint64_t;
|
||||
#endif
|
||||
|
||||
typedef union UnionM128
|
||||
{
|
||||
UnionM128()
|
||||
{
|
||||
}
|
||||
UnionM128(__m128 in)
|
||||
{
|
||||
m128 = in;
|
||||
}
|
||||
|
||||
UnionM128(__m128i in)
|
||||
{
|
||||
m128i = in;
|
||||
}
|
||||
|
||||
operator __m128()
|
||||
{
|
||||
return m128;
|
||||
}
|
||||
|
||||
operator __m128() const
|
||||
{
|
||||
return m128;
|
||||
}
|
||||
|
||||
float m128_f32[4];
|
||||
__int8_t m128_i8[16];
|
||||
__int16_t m128_i16[8];
|
||||
__int32_t m128_i32[4];
|
||||
__int64_t m128_i64[2];
|
||||
__uint16_t m128_u16[8];
|
||||
__uint32_t m128_u32[4];
|
||||
__uint64_t m128_u64[2];
|
||||
__m128 m128;
|
||||
__m128i m128i;
|
||||
} UnionM128;
|
||||
|
||||
typedef __m128 FloatV;
|
||||
typedef __m128 Vec3V;
|
||||
typedef __m128 Vec4V;
|
||||
typedef __m128 BoolV;
|
||||
typedef __m128 QuatV;
|
||||
typedef __m128i VecI32V;
|
||||
typedef UnionM128 VecU32V;
|
||||
typedef UnionM128 VecU16V;
|
||||
typedef UnionM128 VecI16V;
|
||||
typedef UnionM128 VecU8V;
|
||||
|
||||
#define FloatVArg FloatV &
|
||||
#define Vec3VArg Vec3V &
|
||||
#define Vec4VArg Vec4V &
|
||||
#define BoolVArg BoolV &
|
||||
#define VecU32VArg VecU32V &
|
||||
#define VecI32VArg VecI32V &
|
||||
#define VecU16VArg VecU16V &
|
||||
#define VecI16VArg VecI16V &
|
||||
#define VecU8VArg VecU8V &
|
||||
#define QuatVArg QuatV &
|
||||
|
||||
// Optimization for situations in which you cross product multiple vectors with the same vector.
|
||||
// Avoids 2X shuffles per product
|
||||
struct VecCrossV
|
||||
{
|
||||
Vec3V mL1;
|
||||
Vec3V mR1;
|
||||
};
|
||||
|
||||
struct VecShiftV
|
||||
{
|
||||
VecI32V shift;
|
||||
};
|
||||
#define VecShiftVArg VecShiftV &
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat33V
|
||||
{
|
||||
Mat33V()
|
||||
{
|
||||
}
|
||||
Mat33V(const Vec3V& c0, const Vec3V& c1, const Vec3V& c2) : col0(c0), col1(c1), col2(c2)
|
||||
{
|
||||
}
|
||||
Vec3V PX_ALIGN(16, col0);
|
||||
Vec3V PX_ALIGN(16, col1);
|
||||
Vec3V PX_ALIGN(16, col2);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat34V
|
||||
{
|
||||
Mat34V()
|
||||
{
|
||||
}
|
||||
Mat34V(const Vec3V& c0, const Vec3V& c1, const Vec3V& c2, const Vec3V& c3) : col0(c0), col1(c1), col2(c2), col3(c3)
|
||||
{
|
||||
}
|
||||
Vec3V PX_ALIGN(16, col0);
|
||||
Vec3V PX_ALIGN(16, col1);
|
||||
Vec3V PX_ALIGN(16, col2);
|
||||
Vec3V PX_ALIGN(16, col3);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat43V
|
||||
{
|
||||
Mat43V()
|
||||
{
|
||||
}
|
||||
Mat43V(const Vec4V& c0, const Vec4V& c1, const Vec4V& c2) : col0(c0), col1(c1), col2(c2)
|
||||
{
|
||||
}
|
||||
Vec4V PX_ALIGN(16, col0);
|
||||
Vec4V PX_ALIGN(16, col1);
|
||||
Vec4V PX_ALIGN(16, col2);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
PX_ALIGN_PREFIX(16)
|
||||
struct Mat44V
|
||||
{
|
||||
Mat44V()
|
||||
{
|
||||
}
|
||||
Mat44V(const Vec4V& c0, const Vec4V& c1, const Vec4V& c2, const Vec4V& c3) : col0(c0), col1(c1), col2(c2), col3(c3)
|
||||
{
|
||||
}
|
||||
Vec4V PX_ALIGN(16, col0);
|
||||
Vec4V PX_ALIGN(16, col1);
|
||||
Vec4V PX_ALIGN(16, col2);
|
||||
Vec4V PX_ALIGN(16, col3);
|
||||
} PX_ALIGN_SUFFIX(16);
|
||||
|
||||
} // namespace aos
|
||||
} // namespace physx
|
||||
|
||||
#endif // PXFOUNDATION_PXUNIXSSE2AOS_H
|
||||
719
engine/third_party/physx/include/foundation/unix/sse2/PxUnixSse2InlineAoS.h
vendored
Normal file
719
engine/third_party/physx/include/foundation/unix/sse2/PxUnixSse2InlineAoS.h
vendored
Normal file
@@ -0,0 +1,719 @@
|
||||
// Redistribution and use in source and binary forms, with or without
|
||||
// modification, are permitted provided that the following conditions
|
||||
// are met:
|
||||
// * Redistributions of source code must retain the above copyright
|
||||
// notice, this list of conditions and the following disclaimer.
|
||||
// * Redistributions in binary form must reproduce the above copyright
|
||||
// notice, this list of conditions and the following disclaimer in the
|
||||
// documentation and/or other materials provided with the distribution.
|
||||
// * Neither the name of NVIDIA CORPORATION nor the names of its
|
||||
// contributors may be used to endorse or promote products derived
|
||||
// from this software without specific prior written permission.
|
||||
//
|
||||
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ''AS IS'' AND ANY
|
||||
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
// IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
||||
// CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||
// EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
||||
// PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
||||
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
//
|
||||
// Copyright (c) 2008-2025 NVIDIA Corporation. All rights reserved.
|
||||
// Copyright (c) 2004-2008 AGEIA Technologies, Inc. All rights reserved.
|
||||
// Copyright (c) 2001-2004 NovodeX AG. All rights reserved.
|
||||
|
||||
#ifndef PXFOUNDATION_PXUNIXSSE2INLINEAOS_H
|
||||
#define PXFOUNDATION_PXUNIXSSE2INLINEAOS_H
|
||||
|
||||
namespace physx
|
||||
{
|
||||
namespace aos
|
||||
{
|
||||
//////////////////////////////////////////////////////////////////////
|
||||
//Test that Vec3V and FloatV are legal
|
||||
//////////////////////////////////////////////////////////////////////
|
||||
|
||||
#define FLOAT_COMPONENTS_EQUAL_THRESHOLD 0.01f
|
||||
PX_FORCE_INLINE static bool isValidFloatV(const FloatV a)
|
||||
{
|
||||
const PxF32 x = V4ReadX(a);
|
||||
const PxF32 y = V4ReadY(a);
|
||||
const PxF32 z = V4ReadZ(a);
|
||||
const PxF32 w = V4ReadW(a);
|
||||
|
||||
return (x == y && x == z && x == w);
|
||||
|
||||
/*if (
|
||||
(PxAbs(x - y) < FLOAT_COMPONENTS_EQUAL_THRESHOLD) &&
|
||||
(PxAbs(x - z) < FLOAT_COMPONENTS_EQUAL_THRESHOLD) &&
|
||||
(PxAbs(x - w) < FLOAT_COMPONENTS_EQUAL_THRESHOLD)
|
||||
)
|
||||
{
|
||||
return true;
|
||||
}
|
||||
|
||||
if (
|
||||
(PxAbs((x - y) / x) < FLOAT_COMPONENTS_EQUAL_THRESHOLD) &&
|
||||
(PxAbs((x - z) / x) < FLOAT_COMPONENTS_EQUAL_THRESHOLD) &&
|
||||
(PxAbs((x - w) / x) < FLOAT_COMPONENTS_EQUAL_THRESHOLD)
|
||||
)
|
||||
{
|
||||
return true;
|
||||
}
|
||||
return false;*/
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
#include "../../PxVecMathSSE.h"
|
||||
|
||||
namespace physx
|
||||
{
|
||||
namespace aos
|
||||
{
|
||||
|
||||
#define PX_FPCLASS_SNAN 0x0001 /* signaling NaN */
|
||||
#define PX_FPCLASS_QNAN 0x0002 /* quiet NaN */
|
||||
#define PX_FPCLASS_NINF 0x0004 /* negative infinity */
|
||||
#define PX_FPCLASS_PINF 0x0200 /* positive infinity */
|
||||
|
||||
namespace internalSimd
|
||||
{
|
||||
#if !PX_EMSCRIPTEN
|
||||
#if PX_CLANG
|
||||
#if PX_LINUX
|
||||
#pragma clang diagnostic push
|
||||
#pragma clang diagnostic ignored "-Wglobal-constructors"
|
||||
#endif
|
||||
#endif
|
||||
const PX_ALIGN(16, PxF32 gMaskXYZ[4]) = { physx::PxUnionCast<PxF32>(0xffffffff), physx::PxUnionCast<PxF32>(0xffffffff),
|
||||
physx::PxUnionCast<PxF32>(0xffffffff), 0 };
|
||||
#if PX_CLANG
|
||||
#if PX_LINUX
|
||||
#pragma clang diagnostic pop
|
||||
#endif
|
||||
#endif
|
||||
#else
|
||||
// emscripten doesn't like the PxUnionCast data structure
|
||||
// the following is what windows and xbox does -- using these for emscripten
|
||||
const PX_ALIGN(16, PxU32 gMaskXYZ[4]) = { 0xffffffff, 0xffffffff, 0xffffffff, 0 };
|
||||
#endif
|
||||
}
|
||||
|
||||
namespace vecMathTests
|
||||
{
|
||||
PX_FORCE_INLINE bool allElementsEqualBoolV(const BoolV a, const BoolV b)
|
||||
{
|
||||
return internalSimd::BAllTrue4_R(VecI32V_IsEq(internalSimd::m128_F2I(a), internalSimd::m128_F2I(b))) != 0;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE bool allElementsEqualVecI32V(const VecI32V a, const VecI32V b)
|
||||
{
|
||||
BoolV c = internalSimd::m128_I2F(_mm_cmpeq_epi32(a, b));
|
||||
return internalSimd::BAllTrue4_R(c) != 0;
|
||||
}
|
||||
|
||||
#define VECMATH_AOS_EPSILON (1e-3f)
|
||||
|
||||
PX_FORCE_INLINE bool allElementsNearEqualFloatV(const FloatV a, const FloatV b)
|
||||
{
|
||||
ASSERT_ISVALIDFLOATV(a);
|
||||
ASSERT_ISVALIDFLOATV(b);
|
||||
const FloatV c = FSub(a, b);
|
||||
const FloatV minError = FLoad(-VECMATH_AOS_EPSILON);
|
||||
const FloatV maxError = FLoad(VECMATH_AOS_EPSILON);
|
||||
return _mm_comigt_ss(c, minError) && _mm_comilt_ss(c, maxError);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE bool allElementsNearEqualVec3V(const Vec3V a, const Vec3V b)
|
||||
{
|
||||
const Vec3V c = V3Sub(a, b);
|
||||
const Vec3V minError = V3Load(-VECMATH_AOS_EPSILON);
|
||||
const Vec3V maxError = V3Load(VECMATH_AOS_EPSILON);
|
||||
return (_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(0, 0, 0, 0)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(0, 0, 0, 0)), maxError) &&
|
||||
_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(1, 1, 1, 1)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(1, 1, 1, 1)), maxError) &&
|
||||
_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(2, 2, 2, 2)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(2, 2, 2, 2)), maxError));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE bool allElementsNearEqualVec4V(const Vec4V a, const Vec4V b)
|
||||
{
|
||||
const Vec4V c = V4Sub(a, b);
|
||||
const Vec4V minError = V4Load(-VECMATH_AOS_EPSILON);
|
||||
const Vec4V maxError = V4Load(VECMATH_AOS_EPSILON);
|
||||
return (_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(0, 0, 0, 0)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(0, 0, 0, 0)), maxError) &&
|
||||
_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(1, 1, 1, 1)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(1, 1, 1, 1)), maxError) &&
|
||||
_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(2, 2, 2, 2)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(2, 2, 2, 2)), maxError) &&
|
||||
_mm_comigt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(3, 3, 3, 3)), minError) &&
|
||||
_mm_comilt_ss(_mm_shuffle_ps(c, c, _MM_SHUFFLE(3, 3, 3, 3)), maxError));
|
||||
}
|
||||
} //vecMathTests
|
||||
|
||||
/////////////////////////////////////////////////////////////////////
|
||||
////FUNCTIONS USED ONLY FOR ASSERTS IN VECTORISED IMPLEMENTATIONS
|
||||
/////////////////////////////////////////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE bool isFiniteFloatV(const FloatV a)
|
||||
{
|
||||
PxF32 badNumber =
|
||||
physx::PxUnionCast<PxF32, PxU32>(PX_FPCLASS_SNAN | PX_FPCLASS_QNAN | PX_FPCLASS_NINF | PX_FPCLASS_PINF);
|
||||
const FloatV vBadNum = FLoad(badNumber);
|
||||
const BoolV vMask = BAnd(vBadNum, a);
|
||||
return internalSimd::FiniteTestEq(vMask, BFFFF()) == 1;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE bool isFiniteVec3V(const Vec3V a)
|
||||
{
|
||||
PxF32 badNumber =
|
||||
physx::PxUnionCast<PxF32, PxU32>(PX_FPCLASS_SNAN | PX_FPCLASS_QNAN | PX_FPCLASS_NINF | PX_FPCLASS_PINF);
|
||||
const Vec3V vBadNum = V3Load(badNumber);
|
||||
const BoolV vMask = BAnd(BAnd(vBadNum, a), BTTTF());
|
||||
return internalSimd::FiniteTestEq(vMask, BFFFF()) == 1;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE bool isFiniteVec4V(const Vec4V a)
|
||||
{
|
||||
/*Vec4V a;
|
||||
PX_ALIGN(16, PxF32 f[4]);
|
||||
F32Array_Aligned_From_Vec4V(a, f);
|
||||
return PxIsFinite(f[0])
|
||||
&& PxIsFinite(f[1])
|
||||
&& PxIsFinite(f[2])
|
||||
&& PxIsFinite(f[3]);*/
|
||||
|
||||
PxF32 badNumber =
|
||||
physx::PxUnionCast<PxF32, PxU32>(PX_FPCLASS_SNAN | PX_FPCLASS_QNAN | PX_FPCLASS_NINF | PX_FPCLASS_PINF);
|
||||
const Vec4V vBadNum = V4Load(badNumber);
|
||||
const BoolV vMask = BAnd(vBadNum, a);
|
||||
|
||||
return internalSimd::FiniteTestEq(vMask, BFFFF()) == 1;
|
||||
}
|
||||
|
||||
/////////////////////////////////////////////////////////////////////
|
||||
////VECTORISED FUNCTION IMPLEMENTATIONS
|
||||
/////////////////////////////////////////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3LoadA(const PxVec3& f)
|
||||
{
|
||||
ASSERT_ISALIGNED16(&f);
|
||||
#if !PX_EMSCRIPTEN
|
||||
return _mm_and_ps(reinterpret_cast<const Vec3V&>(f), V4LoadA(internalSimd::gMaskXYZ));
|
||||
#else
|
||||
return _mm_and_ps((Vec3V&)f, (VecI32V&)internalSimd::gMaskXYZ);
|
||||
#endif
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3LoadUnsafeA(const PxVec3& f)
|
||||
{
|
||||
ASSERT_ISALIGNED16(&f);
|
||||
return _mm_set_ps(0.0f, f.z, f.y, f.x);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3LoadA(const PxF32* const f)
|
||||
{
|
||||
ASSERT_ISALIGNED16(f);
|
||||
#if !PX_EMSCRIPTEN
|
||||
return _mm_and_ps(V4LoadA(f), V4LoadA(internalSimd::gMaskXYZ));
|
||||
#else
|
||||
return _mm_and_ps((Vec3V&)*f, (VecI32V&)internalSimd::gMaskXYZ);
|
||||
#endif
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE void I4StoreA(const VecI32V iv, PxI32* i)
|
||||
{
|
||||
ASSERT_ISALIGNED16(i);
|
||||
_mm_store_ps(reinterpret_cast<float*>(i), internalSimd::m128_I2F(iv));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE BoolV BLoad(const bool* const f)
|
||||
{
|
||||
const PX_ALIGN(16, PxI32) b[4] = { -PxI32(f[0]), -PxI32(f[1]), -PxI32(f[2]), -PxI32(f[3]) };
|
||||
return _mm_load_ps(reinterpret_cast<const float*>(&b));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE void V3StoreA(const Vec3V a, PxVec3& f)
|
||||
{
|
||||
ASSERT_ISALIGNED16(&f);
|
||||
PX_ALIGN(16, PxF32) f2[4];
|
||||
_mm_store_ps(f2, a);
|
||||
f = PxVec3(f2[0], f2[1], f2[2]);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE void V3StoreU(const Vec3V a, PxVec3& f)
|
||||
{
|
||||
PX_ALIGN(16, PxF32) f2[4];
|
||||
_mm_store_ps(f2, a);
|
||||
f = PxVec3(f2[0], f2[1], f2[2]);
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// FLOATV
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE FloatV FAbs(const FloatV a)
|
||||
{
|
||||
ASSERT_ISVALIDFLOATV(a);
|
||||
PX_ALIGN(16, const PxU32) absMask[4] = { 0x7fFFffFF, 0x7fFFffFF, 0x7fFFffFF, 0x7fFFffFF };
|
||||
return _mm_and_ps(a, _mm_load_ps(reinterpret_cast<const PxF32*>(absMask)));
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// VEC3V
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3UnitX()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) x[4] = { 1.0f, 0.0f, 0.0f, 0.0f };
|
||||
const __m128 x128 = _mm_load_ps(x);
|
||||
return x128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3UnitY()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) y[4] = { 0.0f, 1.0f, 0.0f, 0.0f };
|
||||
const __m128 y128 = _mm_load_ps(y);
|
||||
return y128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec3V V3UnitZ()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) z[4] = { 0.0f, 0.0f, 1.0f, 0.0f };
|
||||
const __m128 z128 = _mm_load_ps(z);
|
||||
return z128;
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// VEC4V
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4UnitW()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) w[4] = { 0.0f, 0.0f, 0.0f, 1.0f };
|
||||
const __m128 w128 = _mm_load_ps(w);
|
||||
return w128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4UnitX()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) x[4] = { 1.0f, 0.0f, 0.0f, 0.0f };
|
||||
const __m128 x128 = _mm_load_ps(x);
|
||||
return x128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4UnitY()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) y[4] = { 0.0f, 1.0f, 0.0f, 0.0f };
|
||||
const __m128 y128 = _mm_load_ps(y);
|
||||
return y128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4UnitZ()
|
||||
{
|
||||
const PX_ALIGN(16, PxF32) z[4] = { 0.0f, 0.0f, 1.0f, 0.0f };
|
||||
const __m128 z128 = _mm_load_ps(z);
|
||||
return z128;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4ClearW(const Vec4V v)
|
||||
{
|
||||
#if !PX_EMSCRIPTEN
|
||||
return _mm_and_ps(v, V4LoadA(internalSimd::gMaskXYZ));
|
||||
#else
|
||||
return _mm_and_ps(v, (VecI32V&)internalSimd::gMaskXYZ);
|
||||
#endif
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// BoolV
|
||||
//////////////////////////////////
|
||||
|
||||
/*
|
||||
template<int index> PX_FORCE_INLINE BoolV BSplatElement(BoolV a)
|
||||
{
|
||||
BoolV result;
|
||||
result[0] = result[1] = result[2] = result[3] = a[index];
|
||||
return result;
|
||||
}
|
||||
*/
|
||||
|
||||
template <int index>
|
||||
BoolV BSplatElement(BoolV a)
|
||||
{
|
||||
float* data = reinterpret_cast<float*>(&a);
|
||||
return V4Load(data[index]);
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// MAT33V
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec3V M33TrnspsMulV3(const Mat33V& a, const Vec3V b)
|
||||
{
|
||||
const FloatV x = V3Dot(a.col0, b);
|
||||
const FloatV y = V3Dot(a.col1, b);
|
||||
const FloatV z = V3Dot(a.col2, b);
|
||||
return V3Merge(x, y, z);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Mat33V M33Trnsps(const Mat33V& a)
|
||||
{
|
||||
return Mat33V(V3Merge(V3GetX(a.col0), V3GetX(a.col1), V3GetX(a.col2)),
|
||||
V3Merge(V3GetY(a.col0), V3GetY(a.col1), V3GetY(a.col2)),
|
||||
V3Merge(V3GetZ(a.col0), V3GetZ(a.col1), V3GetZ(a.col2)));
|
||||
}
|
||||
|
||||
/*PX_FORCE_INLINE Mat33V PromoteVec3V(const Vec3V v)
|
||||
{
|
||||
const BoolV bTFFF = BTFFF();
|
||||
const BoolV bFTFF = BFTFF();
|
||||
const BoolV bFFTF = BTFTF();
|
||||
|
||||
const Vec3V zero = V3Zero();
|
||||
|
||||
return Mat33V(V3Sel(bTFFF, v, zero), V3Sel(bFTFF, v, zero), V3Sel(bFFTF, v, zero));
|
||||
}*/
|
||||
|
||||
//////////////////////////////////
|
||||
// MAT34V
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec3V M34TrnspsMul33V3(const Mat34V& a, const Vec3V b)
|
||||
{
|
||||
const FloatV x = V3Dot(a.col0, b);
|
||||
const FloatV y = V3Dot(a.col1, b);
|
||||
const FloatV z = V3Dot(a.col2, b);
|
||||
return V3Merge(x, y, z);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Mat33V M34Trnsps33(const Mat34V& a)
|
||||
{
|
||||
return Mat33V(V3Merge(V3GetX(a.col0), V3GetX(a.col1), V3GetX(a.col2)),
|
||||
V3Merge(V3GetY(a.col0), V3GetY(a.col1), V3GetY(a.col2)),
|
||||
V3Merge(V3GetZ(a.col0), V3GetZ(a.col1), V3GetZ(a.col2)));
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// MAT44V
|
||||
//////////////////////////////////
|
||||
|
||||
PX_FORCE_INLINE Vec4V M44TrnspsMulV4(const Mat44V& a, const Vec4V b)
|
||||
{
|
||||
PX_ALIGN(16, FloatV) dotProdArray[4] = { V4Dot(a.col0, b), V4Dot(a.col1, b), V4Dot(a.col2, b), V4Dot(a.col3, b) };
|
||||
return V4Merge(dotProdArray);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Mat44V M44Trnsps(const Mat44V& a)
|
||||
{
|
||||
const Vec4V v0 = _mm_unpacklo_ps(a.col0, a.col2);
|
||||
const Vec4V v1 = _mm_unpackhi_ps(a.col0, a.col2);
|
||||
const Vec4V v2 = _mm_unpacklo_ps(a.col1, a.col3);
|
||||
const Vec4V v3 = _mm_unpackhi_ps(a.col1, a.col3);
|
||||
return Mat44V(_mm_unpacklo_ps(v0, v2), _mm_unpackhi_ps(v0, v2), _mm_unpacklo_ps(v1, v3), _mm_unpackhi_ps(v1, v3));
|
||||
}
|
||||
|
||||
//////////////////////////////////
|
||||
// Misc
|
||||
//////////////////////////////////
|
||||
|
||||
/*
|
||||
// AP: work in progress - use proper SSE intrinsics where possible
|
||||
PX_FORCE_INLINE VecU16V V4U32PK(VecU32V a, VecU32V b)
|
||||
{
|
||||
VecU16V result;
|
||||
result.m128_u16[0] = PxU16(PxClamp<PxU32>((a).m128_u32[0], 0, 0xFFFF));
|
||||
result.m128_u16[1] = PxU16(PxClamp<PxU32>((a).m128_u32[1], 0, 0xFFFF));
|
||||
result.m128_u16[2] = PxU16(PxClamp<PxU32>((a).m128_u32[2], 0, 0xFFFF));
|
||||
result.m128_u16[3] = PxU16(PxClamp<PxU32>((a).m128_u32[3], 0, 0xFFFF));
|
||||
result.m128_u16[4] = PxU16(PxClamp<PxU32>((b).m128_u32[0], 0, 0xFFFF));
|
||||
result.m128_u16[5] = PxU16(PxClamp<PxU32>((b).m128_u32[1], 0, 0xFFFF));
|
||||
result.m128_u16[6] = PxU16(PxClamp<PxU32>((b).m128_u32[2], 0, 0xFFFF));
|
||||
result.m128_u16[7] = PxU16(PxClamp<PxU32>((b).m128_u32[3], 0, 0xFFFF));
|
||||
return result;
|
||||
}
|
||||
*/
|
||||
|
||||
/*
|
||||
PX_FORCE_INLINE VecU16V V4U16Or(VecU16V a, VecU16V b)
|
||||
{
|
||||
return m128_I2F(_mm_or_si128(m128_F2I(a), m128_F2I(b)));
|
||||
}
|
||||
*/
|
||||
|
||||
/*
|
||||
PX_FORCE_INLINE VecU16V V4U16And(VecU16V a, VecU16V b)
|
||||
{
|
||||
return m128_I2F(_mm_and_si128(m128_F2I(a), m128_F2I(b)));
|
||||
}
|
||||
*/
|
||||
|
||||
/*
|
||||
PX_FORCE_INLINE VecU16V V4U16Andc(VecU16V a, VecU16V b)
|
||||
{
|
||||
return m128_I2F(_mm_andnot_si128(m128_F2I(b), m128_F2I(a)));
|
||||
}
|
||||
*/
|
||||
|
||||
PX_FORCE_INLINE VecI32V I4LoadXYZW(const PxI32& x, const PxI32& y, const PxI32& z, const PxI32& w)
|
||||
{
|
||||
return _mm_set_epi32(w, z, y, x);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V I4Load(const PxI32 i)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_load1_ps(reinterpret_cast<const PxF32*>(&i)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V I4LoadU(const PxI32* i)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_loadu_ps(reinterpret_cast<const PxF32*>(i)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V I4LoadA(const PxI32* i)
|
||||
{
|
||||
ASSERT_ISALIGNED16(i);
|
||||
return internalSimd::m128_F2I(_mm_load_ps(reinterpret_cast<const PxF32*>(i)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Add(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return _mm_add_epi32(a, b);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Sub(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return _mm_sub_epi32(a, b);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE BoolV VecI32V_IsGrtr(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return internalSimd::m128_I2F(_mm_cmpgt_epi32(a, b));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE BoolV VecI32V_IsEq(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return internalSimd::m128_I2F(_mm_cmpeq_epi32(a, b));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V V4I32Sel(const BoolV c, const VecI32V a, const VecI32V b)
|
||||
{
|
||||
return _mm_or_si128(_mm_andnot_si128(internalSimd::m128_F2I(c), b), _mm_and_si128(internalSimd::m128_F2I(c), a));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Zero()
|
||||
{
|
||||
return _mm_setzero_si128();
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Sel(const BoolV c, const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return _mm_or_si128(_mm_andnot_si128(internalSimd::m128_F2I(c), b), _mm_and_si128(internalSimd::m128_F2I(c), a));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecShiftV VecI32V_PrepareShift(const VecI32VArg shift)
|
||||
{
|
||||
VecShiftV s;
|
||||
s.shift = VecI32V_Sel(BTFFF(), shift, VecI32V_Zero());
|
||||
return s;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_LeftShift(const VecI32VArg a, const VecShiftVArg count)
|
||||
{
|
||||
return _mm_sll_epi32(a, count.shift);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_RightShift(const VecI32VArg a, const VecShiftVArg count)
|
||||
{
|
||||
return _mm_srl_epi32(a, count.shift);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_LeftShift(const VecI32VArg a, const PxU32 count)
|
||||
{
|
||||
return _mm_slli_epi32(a, PxI32(count));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_RightShift(const VecI32VArg a, const PxU32 count)
|
||||
{
|
||||
return _mm_srai_epi32(a, PxI32(count));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_And(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return _mm_and_si128(a, b);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Or(const VecI32VArg a, const VecI32VArg b)
|
||||
{
|
||||
return _mm_or_si128(a, b);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_GetX(const VecI32VArg a)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_shuffle_ps(internalSimd::m128_I2F(a), internalSimd::m128_I2F(a), _MM_SHUFFLE(0, 0, 0, 0)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_GetY(const VecI32VArg a)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_shuffle_ps(internalSimd::m128_I2F(a), internalSimd::m128_I2F(a), _MM_SHUFFLE(1, 1, 1, 1)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_GetZ(const VecI32VArg a)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_shuffle_ps(internalSimd::m128_I2F(a), internalSimd::m128_I2F(a), _MM_SHUFFLE(2, 2, 2, 2)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_GetW(const VecI32VArg a)
|
||||
{
|
||||
return internalSimd::m128_F2I(_mm_shuffle_ps(internalSimd::m128_I2F(a), internalSimd::m128_I2F(a), _MM_SHUFFLE(3, 3, 3, 3)));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE void PxI32_From_VecI32V(const VecI32VArg a, PxI32* i)
|
||||
{
|
||||
_mm_store_ss(reinterpret_cast<PxF32*>(i), internalSimd::m128_I2F(a));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_From_BoolV(const BoolVArg a)
|
||||
{
|
||||
return internalSimd::m128_F2I(a);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecU32V VecU32V_From_BoolV(const BoolVArg a)
|
||||
{
|
||||
return a;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_Merge(const VecI32VArg x, const VecI32VArg y, const VecI32VArg z, const VecI32VArg w)
|
||||
{
|
||||
const __m128 xw = _mm_move_ss(internalSimd::m128_I2F(y), internalSimd::m128_I2F(x)); // y, y, y, x
|
||||
const __m128 yz = _mm_move_ss(internalSimd::m128_I2F(z), internalSimd::m128_I2F(w)); // z, z, z, w
|
||||
return internalSimd::m128_F2I(_mm_shuffle_ps(xw, yz, _MM_SHUFFLE(0, 2, 1, 0)));
|
||||
}
|
||||
|
||||
/*
|
||||
template<int a> PX_FORCE_INLINE VecI32V V4ISplat()
|
||||
{
|
||||
VecI32V result;
|
||||
result.m128_i32[0] = a;
|
||||
result.m128_i32[1] = a;
|
||||
result.m128_i32[2] = a;
|
||||
result.m128_i32[3] = a;
|
||||
return result;
|
||||
}
|
||||
|
||||
template<PxU32 a> PX_FORCE_INLINE VecU32V V4USplat()
|
||||
{
|
||||
VecU32V result;
|
||||
result.m128_u32[0] = a;
|
||||
result.m128_u32[1] = a;
|
||||
result.m128_u32[2] = a;
|
||||
result.m128_u32[3] = a;
|
||||
return result;
|
||||
}
|
||||
*/
|
||||
|
||||
/*
|
||||
PX_FORCE_INLINE void V4U16StoreAligned(VecU16V val, VecU16V* address)
|
||||
{
|
||||
*address = val;
|
||||
}
|
||||
*/
|
||||
|
||||
PX_FORCE_INLINE void V4U32StoreAligned(VecU32V val, VecU32V* address)
|
||||
{
|
||||
*address = val;
|
||||
}
|
||||
|
||||
/*PX_FORCE_INLINE Vec4V V4LoadAligned(Vec4V* addr)
|
||||
{
|
||||
return *addr;
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4LoadUnaligned(Vec4V* addr)
|
||||
{
|
||||
return V4LoadU(reinterpret_cast<float*>(addr));
|
||||
}*/
|
||||
|
||||
PX_FORCE_INLINE Vec4V Vec4V_From_VecI32V(VecI32V in)
|
||||
{
|
||||
return _mm_cvtepi32_ps(in);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_From_Vec4V(Vec4V a)
|
||||
{
|
||||
return _mm_cvttps_epi32(a);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V Vec4V_ReinterpretFrom_VecU32V(VecU32V a)
|
||||
{
|
||||
return Vec4V(a);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V Vec4V_ReinterpretFrom_VecI32V(VecI32V a)
|
||||
{
|
||||
return internalSimd::m128_I2F(a);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecU32V VecU32V_ReinterpretFrom_Vec4V(Vec4V a)
|
||||
{
|
||||
return VecU32V(a);
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecI32V VecI32V_ReinterpretFrom_Vec4V(Vec4V a)
|
||||
{
|
||||
return internalSimd::m128_F2I(a);
|
||||
}
|
||||
|
||||
template <int index>
|
||||
PX_FORCE_INLINE VecU32V V4U32SplatElement(VecU32V a)
|
||||
{
|
||||
VecU32V result;
|
||||
result.m128_u32[0] = result.m128_u32[1] = result.m128_u32[2] = result.m128_u32[3] = a.m128_u32[index];
|
||||
return result;
|
||||
}
|
||||
|
||||
template <int index>
|
||||
PX_FORCE_INLINE Vec4V V4SplatElement(Vec4V a)
|
||||
{
|
||||
float* data = reinterpret_cast<float*>(&a);
|
||||
return V4Load(data[index]);
|
||||
}
|
||||
|
||||
/*PX_FORCE_INLINE Vec4V V4Ceil(const Vec4V in)
|
||||
{
|
||||
UnionM128 a(in);
|
||||
return V4LoadXYZW(PxCeil(a.m128_f32[0]), PxCeil(a.m128_f32[1]), PxCeil(a.m128_f32[2]), PxCeil(a.m128_f32[3]));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE Vec4V V4Floor(const Vec4V in)
|
||||
{
|
||||
UnionM128 a(in);
|
||||
return V4LoadXYZW(PxFloor(a.m128_f32[0]), PxFloor(a.m128_f32[1]), PxFloor(a.m128_f32[2]), PxFloor(a.m128_f32[3]));
|
||||
}
|
||||
|
||||
PX_FORCE_INLINE VecU32V V4ConvertToU32VSaturate(const Vec4V in, PxU32 power)
|
||||
{
|
||||
PX_ASSERT(power == 0 && "Non-zero power not supported in convertToU32VSaturate");
|
||||
PX_UNUSED(power); // prevent warning in release builds
|
||||
PxF32 ffffFFFFasFloat = PxF32(0xFFFF0000);
|
||||
UnionM128 a(in);
|
||||
VecU32V result;
|
||||
result.m128_u32[0] = PxU32(PxClamp<PxF32>((a).m128_f32[0], 0.0f, ffffFFFFasFloat));
|
||||
result.m128_u32[1] = PxU32(PxClamp<PxF32>((a).m128_f32[1], 0.0f, ffffFFFFasFloat));
|
||||
result.m128_u32[2] = PxU32(PxClamp<PxF32>((a).m128_f32[2], 0.0f, ffffFFFFasFloat));
|
||||
result.m128_u32[3] = PxU32(PxClamp<PxF32>((a).m128_f32[3], 0.0f, ffffFFFFasFloat));
|
||||
return result;
|
||||
}*/
|
||||
|
||||
} // namespace aos
|
||||
} // namespace physx
|
||||
|
||||
#endif // PXFOUNDATION_PXUNIXSSE2INLINEAOS_H
|
||||
Reference in New Issue
Block a user