/* Copyright (c) 2018-2019, tevador All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include "common.hpp" #include "intrin_portable.h" #include "blake2/endian.h" #if defined(__SIZEOF_INT128__) typedef unsigned __int128 uint128_t; typedef __int128 int128_t; uint64_t mulh(uint64_t a, uint64_t b) { return ((uint128_t)a * b) >> 64; } int64_t smulh(int64_t a, int64_t b) { return ((int128_t)a * b) >> 64; } #define HAVE_MULH #define HAVE_SMULH #endif #if defined(_MSC_VER) #define HAS_VALUE(X) X ## 0 #define EVAL_DEFINE(X) HAS_VALUE(X) #include #include uint64_t rotl(uint64_t x, int c) { return _rotl64(x, c); } uint64_t rotr(uint64_t x , int c) { return _rotr64(x, c); } #define HAVE_ROTL #define HAVE_ROTR #if EVAL_DEFINE(__MACHINEARM64_X64(1)) uint64_t mulh(uint64_t a, uint64_t b) { return __umulh(a, b); } #define HAVE_MULH #endif #if EVAL_DEFINE(__MACHINEX64(1)) int64_t smulh(int64_t a, int64_t b) { int64_t hi; _mul128(a, b, &hi); return hi; } #define HAVE_SMULH #endif static void setRoundMode_(uint32_t mode) { _controlfp(mode, _MCW_RC); } #define HAVE_SETROUNDMODE_IMPL #endif #ifndef HAVE_SETROUNDMODE_IMPL static void setRoundMode_(uint32_t mode) { fesetround(mode); } #endif #ifndef HAVE_ROTR uint64_t rotr(uint64_t a, int b) { return (a >> b) | (a << (64 - b)); } #define HAVE_ROTR #endif #ifndef HAVE_ROTL uint64_t rotl(uint64_t a, int b) { return (a << b) | (a >> (64 - b)); } #define HAVE_ROTL #endif #ifndef HAVE_MULH #define LO(x) ((x)&0xffffffff) #define HI(x) ((x)>>32) uint64_t mulh(uint64_t a, uint64_t b) { uint64_t ah = HI(a), al = LO(a); uint64_t bh = HI(b), bl = LO(b); uint64_t x00 = al * bl; uint64_t x01 = al * bh; uint64_t x10 = ah * bl; uint64_t x11 = ah * bh; uint64_t m1 = LO(x10) + LO(x01) + HI(x00); uint64_t m2 = HI(x10) + HI(x01) + LO(x11) + HI(m1); uint64_t m3 = HI(x11) + HI(m2); return (m3 << 32) + LO(m2); } #define HAVE_MULH #endif #ifndef HAVE_SMULH int64_t smulh(int64_t a, int64_t b) { int64_t hi = mulh(a, b); if (a < 0LL) hi -= b; if (b < 0LL) hi -= a; return hi; } #define HAVE_SMULH #endif #ifdef RANDOMX_DEFAULT_FENV void rx_reset_float_state() { setRoundMode_(FE_TONEAREST); rx_set_double_precision(); //set precision to 53 bits if needed by the platform } void rx_set_rounding_mode(uint32_t mode) { switch (mode & 3) { case RoundDown: setRoundMode_(FE_DOWNWARD); break; case RoundUp: setRoundMode_(FE_UPWARD); break; case RoundToZero: setRoundMode_(FE_TOWARDZERO); break; case RoundToNearest: setRoundMode_(FE_TONEAREST); break; default: UNREACHABLE; } } #endif #ifdef RANDOMX_USE_X87 #ifdef _M_IX86 void rx_set_double_precision() { _control87(_PC_53, _MCW_PC); } #elif defined(__i386) void rx_set_double_precision() { uint16_t volatile x87cw; asm volatile("fstcw %0" : "=m" (x87cw)); x87cw &= ~0x300; x87cw |= 0x200; asm volatile("fldcw %0" : : "m" (x87cw)); } #endif #endif //RANDOMX_USE_X87 union double_ser_t { double f; uint64_t i; }; double loadDoublePortable(const void* addr) { double_ser_t ds; ds.i = load64(addr); return ds.f; }