diff options
Diffstat (limited to 'vendor/modernc.org/libc/libc.go')
-rw-r--r-- | vendor/modernc.org/libc/libc.go | 293 |
1 files changed, 264 insertions, 29 deletions
diff --git a/vendor/modernc.org/libc/libc.go b/vendor/modernc.org/libc/libc.go index f122eba47..031e7119c 100644 --- a/vendor/modernc.org/libc/libc.go +++ b/vendor/modernc.org/libc/libc.go @@ -2,7 +2,7 @@ // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. -//go:build !(linux && (amd64 || loong64)) +//go:build !(linux && (amd64 || arm64 || loong64)) //go.generate echo package libc > ccgo.go //go:generate go fmt ./... @@ -1002,6 +1002,13 @@ func Xacos(t *TLS, x float64) float64 { return math.Acos(x) } +func Xacosf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Acos(float64(x))) +} + func Xacosh(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1009,6 +1016,13 @@ func Xacosh(t *TLS, x float64) float64 { return math.Acosh(x) } +func Xacoshf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Acosh(float64(x))) +} + func Xasin(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1016,6 +1030,13 @@ func Xasin(t *TLS, x float64) float64 { return math.Asin(x) } +func Xasinf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Asin(float64(x))) +} + func Xasinh(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1023,6 +1044,13 @@ func Xasinh(t *TLS, x float64) float64 { return math.Asinh(x) } +func Xasinhf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Asinh(float64(x))) +} + func Xatan(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1030,6 +1058,13 @@ func Xatan(t *TLS, x float64) float64 { return math.Atan(x) } +func Xatanf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Atan(float64(x))) +} + func Xatan2(t *TLS, x, y float64) float64 { if __ccgo_strace { trc("t=%v y=%v, (%v:)", t, y, origin(2)) @@ -1037,6 +1072,13 @@ func Xatan2(t *TLS, x, y float64) float64 { return math.Atan2(x, y) } +func Xatan2f(t *TLS, x, y float32) float32 { + if __ccgo_strace { + trc("t=%v y=%v, (%v:)", t, y, origin(2)) + } + return float32(math.Atan2(float64(x), float64(y))) +} + func Xatanh(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1044,6 +1086,13 @@ func Xatanh(t *TLS, x float64) float64 { return math.Atanh(x) } +func Xatanhf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Atanh(float64(x))) +} + func Xceil(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1093,6 +1142,13 @@ func Xcosh(t *TLS, x float64) float64 { return math.Cosh(x) } +func Xcoshf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Cosh(float64(x))) +} + func Xexp(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1100,6 +1156,13 @@ func Xexp(t *TLS, x float64) float64 { return math.Exp(x) } +func Xexpf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Exp(float64(x))) +} + func Xfabs(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1121,6 +1184,13 @@ func Xfloor(t *TLS, x float64) float64 { return math.Floor(x) } +func Xfloorf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Floor(float64(x))) +} + func Xfmod(t *TLS, x, y float64) float64 { if __ccgo_strace { trc("t=%v y=%v, (%v:)", t, y, origin(2)) @@ -1128,6 +1198,17 @@ func Xfmod(t *TLS, x, y float64) float64 { return math.Mod(x, y) } +func Xfmodf(t *TLS, x, y float32) float32 { + if __ccgo_strace { + trc("t=%v y=%v, (%v:)", t, y, origin(2)) + } + return float32(math.Mod(float64(x), float64(y))) +} + +func X__builtin_hypot(t *TLS, x float64, y float64) (r float64) { + return Xhypot(t, x, y) +} + func Xhypot(t *TLS, x, y float64) float64 { if __ccgo_strace { trc("t=%v y=%v, (%v:)", t, y, origin(2)) @@ -1135,6 +1216,13 @@ func Xhypot(t *TLS, x, y float64) float64 { return math.Hypot(x, y) } +func Xhypotf(t *TLS, x, y float32) float32 { + if __ccgo_strace { + trc("t=%v y=%v, (%v:)", t, y, origin(2)) + } + return float32(math.Hypot(float64(x), float64(y))) +} + func Xisnan(t *TLS, x float64) int32 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1170,6 +1258,13 @@ func Xlog(t *TLS, x float64) float64 { return math.Log(x) } +func Xlogf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Log(float64(x))) +} + func Xlog10(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1177,6 +1272,13 @@ func Xlog10(t *TLS, x float64) float64 { return math.Log10(x) } +func Xlog10f(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Log10(float64(x))) +} + func X__builtin_log2(t *TLS, x float64) float64 { return Xlog2(t, x) } @@ -1188,6 +1290,13 @@ func Xlog2(t *TLS, x float64) float64 { return math.Log2(x) } +func Xlog2f(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Log2(float64(x))) +} + func Xround(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1195,6 +1304,13 @@ func Xround(t *TLS, x float64) float64 { return math.Round(x) } +func Xroundf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Round(float64(x))) +} + func X__builtin_round(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1202,6 +1318,13 @@ func X__builtin_round(t *TLS, x float64) float64 { return math.Round(x) } +func X__builtin_roundf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Round(float64(x))) +} + func Xsin(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1223,6 +1346,13 @@ func Xsinh(t *TLS, x float64) float64 { return math.Sinh(x) } +func Xsinhf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Sinh(float64(x))) +} + func Xsqrt(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1230,6 +1360,13 @@ func Xsqrt(t *TLS, x float64) float64 { return math.Sqrt(x) } +func Xsqrtf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Sqrt(float64(x))) +} + func Xtan(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1237,6 +1374,13 @@ func Xtan(t *TLS, x float64) float64 { return math.Tan(x) } +func Xtanf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Tan(float64(x))) +} + func Xtanh(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1244,6 +1388,13 @@ func Xtanh(t *TLS, x float64) float64 { return math.Tanh(x) } +func Xtanhf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Tanh(float64(x))) +} + func Xtrunc(t *TLS, x float64) float64 { if __ccgo_strace { trc("t=%v x=%v, (%v:)", t, x, origin(2)) @@ -1251,6 +1402,13 @@ func Xtrunc(t *TLS, x float64) float64 { return math.Trunc(x) } +func Xtruncf(t *TLS, x float32) float32 { + if __ccgo_strace { + trc("t=%v x=%v, (%v:)", t, x, origin(2)) + } + return float32(math.Trunc(float64(x))) +} + var nextRand = uint64(1) // int rand(void); @@ -1273,6 +1431,13 @@ func Xpow(t *TLS, x, y float64) float64 { return r } +func Xpowf(t *TLS, x, y float32) float32 { + if __ccgo_strace { + trc("t=%v y=%v, (%v:)", t, y, origin(2)) + } + return float32(math.Pow(float64(x), float64(y))) +} + func Xfrexp(t *TLS, x float64, exp uintptr) float64 { if __ccgo_strace { trc("t=%v x=%v exp=%v, (%v:)", t, x, exp, origin(2)) @@ -1282,6 +1447,15 @@ func Xfrexp(t *TLS, x float64, exp uintptr) float64 { return f } +func Xfrexpf(t *TLS, x float32, exp uintptr) float32 { + if __ccgo_strace { + trc("t=%v x=%v exp=%v, (%v:)", t, x, exp, origin(2)) + } + f, e := math.Frexp(float64(x)) + *(*int32)(unsafe.Pointer(exp)) = int32(e) + return float32(f) +} + func Xmodf(t *TLS, x float64, iptr uintptr) float64 { if __ccgo_strace { trc("t=%v x=%v iptr=%v, (%v:)", t, x, iptr, origin(2)) @@ -1291,6 +1465,15 @@ func Xmodf(t *TLS, x float64, iptr uintptr) float64 { return f } +func Xmodff(t *TLS, x float32, iptr uintptr) float32 { + if __ccgo_strace { + trc("t=%v x=%v iptr=%v, (%v:)", t, x, iptr, origin(2)) + } + i, f := math.Modf(float64(x)) + *(*float32)(unsafe.Pointer(iptr)) = float32(i) + return float32(f) +} + // char *strncpy(char *dest, const char *src, size_t n) func Xstrncpy(t *TLS, dest, src uintptr, n types.Size_t) (r uintptr) { if __ccgo_strace { @@ -1465,39 +1648,91 @@ func Xstrrchr(t *TLS, s uintptr, c int32) (r uintptr) { } // void *memset(void *s, int c, size_t n) -func Xmemset(t *TLS, s uintptr, c int32, n types.Size_t) uintptr { +func Xmemset(t *TLS, dest uintptr, c int32, n types.Size_t) uintptr { if __ccgo_strace { - trc("t=%v s=%v c=%v n=%v, (%v:)", t, s, c, n, origin(2)) + trc("t=%v s=%v c=%v n=%v, (%v:)", t, dest, c, n, origin(2)) } - if n != 0 { - c := byte(c & 0xff) + var c8 uint8 + var c32 uint32 + var c64 uint64 + var k types.Size_t + var s uintptr - // This will make sure that on platforms where they are not equally aligned we - // clear out the first few bytes until allignment - bytesBeforeAllignment := s % unsafe.Alignof(uint64(0)) - if bytesBeforeAllignment > uintptr(n) { - bytesBeforeAllignment = uintptr(n) - } - b := (*RawMem)(unsafe.Pointer(s))[:bytesBeforeAllignment:bytesBeforeAllignment] - n -= types.Size_t(bytesBeforeAllignment) - for i := range b { - b[i] = c - } - if n >= 8 { - i64 := uint64(c) + uint64(c)<<8 + uint64(c)<<16 + uint64(c)<<24 + uint64(c)<<32 + uint64(c)<<40 + uint64(c)<<48 + uint64(c)<<56 - b8 := (*RawMem64)(unsafe.Pointer(s + bytesBeforeAllignment))[: n/8 : n/8] - for i := range b8 { - b8[i] = i64 - } - } - if n%8 != 0 { - b = (*RawMem)(unsafe.Pointer(s + bytesBeforeAllignment + uintptr(n-n%8)))[: n%8 : n%8] - for i := range b { - b[i] = c - } + s = dest + /* Fill head and tail with minimal branching. Each + * conditional ensures that all the subsequently used + * offsets are well-defined and in the dest region. */ + if n == 0 { + return dest + } + c8 = uint8(c) + *(*uint8)(unsafe.Pointer(s)) = c8 + *(*uint8)(unsafe.Pointer(s + uintptr(n-1))) = c8 + if n <= types.Size_t(2) { + return dest + } + *(*uint8)(unsafe.Pointer(s + 1)) = c8 + *(*uint8)(unsafe.Pointer(s + 2)) = c8 + *(*uint8)(unsafe.Pointer(s + uintptr(n-2))) = c8 + *(*uint8)(unsafe.Pointer(s + uintptr(n-3))) = c8 + if n <= types.Size_t(6) { + return dest + } + *(*uint8)(unsafe.Pointer(s + 3)) = c8 + *(*uint8)(unsafe.Pointer(s + uintptr(n-4))) = c8 + if n <= types.Size_t(8) { + return dest + } + /* Advance pointer to align it at a 4-byte boundary, + * and truncate n to a multiple of 4. The previous code + * already took care of any head/tail that get cut off + * by the alignment. */ + k = -types.Size_t(s) & types.Size_t(3) + s += uintptr(k) + n -= k + n &= types.Size_t(-Int32FromInt32(4)) + c32 = uint32(0x01010101) * uint32(c8) + /* In preparation to copy 32 bytes at a time, aligned on + * an 8-byte bounary, fill head/tail up to 28 bytes each. + * As in the initial byte-based head/tail fill, each + * conditional below ensures that the subsequent offsets + * are valid (e.g. !(n<=24) implies n>=28). */ + *(*uint32)(unsafe.Pointer(s + uintptr(0))) = c32 + *(*uint32)(unsafe.Pointer(s + uintptr(n-4))) = c32 + if n <= types.Size_t(8) { + return dest + } + c64 = uint64(c32) | (uint64(c32) << 32) + *(*uint64)(unsafe.Pointer(s + uintptr(4))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(n-12))) = c64 + if n <= types.Size_t(24) { + return dest + } + *(*uint64)(unsafe.Pointer(s + uintptr(12))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(20))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(n-28))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(n-20))) = c64 + /* Align to a multiple of 8 so we can fill 64 bits at a time, + * and avoid writing the same bytes twice as much as is + * practical without introducing additional branching. */ + k = types.Size_t(24) + types.Size_t(s)&types.Size_t(4) + s += uintptr(k) + n -= k + /* If this loop is reached, 28 tail bytes have already been + * filled, so any remainder when n drops below 32 can be + * safely ignored. */ + for { + if !(n >= types.Size_t(32)) { + break } + *(*uint64)(unsafe.Pointer(s + uintptr(0))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(8))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(16))) = c64 + *(*uint64)(unsafe.Pointer(s + uintptr(24))) = c64 + n -= types.Size_t(32) + s += uintptr(32) } - return s + return dest } // void *memcpy(void *dest, const void *src, size_t n); |