@stdlib/strided-base-cmap
Apply a unary function to a single-precision complex floating-point strided input array and assign results to a single-precision complex floating-point strided output array.
Apply a unary function to a single-precision complex floating-point strided input array and assign results to a single-precision complex floating-point strided output array.
Reinterpret a complex-valued floating-point array as a real-valued floating-point array having the same precision.
Reinterpret a Complex64Array as a Float32Array.
Apply a unary function to a single-precision floating-point strided input array according to a strided mask array and assign results to a single-precision floating-point strided output array.
Apply a binary function to single-precision floating-point strided input arrays according to a strided mask array and assign results to a single-precision floating-point strided output array.
Apply a unary function to a single-precision floating-point strided input array and assign results to a single-precision floating-point strided output array.
Apply a binary function to single-precision floating-point strided input arrays and assign results to a single-precision floating-point strided output array.
Test if two arguments are both Float32Arrays and have the same values.
Test if a value is a Float32Array.
Test if a value is a 2-dimensional ndarray-like object containing single-precision floating-point numbers.
Utilities for single-precision floating-point numbers.
Base utilities for single-precision floating-point numbers.
Base double-precision floating-point number assert functions.
Return an integer corresponding to the unbiased exponent of a single-precision floating-point number.
Create a single-precision floating-point number from a literal bit representation.
Return a boolean indicating if the sign bit for a single-precision floating-point number is on (true) or off (false).
Return an integer corresponding to the significand of a single-precision floating-point number.
Return a string giving the literal bit representation of a single-precision floating-point number.
Convert a single-precision floating-point number to a signed 32-bit integer.
Convert a single-precision floating-point number to an unsigned 32-bit integer.