diff options
Diffstat (limited to 'libs/libc')
-rw-r--r-- | libs/libc/inc/math.h | 36 | ||||
-rw-r--r-- | libs/libc/inc/vec.h | 38 | ||||
-rw-r--r-- | libs/libc/math.c | 154 | ||||
-rw-r--r-- | libs/libc/print.c | 19 |
4 files changed, 175 insertions, 72 deletions
diff --git a/libs/libc/inc/math.h b/libs/libc/inc/math.h index 813ebf1..4d21e88 100644 --- a/libs/libc/inc/math.h +++ b/libs/libc/inc/math.h @@ -17,21 +17,25 @@ #define M_PI_2 1.57079632679489661923 #define M_PI_4 0.78539816339744830962 -f32 powf(f32 base, f32 exp); -f64 pow(f64 base, f64 exp); -f32 sqrtf(f32 num); -f64 sqrt(f64 num); - -f32 lerpf(f32 from, f32 to, f32 trans); -f64 lerp(f64 from, f64 to, f64 trans); -f32 blerpf(f32 a, f32 b, f32 c, f32 d, f32 transx, f32 transy); -f64 blerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy); - -f32 sinf(f32 angle); -f64 sin(f64 angle); -f32 cosf(f32 angle); -f64 cos(f64 angle); -f32 tanf(f32 angle); -f64 tan(f64 angle); +f64 mceil(f64 x); +f64 mfloor(f64 x); + +f64 mexp(f64 exp); +f64 mexp2(f64 exp); + +f64 mlog(f64 x); +f64 mlog2(f64 x); + +f64 mpow(f64 base, f64 exp); +f64 msqrt(f64 num); + +f64 mcubic(f64 x, f64 a, f64 b, f64 c, f64 d); + +f64 mlerp(f64 from, f64 to, f64 trans); +f64 mblerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy); + +f64 msin(f64 angle); +f64 mcos(f64 angle); +f64 mtan(f64 angle); #endif diff --git a/libs/libc/inc/vec.h b/libs/libc/inc/vec.h index bcd4cca..b2755ae 100644 --- a/libs/libc/inc/vec.h +++ b/libs/libc/inc/vec.h @@ -5,6 +5,10 @@ #include <def.h> +/** + * Unsigned 32 + */ + typedef struct vec2 { u32 x, y; } vec2; @@ -35,4 +39,38 @@ typedef struct vec3 { ((vec3){ a.y * b.z - a.z * b.y, a.z * b.x - a.x * b.z, a.x * b.y - a.y * b.x }) #define vec3_sum(a) ((u32)(a.x + a.y + a.z)) +/** + * Floating 32 + */ + +typedef struct vec2f { + f64 x, y; +} vec2f; + +typedef struct vec3f { + f64 x, y, z; +} vec3f; + +#define vec2f(x, y) ((vec2f){ (x), (y) }) +#define vec2fto3f(a, z) ((vec3f){ a.x, a.y, (z) }) +#define vec2f_add(a, b) ((vec2f){ a.x + b.x, a.y + b.y }) +#define vec2f_sub(a, b) ((vec2f){ a.x - b.x, a.y - b.y }) +#define vec2f_mul(a, b) ((vec2f){ a.x * (b), a.y * (b) }) +#define vec2f_div(a, b) ((vec2f){ a.x / (b), a.y / (b) }) +#define vec2f_dot(a, b) ((f64)(a.x * b.x + a.y * b.y)) +#define vec2f_eq(a, b) (a.x == b.x && a.y == b.y) +#define vec2f_sum(a) ((f64)(a.x + a.y)) + +#define vec3f(x, y, z) ((vec3f){ (x), (y), (z) }) +#define vec3fto2f(a) ((vec2f){ a.x, a.y }) +#define vec3f_add(a, b) ((vec3f){ a.x + b.x, a.y + b.y, a.z + b.z }) +#define vec3f_sub(a, b) ((vec3f){ a.x - b.x, a.y - b.y, a.z - b.z }) +#define vec3f_mul(a, b) ((vec3f){ a.x * (b), a.y * (b), a.z * (b) }) +#define vec3f_div(a, b) ((vec3f){ a.x / (b), a.y / (b), a.z / (b) }) +#define vec3f_dot(a, b) ((f64f)(a.x * b.x + a.y * b.y + a.z * b.z)) +#define vec3f_eq(a, b) (a.x == b.x && a.y == b.y && a.z == c.z) +#define vec3f_cross(a, b) \ + ((vec3f){ a.y * b.z - a.z * b.y, a.z * b.x - a.x * b.z, a.x * b.y - a.y * b.x }) +#define vec3f_sum(a) ((f64)(a.x + a.y + a.z)) + #endif diff --git a/libs/libc/math.c b/libs/libc/math.c index 606281c..4d31e04 100644 --- a/libs/libc/math.c +++ b/libs/libc/math.c @@ -2,104 +2,148 @@ #include <math.h> -f32 powf(f32 base, f32 exp) +f64 mceil(f64 x) { - return (f32)pow(base, exp); + if (x == 0.0) + return x; + + f64 out; + __asm__ volatile("frndint\n" : "=t"(out) : "0"(x)); + if (out < x) + return out + 1.0; + return out; } -f64 pow(f64 base, f64 exp) +f64 mfloor(f64 x) { + if (x == 0.0) + return x; + f64 out; - __asm__ volatile("fyl2x;" - "fld %%st;" - "frndint;" - "fsub %%st,%%st(1);" - "fxch;" - "fchs;" - "f2xm1;" - "fld1;" - "faddp;" - "fxch;" - "fld1;" - "fscale;" - "fstp %%st(1);" - "fmulp;" - : "=t"(out) - : "0"(base), "u"(exp) - : "st(1)"); + __asm__ volatile("frndint\n" : "=t"(out) : "0"(x)); + if (out > x) + return out - 1.0; return out; } -// TODO: More efficient sqrt? - -f32 sqrtf(f32 num) +f64 mexp(f64 exp) { - return powf(num, .5); + f64 out; + __asm__ volatile("fldl2e\n" + "fmulp\n" + "fld1\n" + "fld %%st(1)\n" + "fprem\n" + "f2xm1\n" + "faddp\n" + "fscale\n" + "fstp %%st(1)" + : "=t"(out) + : "0"(exp)); + return out; } -f64 sqrt(f64 num) +f64 mexp2(f64 exp) { - return pow(num, .5); + f64 out; + __asm__ volatile("fld1\n" + "fld %%st(1)\n" + "fprem\n" + "f2xm1\n" + "faddp\n" + "fscale\n" + "fstp %%st(1)" + : "=t"(out) + : "0"(exp)); + return out; } -/** - * Interpolations - */ +f64 mlog(f64 x) +{ + f64 out; + __asm__ volatile("fldln2\n" + "fld %%st(1)\n" + "fyl2x\n" + "fstp %%st(1)" + : "=t"(out) + : "0"(x)); + return out; +} -f32 lerpf(f32 from, f32 to, f32 trans) +f64 mlog2(f64 x) { - return from + (to - from) * trans; + f64 out; + __asm__ volatile("fld1\n" + "fld %%st(1)\n" + "fyl2x\n" + "fstp %%st(1)" + : "=t"(out) + : "0"(x)); + return out; } -f64 lerp(f64 from, f64 to, f64 trans) +f64 mpow(f64 base, f64 exp) { - return from + (to - from) * trans; + if (exp == 0) + return 1; + if (exp == 1) + return base; + if (base == 0) + return 0; + if (exp == (f64)((s32)exp)) { + f64 out = base; + for (u32 i = 0; i < FABS(exp) - 1; i++) + out *= base; + if (exp < 0) + out = 1.0 / out; + return out; + } + return mexp2(exp * mlog2(base)); } -f32 blerpf(f32 a, f32 b, f32 c, f32 d, f32 transx, f32 transy) +// TODO: More efficient sqrt? + +f64 msqrt(f64 num) { - return lerpf(lerpf(a, b, transx), lerpf(c, d, transx), transy); + return mpow(num, .5); } -f64 blerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy) +f64 mcubic(f64 x, f64 a, f64 b, f64 c, f64 d) { - return lerp(lerp(a, b, transx), lerp(c, d, transx), transy); + return a * mpow(x, 3) + b * mpow(x, 2) + c * x + d; } /** - * Trigonometric functions + * Interpolations */ -f32 sinf(f32 angle) +f64 mlerp(f64 from, f64 to, f64 trans) { - f32 ret = 0.0; - __asm__ volatile("fsin" : "=t"(ret) : "0"(angle)); - return ret; + return from + (to - from) * trans; } -f64 sin(f64 angle) +f64 mblerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy) { - f64 ret = 0.0; - __asm__ volatile("fsin" : "=t"(ret) : "0"(angle)); - return ret; + return mlerp(mlerp(a, b, transx), mlerp(c, d, transx), transy); } -f32 cosf(f32 angle) -{ - return sinf(angle + (f32)M_PI_2); -} +/** + * Trigonometric functions + */ -f64 cos(f64 angle) +f64 msin(f64 angle) { - return sin(angle + (f64)M_PI_2); + f64 ret = 0.0; + __asm__ volatile("fsin" : "=t"(ret) : "0"(angle)); + return ret; } -f32 tanf(f32 angle) +f64 mcos(f64 angle) { - return (f32)tan(angle); + return msin(angle + (f64)M_PI_2); } -f64 tan(f64 angle) +f64 mtan(f64 angle) { f64 ret = 0.0, one; __asm__ volatile("fptan" : "=t"(one), "=u"(ret) : "0"(angle)); diff --git a/libs/libc/print.c b/libs/libc/print.c index 00b47c3..4ad1021 100644 --- a/libs/libc/print.c +++ b/libs/libc/print.c @@ -23,10 +23,12 @@ int vsnprintf(char *str, u32 size, const char *format, va_list ap) int temp_int; char temp_ch; char *temp_str; - f64 temp_double; + double temp_double; + vec2 temp_vec2; char buffer[64] = { 0 }; + // TODO: Fix potential memory overflows because of str[length++]=xxx char ch; while ((ch = *format++)) { if (ch == '%') { @@ -42,11 +44,26 @@ int vsnprintf(char *str, u32 size, const char *format, va_list ap) temp_str = va_arg(ap, char *); length += strlcpy(&str[length], temp_str, size - length); break; + case 'b': + temp_int = va_arg(ap, int); + itoa(temp_int, buffer, 2); + length += strlcpy(&str[length], buffer, size - length); + break; case 'd': temp_int = va_arg(ap, int); itoa(temp_int, buffer, 10); length += strlcpy(&str[length], buffer, size - length); break; + case 'v': + str[length++] = '('; + temp_vec2 = va_arg(ap, vec2); + itoa(temp_vec2.x, buffer, 10); + length += strlcpy(&str[length], buffer, size - length); + str[length++] = '/'; + itoa(temp_vec2.y, buffer, 10); + length += strlcpy(&str[length], buffer, size - length); + str[length++] = ')'; + break; case 'x': temp_int = va_arg(ap, int); itoa(temp_int, buffer, 16); |