aboutsummaryrefslogtreecommitdiff
path: root/libs/libc
diff options
context:
space:
mode:
Diffstat (limited to 'libs/libc')
-rw-r--r--libs/libc/inc/math.h36
-rw-r--r--libs/libc/inc/vec.h38
-rw-r--r--libs/libc/math.c154
-rw-r--r--libs/libc/print.c19
4 files changed, 175 insertions, 72 deletions
diff --git a/libs/libc/inc/math.h b/libs/libc/inc/math.h
index 813ebf1..4d21e88 100644
--- a/libs/libc/inc/math.h
+++ b/libs/libc/inc/math.h
@@ -17,21 +17,25 @@
#define M_PI_2 1.57079632679489661923
#define M_PI_4 0.78539816339744830962
-f32 powf(f32 base, f32 exp);
-f64 pow(f64 base, f64 exp);
-f32 sqrtf(f32 num);
-f64 sqrt(f64 num);
-
-f32 lerpf(f32 from, f32 to, f32 trans);
-f64 lerp(f64 from, f64 to, f64 trans);
-f32 blerpf(f32 a, f32 b, f32 c, f32 d, f32 transx, f32 transy);
-f64 blerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy);
-
-f32 sinf(f32 angle);
-f64 sin(f64 angle);
-f32 cosf(f32 angle);
-f64 cos(f64 angle);
-f32 tanf(f32 angle);
-f64 tan(f64 angle);
+f64 mceil(f64 x);
+f64 mfloor(f64 x);
+
+f64 mexp(f64 exp);
+f64 mexp2(f64 exp);
+
+f64 mlog(f64 x);
+f64 mlog2(f64 x);
+
+f64 mpow(f64 base, f64 exp);
+f64 msqrt(f64 num);
+
+f64 mcubic(f64 x, f64 a, f64 b, f64 c, f64 d);
+
+f64 mlerp(f64 from, f64 to, f64 trans);
+f64 mblerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy);
+
+f64 msin(f64 angle);
+f64 mcos(f64 angle);
+f64 mtan(f64 angle);
#endif
diff --git a/libs/libc/inc/vec.h b/libs/libc/inc/vec.h
index bcd4cca..b2755ae 100644
--- a/libs/libc/inc/vec.h
+++ b/libs/libc/inc/vec.h
@@ -5,6 +5,10 @@
#include <def.h>
+/**
+ * Unsigned 32
+ */
+
typedef struct vec2 {
u32 x, y;
} vec2;
@@ -35,4 +39,38 @@ typedef struct vec3 {
((vec3){ a.y * b.z - a.z * b.y, a.z * b.x - a.x * b.z, a.x * b.y - a.y * b.x })
#define vec3_sum(a) ((u32)(a.x + a.y + a.z))
+/**
+ * Floating 32
+ */
+
+typedef struct vec2f {
+ f64 x, y;
+} vec2f;
+
+typedef struct vec3f {
+ f64 x, y, z;
+} vec3f;
+
+#define vec2f(x, y) ((vec2f){ (x), (y) })
+#define vec2fto3f(a, z) ((vec3f){ a.x, a.y, (z) })
+#define vec2f_add(a, b) ((vec2f){ a.x + b.x, a.y + b.y })
+#define vec2f_sub(a, b) ((vec2f){ a.x - b.x, a.y - b.y })
+#define vec2f_mul(a, b) ((vec2f){ a.x * (b), a.y * (b) })
+#define vec2f_div(a, b) ((vec2f){ a.x / (b), a.y / (b) })
+#define vec2f_dot(a, b) ((f64)(a.x * b.x + a.y * b.y))
+#define vec2f_eq(a, b) (a.x == b.x && a.y == b.y)
+#define vec2f_sum(a) ((f64)(a.x + a.y))
+
+#define vec3f(x, y, z) ((vec3f){ (x), (y), (z) })
+#define vec3fto2f(a) ((vec2f){ a.x, a.y })
+#define vec3f_add(a, b) ((vec3f){ a.x + b.x, a.y + b.y, a.z + b.z })
+#define vec3f_sub(a, b) ((vec3f){ a.x - b.x, a.y - b.y, a.z - b.z })
+#define vec3f_mul(a, b) ((vec3f){ a.x * (b), a.y * (b), a.z * (b) })
+#define vec3f_div(a, b) ((vec3f){ a.x / (b), a.y / (b), a.z / (b) })
+#define vec3f_dot(a, b) ((f64f)(a.x * b.x + a.y * b.y + a.z * b.z))
+#define vec3f_eq(a, b) (a.x == b.x && a.y == b.y && a.z == c.z)
+#define vec3f_cross(a, b) \
+ ((vec3f){ a.y * b.z - a.z * b.y, a.z * b.x - a.x * b.z, a.x * b.y - a.y * b.x })
+#define vec3f_sum(a) ((f64)(a.x + a.y + a.z))
+
#endif
diff --git a/libs/libc/math.c b/libs/libc/math.c
index 606281c..4d31e04 100644
--- a/libs/libc/math.c
+++ b/libs/libc/math.c
@@ -2,104 +2,148 @@
#include <math.h>
-f32 powf(f32 base, f32 exp)
+f64 mceil(f64 x)
{
- return (f32)pow(base, exp);
+ if (x == 0.0)
+ return x;
+
+ f64 out;
+ __asm__ volatile("frndint\n" : "=t"(out) : "0"(x));
+ if (out < x)
+ return out + 1.0;
+ return out;
}
-f64 pow(f64 base, f64 exp)
+f64 mfloor(f64 x)
{
+ if (x == 0.0)
+ return x;
+
f64 out;
- __asm__ volatile("fyl2x;"
- "fld %%st;"
- "frndint;"
- "fsub %%st,%%st(1);"
- "fxch;"
- "fchs;"
- "f2xm1;"
- "fld1;"
- "faddp;"
- "fxch;"
- "fld1;"
- "fscale;"
- "fstp %%st(1);"
- "fmulp;"
- : "=t"(out)
- : "0"(base), "u"(exp)
- : "st(1)");
+ __asm__ volatile("frndint\n" : "=t"(out) : "0"(x));
+ if (out > x)
+ return out - 1.0;
return out;
}
-// TODO: More efficient sqrt?
-
-f32 sqrtf(f32 num)
+f64 mexp(f64 exp)
{
- return powf(num, .5);
+ f64 out;
+ __asm__ volatile("fldl2e\n"
+ "fmulp\n"
+ "fld1\n"
+ "fld %%st(1)\n"
+ "fprem\n"
+ "f2xm1\n"
+ "faddp\n"
+ "fscale\n"
+ "fstp %%st(1)"
+ : "=t"(out)
+ : "0"(exp));
+ return out;
}
-f64 sqrt(f64 num)
+f64 mexp2(f64 exp)
{
- return pow(num, .5);
+ f64 out;
+ __asm__ volatile("fld1\n"
+ "fld %%st(1)\n"
+ "fprem\n"
+ "f2xm1\n"
+ "faddp\n"
+ "fscale\n"
+ "fstp %%st(1)"
+ : "=t"(out)
+ : "0"(exp));
+ return out;
}
-/**
- * Interpolations
- */
+f64 mlog(f64 x)
+{
+ f64 out;
+ __asm__ volatile("fldln2\n"
+ "fld %%st(1)\n"
+ "fyl2x\n"
+ "fstp %%st(1)"
+ : "=t"(out)
+ : "0"(x));
+ return out;
+}
-f32 lerpf(f32 from, f32 to, f32 trans)
+f64 mlog2(f64 x)
{
- return from + (to - from) * trans;
+ f64 out;
+ __asm__ volatile("fld1\n"
+ "fld %%st(1)\n"
+ "fyl2x\n"
+ "fstp %%st(1)"
+ : "=t"(out)
+ : "0"(x));
+ return out;
}
-f64 lerp(f64 from, f64 to, f64 trans)
+f64 mpow(f64 base, f64 exp)
{
- return from + (to - from) * trans;
+ if (exp == 0)
+ return 1;
+ if (exp == 1)
+ return base;
+ if (base == 0)
+ return 0;
+ if (exp == (f64)((s32)exp)) {
+ f64 out = base;
+ for (u32 i = 0; i < FABS(exp) - 1; i++)
+ out *= base;
+ if (exp < 0)
+ out = 1.0 / out;
+ return out;
+ }
+ return mexp2(exp * mlog2(base));
}
-f32 blerpf(f32 a, f32 b, f32 c, f32 d, f32 transx, f32 transy)
+// TODO: More efficient sqrt?
+
+f64 msqrt(f64 num)
{
- return lerpf(lerpf(a, b, transx), lerpf(c, d, transx), transy);
+ return mpow(num, .5);
}
-f64 blerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy)
+f64 mcubic(f64 x, f64 a, f64 b, f64 c, f64 d)
{
- return lerp(lerp(a, b, transx), lerp(c, d, transx), transy);
+ return a * mpow(x, 3) + b * mpow(x, 2) + c * x + d;
}
/**
- * Trigonometric functions
+ * Interpolations
*/
-f32 sinf(f32 angle)
+f64 mlerp(f64 from, f64 to, f64 trans)
{
- f32 ret = 0.0;
- __asm__ volatile("fsin" : "=t"(ret) : "0"(angle));
- return ret;
+ return from + (to - from) * trans;
}
-f64 sin(f64 angle)
+f64 mblerp(f64 a, f64 b, f64 c, f64 d, f64 transx, f64 transy)
{
- f64 ret = 0.0;
- __asm__ volatile("fsin" : "=t"(ret) : "0"(angle));
- return ret;
+ return mlerp(mlerp(a, b, transx), mlerp(c, d, transx), transy);
}
-f32 cosf(f32 angle)
-{
- return sinf(angle + (f32)M_PI_2);
-}
+/**
+ * Trigonometric functions
+ */
-f64 cos(f64 angle)
+f64 msin(f64 angle)
{
- return sin(angle + (f64)M_PI_2);
+ f64 ret = 0.0;
+ __asm__ volatile("fsin" : "=t"(ret) : "0"(angle));
+ return ret;
}
-f32 tanf(f32 angle)
+f64 mcos(f64 angle)
{
- return (f32)tan(angle);
+ return msin(angle + (f64)M_PI_2);
}
-f64 tan(f64 angle)
+f64 mtan(f64 angle)
{
f64 ret = 0.0, one;
__asm__ volatile("fptan" : "=t"(one), "=u"(ret) : "0"(angle));
diff --git a/libs/libc/print.c b/libs/libc/print.c
index 00b47c3..4ad1021 100644
--- a/libs/libc/print.c
+++ b/libs/libc/print.c
@@ -23,10 +23,12 @@ int vsnprintf(char *str, u32 size, const char *format, va_list ap)
int temp_int;
char temp_ch;
char *temp_str;
- f64 temp_double;
+ double temp_double;
+ vec2 temp_vec2;
char buffer[64] = { 0 };
+ // TODO: Fix potential memory overflows because of str[length++]=xxx
char ch;
while ((ch = *format++)) {
if (ch == '%') {
@@ -42,11 +44,26 @@ int vsnprintf(char *str, u32 size, const char *format, va_list ap)
temp_str = va_arg(ap, char *);
length += strlcpy(&str[length], temp_str, size - length);
break;
+ case 'b':
+ temp_int = va_arg(ap, int);
+ itoa(temp_int, buffer, 2);
+ length += strlcpy(&str[length], buffer, size - length);
+ break;
case 'd':
temp_int = va_arg(ap, int);
itoa(temp_int, buffer, 10);
length += strlcpy(&str[length], buffer, size - length);
break;
+ case 'v':
+ str[length++] = '(';
+ temp_vec2 = va_arg(ap, vec2);
+ itoa(temp_vec2.x, buffer, 10);
+ length += strlcpy(&str[length], buffer, size - length);
+ str[length++] = '/';
+ itoa(temp_vec2.y, buffer, 10);
+ length += strlcpy(&str[length], buffer, size - length);
+ str[length++] = ')';
+ break;
case 'x':
temp_int = va_arg(ap, int);
itoa(temp_int, buffer, 16);