Attached: 1 image
Wait, 20 *milliseconds*? Either their kernel scheduler config is completely out of whack, or ARM/Qualcomm really screwed this one up.
Apple cores can boost to max in around 50 *micro* seconds. 20 milliseconds is just broken. That's more than one frame, and that's how you get janky UI response and dropped frames. I hope this was a config issue and these cores/designs really don't take 20ms to increase clocks (I could see that with a really bad regulator/power delivery system...)
From: https://chipsandcheese.com/2023/08/11/arms-cortex-a710-winning-by-default/
Android doesn’t use Java at a byte-code level and never has, as far as I can tell. Source code was written in Java since mobile developers were so used to it but Android never ran the JVM, they do their own thing with Java source.
You can dislike Java syntax but the software stack on Android wasn’t Java’s.
They compile Java Bytecode to Dalvik Bytecode and run that on the Android Runtime which is a tiered JIT compiler.
It still inherits the issues of Java such as the GC, no stack allocated value types, poor cache locality, etc. Although tbf the GC on Android is pretty fucking good these days and doesn’t pause the world anymore.
Interesting, I always thought it had to do with Android’s ungodly software stack which at some point involves, of all things, fucking java.
Android doesn’t use Java at a byte-code level and never has, as far as I can tell. Source code was written in Java since mobile developers were so used to it but Android never ran the JVM, they do their own thing with Java source.
You can dislike Java syntax but the software stack on Android wasn’t Java’s.
They compile Java Bytecode to Dalvik Bytecode and run that on the Android Runtime which is a tiered JIT compiler.
It still inherits the issues of Java such as the GC, no stack allocated value types, poor cache locality, etc. Although tbf the GC on Android is pretty fucking good these days and doesn’t pause the world anymore.