How much of that is cached state based on the percentage of ram available?
How much of that is cached state based on the percentage of ram available?
Base it off of total sqft?
I’m struggling to see how someone would need a combined 40000 sqft of residential living space either…
Yeah I completely forgot about the consumer side of things. I was expecting there being Cisco iOS/FRR router configs, not a full web dashboard.
As someone who works with 100Gbps networking:
Nah, just grab the domain and redirect it to X. Watch him explode.
How good are the RISC-V vector instructions implementations IRL? I’ve never heard of them. My experience with ARM is that even on certain data center chips the performance gains are abyssal (when using highly optimized libraries such as dpdk)
Harder to write compilers for RISC? I would argue that CISC is much harder to design a compiler for.
That being said there’s a lack of standardized vector/streaming instructions in out-of-the-box RISC-V that may hurt performance, but compiler design wise it’s much easier to write a functional compiler than for the nightmare that is x86.
Oh nice! A new tool! Do you happen to know how this compares to win10privacy?
I assert that this tech is biased towards bears and racoons.
To anyone who doesn’t know who they are, here’s a nice piece of investigative journalism about them: https://newrepublic.com/article/176811/united-daughters-confederacy-racist-ladies
Join us at !longreads@sh.itjust.works !
A tumbleweed rolls in the distance…
Both Bluetooth and BLE are perfectly fine protocols. You won’t be able to design much for short distance with that much power savings otherwise. The main issue is that for any protocols like this you would most likely need to put it in the 2.4ghz unlicensed band. And that’s predominantly used by wifi these days.
I’m familiar with the Apollo retro-reflectors. Though in all seriousness I doubt a laser would provide a substantial amount of power (unless you have a specialty designed energy collector like in RFID)
*stares at the intern’s 400 line bash script*
There are totally more flexible options. Just don’t mind the front falling off. It’s totally normal!
Wouldn’t shining back be counterproductive for this? You want the solar panels to harness the energy, not returning it to sender
Simply changing the binary worked for me. Been more than 1 month and no migration issues.
It does still show gitea branding, however.
I believe it. Linux is not a good measure of efficiency (see kernel bypass tcp stacks, af_xdp, dpdk, spdk, etc). You can almost always make something more efficient/faster than Linux for a given task. The problem is doing that while having support for almost all hardware/configurations/uses cases under the sun.
My suggestion would be to try compiling the kernel locally.its highly likely the one packaged in your distro contains extensions that you don’t have. Doing a local native compile should rule that out pretty quickly without having to disable any additional features.
Look at the line with the asm_exc_invalid_op. That seems like a hardware fault caused by an invalid asm instruction to me. Either something wrong is being interpreted as an opcode (unlikely) or maybe the driver was compiled with extensions not available on the current machine.
OP, how old is your CPU? And how old is the nic you are using?
Edit: did you use a custom driver for the NIC? I’m looking at the Linux src and rt_mutex_schedule does not exist. Nevermind. Was checking 4.18 instead of 6.7. found it now. The bug is most likely inside a macro called preempt_disable(). Unfortunately most of the functions are pretty heavily inlined and architecture dependent so you won’t get much out of it. But it is likely any changes you made in terms of premption might also be causing the bug.
Not sure about GreaseMonkey, but V8 compiles JS to an IL.
Nodejs has an emit IL debugging feature to see the emitted IL code.