Static analysis would probably work in this case because the in-order core would be very GPU-like while the other core would not.
In cases where performance characteristics are closer, the OS could switch cores, monitor the runtimes, and add metadata about which core worked best (potentially even about which core worked best at which times).
There's no point in shuffling background processes to a subset of cores if it doesn't provide power-savings.
It sounds like upcoming Intel chips will have the ability to run efficiency-mode on some cores, at which case OS-level code to do shuffling makes sense.
Well it would make some sense to have perhaps 2 or 4 single core speed focused cores for the main process that you're currently working on, and then add as many of the smaller E cores as you can, to handle the hundreds of background processes so that the P cores can run uninterrupted by them.
If this works that way in practice is anyone's guess though.
One option is to have multiple cores dedicated to the same workload and then switch between them as they heat up. There are many issues with this approach which make it impractical, but if your really do want speed at any cost then it may be worth it.
Modern CPUs have a very large bag of tricks to make the code they execute run fast. These tricks are usually designed for some optimal workload, and (especially x86) the real deal can't actually keep the backend busy. Solution: Duplicate some of the CPU to feed the backend with more stuff to do. This can get you a performance win without the cost of a whole core.
All good so far, for (say) HPC where you have people with guns guarding the machines you can turn on all the performance. The issue is that as CPUs have got more complicated, it's increasingly easy to attack the internal state of the processor. This is already possible in a single thread (i.e. the prototypical spectre and meltdown implementations), however imagine what you can do next door to a completely different "core". The processor maintains separation of structure but not of state, so you can use (say) timing side channels to extract information about what your neighbour is up to.
FWIW these issues are more of a problem for Cloud vendors than you or I as they usually require pretty specific knowledge of the hardware being attacked and even then are not easy to pull off.
It might actually be useful to have a second core, to service hardware and peripherals. This way you'd have smoother performance, because you would not need to flush the cache on every interrupt.
Not really, at least with how it currently works. Most OSes will redirect threads to the different cores based on their resource usage, which is impossible if each core was running a different architecture. No amount of magic will make it work elegantly, I'm afraid.
I wonder if certain applications would benefit from redundant setups in different cpu arches to check for discrepancies.. kinda like NASA stuff but different.
You could probably do something like using idle cores (or idle hyperthreads) to duplicate instructions on an opportunistic basis to verify outcomes on a less than complete basis. There would be thermal and power consequences but some situations care about that less than others.
reply