Jun. 27th, 2008

izard: (Default)
[livejournal.com profile] avlasov had guessed right: The main problem is "one parallelizes the emulation process, which is not possible too, because of data dependencies between instructions."
That's really the main caveat that ruins the assumption of "60% "scalability"". However, there is a well known approach to extracting parallelism in spite of data and control dependencies : OOO superscalar pipelined execution.

That means that if I take a cycle-accurate emulator e.g. PTLSim which imposes 99.9% overhead, manage to tackle overhead down to 80% by somewhat reducing accuracy and parallelize for "60% scalability" then than we are done.

Extracting as much IPC as possible from x86 code stream was already done by Intel engineers, and cycle accurate simulator should be parallelizable e.g. by running the code simulating different pipeline stages and execution units on different cores. Unfortunately the theoretical IPC limit of 4.0 and practical - just about 2.0 makes "60% scalability on 16 cores" absolutely impossible.

P.S. The statement above about IPC limit also has a major logical flaw in it :)

Profile

izard: (Default)
izard

August 2025

S M T W T F S
     12
3456789
10111213 141516
17181920212223
24252627282930
31      

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Aug. 19th, 2025 04:30 pm
Powered by Dreamwidth Studios