[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
performance with >> 9 MW of memory
Date: Mon, 9 Dec 1991 16:06-0000
From: Scott McKay <SWM@SAPSUCKER.SCRC.Symbolics.COM>
Date: Fri, 6 Dec 1991 17:48 EST
From: firstname.lastname@example.org (Peter Paine)
Date: Fri, 6 Dec 1991 20:24-0000
From: SWM@SAPSUCKER.SCRC.Symbolics.COM (Scott McKay)
Date: Fri, 6 Dec 1991 14:42 EST
From: pan@ATHENA.PANGARO.dialnet.symbolics.com (Paul Pangaro)
I have heard second hand that performance of symbolics machines
(certainly this was about L machines, maybe G's too, that is part of the
question) "degrades" with more than 9 MW of memory. Anybody know
anything about this? I have no other info
Back in about 1986, a bug was discovered in the virtual memory system
that caused performance to degrade in machines with tons of memory.
It was fixed the same afternoon. I know of no such problem any more,
but the story of that ancient bug is sometimes told for its amusement
value. We have some substantially loaded machines in-house that have
the expected performance: fast.
How loaded is loaded?
What are the limits?
I dunno, something like 14.5MW on a G-machine (3620, 3650).
L-machines are smaller.
I-machines are bigger.
Is the correct equation for G-mach = (* 14.5MW 36/8) 65.25 Mbytes?
I see on the MacIvory II Show Machine Configuration:
NuBus slot C: National Semiconductor NS8/16 Memory Expansion Card p/n
NS8/16 (id #x10F) rev Rev. 2.21, 16MB total, 2.6MW in use
(* 2.6 40/8) = 13 Mbytes - where did the other 3 Mbytes go?
It would be very useful to know the actual/precise limits for memory