Computer Chess Club Archives


Search

Terms

Messages

Subject: Re: Who can update about new 64 bits chip?

Author: Matthew Hull

Date: 10:03:39 08/27/02

Go up one level in this thread


On August 27, 2002 at 11:28:43, Robert Hyatt wrote:

>On August 27, 2002 at 09:42:15, Matthew Hull wrote:
>
>>On August 26, 2002 at 22:36:13, Robert Hyatt wrote:
>>
>>>On August 26, 2002 at 21:33:18, Jeremiah Penery wrote:
>>>
>>>>On August 26, 2002 at 15:06:05, Robert Hyatt wrote:
>>>>
>>>>>On August 26, 2002 at 13:12:42, Jeremiah Penery wrote:
>>>>
>>>>>>On August 26, 2002 at 11:16:47, Robert Hyatt wrote:
>>>>
>>>>>>As I said, much of that time must come from the long travel path, slow memory
>>>>>>controller, etc.
>>>>>
>>>>>We just disagree on where the delay is.  I believe that at _least_ 80% of
>>>>>the latency is in the chip.  Not in the controller/bus...
>>>>
>>>>>>If you see 140ns today (average), you don't believe that almost half of that
>>>>>>latency is caused by the travel path from CPU/controller/memory and back?  If
>>>>>>the memory controller runs at bus speed (133MHz), it has 7.5ns/clock cycle.
>>>>>>That alone is significant latency added to the process.
>>>>>
>>>>>I don't believe it, no.  I believe that most of the latency is in the DRAM
>>>>>itself, not in the controller.  The controller has no "capacitors" to deal
>>>>>with, it is made up of SRAM buffers and some form of hardware logic (such
>>>>>as TTL) which means switching times are at the picosecond level.  It takes
>>>>>a _bunch_ of picoseconds to add up to a nanosecond...
>>>>
>>>>The clock cycle of the memory controller is some 7.5ns.  It can only send one
>>>>request/clock, AFAIK.  That is already adding significant latency.
>>>
>>>
>>>OK... same issue I talk about in parallel programming classes...  Amdahl's
>>>Law, slightly paraphrased to fit here.  If the total access time is > 100ns,
>>>and the controller has a 7.5ns cycle time,  what happens when you design a
>>>controller that takes _zero_ nanoseconds?  Answer:  you are _still_ near
>>>100ns in total access time as you only eliminated 7.5...
>>>
>>>That is the problem I see.  DRAM is simply inherently slow in terms of
>>>latency.  It is hard to take a static charge, shunt it somewhere to measure
>>>it, and do that quickly.  Because of resistance, capacitance and inductance
>>>as I mentioned earlier..  IE replace a capacitor with a balloon filled with
>>>water.  It is _hard_ to get that water out and over to something that can
>>>measure it, when you have millions of balloons sharing common water tubes...
>>>
>>>
>>>
>>>
>>>
>>>>
>>>>>>Even a few 10s reduces your number of 120ns to the claimed 80ns of Hammer. :)
>>>>>
>>>>>I'll believe 80 when I actually get my hands on it.  :)  Because that will
>>>>>be faster than any Cray ever made (that used DRAM, older crays used bipolar
>>>>>but the memory was not nearly as dense).
>>>>
>>>>>I was talking about cray from the perspective that they have never had an 80ns
>>>>>memory access time.  It has _always_ been over 100 since they moved away from
>>>>>bipolar memory to DRAM for density.  And their controllers have >_never_ "sucked"
>>>>
>>>>It's difficult to find really accurate data on this.  I've read more than a few
>>>>different things.  But from what I can tell, the latency (cycle time) of DRAM c.
>>>>1994 was on the order of 100ns. (In 1980, it was 250ns; 1989 was nearer 170ns).
>>>>It hasn't been lowered at that pace since then, but it has gotten lower.  As
>>>>I've said, current figures I've seen place it at 70ns today.  _Any additional
>>>>latencies_ seen are/were caused by controller, path length, and whatever else.
>>>
>>>
>>>I don't know of any way to lower latency.  How do you make a charge leave a
>>>capacitor faster?  Maybe at super-conducting temps, perhaps.  But you have to
>>>deal with the magnetic field that has to propogate as the current attempts to
>>>move along a wire...  Smaller distances can help _some_.  But this is a "square"
>>>problem.  Not a linear problem.
>>>
>>>
>>>
>>>
>>>>
>>>>If you don't think path length can influence latency very much, then you must
>>>>not talk about RDRAM having bad latency. :)  The only reason it has higher
>>>>latency is because of a much longer path length for the memory request signal.
>>>>(And sometimes the banks can go into sleep mode, and take a while to wake.)
>>>
>>>I understand how they chain things together and how it added to the serial gate
>>>count.  The problem with RDRAM is that they couldn't make the DRAM dump any
>>>faster, and they added more gates in an effort to improve streaming performance,
>>>which has nothing to do with random access...
>>>
>>>
>>>
>>>
>>>>
>>>>Cray machines probably have some additional issues because they're
>>>>super-multi-ported designs, with a lot of processors trying to concurrently
>>>>access the same memory banks. (I'm talking about their vector machines, not
>>>>stuff like the T3E, which is some kind of Alpha cluster.)
>>>
>>>They have lots of issues.  But even going back to single-cpu machines, latency
>>>was no better than it is today...  and with every new generation of cpu, the
>>>memory fell that much farther behind...  ie cycletime of cpu * number of cycles
>>>to do a read has effectively been constant since they first went from Bipolar to
>>>DRAM.
>>
>>
>>How much is latency improved with bipolar?
>>
>>Also, what is the future of RAM.  Now it seems hopelessly behind, and getting
>>"behinder".
>
>
>Bipolar has no actual read latency.  So all the latency is in moving the
>data around...  IE think "JK flip-flop" and you'll get the idea.  There is
>a zero/one value that is always available without having to dump a capacitor
>and rewrite it.  The only latency issues are latency from the point where the
>bit is presented to the point where the bit has to be moved...
>
>DRAM is definitely a slow item.  But it is also, by far, the most dense memory
>system we can build with reasonable access times.  Cray went that way so that
>they could build compact 32 gig memory systems.


Where I work, we still have some old bipolar mainframes.  But they still have
high speed buffers (is that the same thing we call cache on a PC?).  If memory
latency is not an issue, why the high speed buffers?

Also, why can't all memory be as fast as cache memory, or why can't all memory
just be cache memory?

Thanks.



This page took 0 seconds to execute

Last modified: Thu, 15 Apr 21 08:11:13 -0700

Current Computer Chess Club Forums at Talkchess. This site by Sean Mintz.