From: Robert Myers on
On Oct 25, 6:36 am, Mayan Moudgill <ma...(a)bestweb.net> wrote:
> Robert Myers wrote:

>
> > even if, in some rare awkward instances, it might
> > take a while.   Something about large prime numbers.
>
> I'm kind of lost here; which large prime numbers are involved? If you're
> talking loops with fractional II, then those might behave better on an
> OoO processors. Terje?

Poor Terje, to whom I really do owe an apology. I'm pretty sure this
part of the conversation makes sense only in the context of a
processor with a trace cache where (given my primitive understanding
of the Pentium 4 trace cache) in some circumstances it will sometimes
take a long time for the OoO scheduler to find a stable "optimal"
schedule for a given loop iteration.

You have kindly pointed out that there are other circumstances where
there are important differences between OoO and software pipelining.
It probably would have been best if the discussion had been more
careful from the very beginning.

Robert.

From: "Andy "Krazy" Glew" on
Bernd Paysan wrote:
> Del Cecchi wrote:
>> You could use SOI, no bulk. :-)
>
> There still is a bulk, there is just no substrate, so the bulk is left
> floating. The diodes I mentioned are sill there, supplying the bulk
> when forward biased (this is the well-known effect of SOI to have
> variable gate thresholds through charging and discharging the bulk below
> the diodes threshold, unless you add in a real bulk contact like on
> stock silicon wafers).
>
>> I don't get the point of the AC. Light bulbs and space heaters are AC
>> powered and still disipate power. What did I miss?
>
> I can't tell you. Andy apparently doesn't care much about the physics
> behind integrated circuits, his knowledge stops at the gate level. This
> is completely ok for digital design, but I wonder why he makes that sort
> of suggestions ;-).

Och! Wounded I am, ad-hominem'ed. What did I ever do to you, Bernd?
Have you been taking lessons from Nick?

Anyway, to address Del's question:

In present electronics, electrons flow from the power supply, across
lots of wires, to the switching circuit, across the circuit, and then
back to the power supply. (Or holes if you want to get pedantic, but if
you get pedantic, that's a whole other story behind holes.)
The idea behind charge recovery is that, instead of making this
round trip, you accumulate the charge that has flowed from high to low -
and then somehow find a way to do more work with it. Have it flow across
the circuit, get stored for a while, and then have it flow back again.
This flowing first one way, then back the other, is why I call it
AC. Alternating current (directions). This is not that common a
terminology, although as I explained in my original post I am not the
only person to think of it this way.

Bernd apparently thought that I was espousing having the power rails for
the entire chip reverse polarity - swapping, e.g. (+1,-1) or (+1,0).
This doesn't work, for reasons Bernd explained.
However, what I am talking about - what I have picked up from
researchers in this field - is to still have power and ground rails at
(1.5V,0) - but to have signal rails that are at (+1.3,+0.7). Then
signal rails periodically reverse, to (+0.7,+1.3).
This is why, in my original post, I was careful to say `The devices
will have to work within a range of changing voltages, say when "high"
is between k=1/3 and k=1 Vmax=(Vmean+k*Vswing), and "low" is between
k=1/3 and k=1 Vmin=(Vmean-k*Vswing).'
Where signal rails at (+0.7,+1.3) range, with power rails at
(+1.5,0), Vmean=1V, VSwing = 0.3V, Vmax=1.3V, Vmin=0.7V.

Can this be made to work? I'm sure that Bernd will tell me an answer;
and, if I am lucky, we may get an answer frm somebody with more
knowledge than either I or Bernd have. I think you can keep the devices
properly biased under such a scheme.
But then again, this is not my specialty.

Even if it can be made to work, is there any advantage to doing it?
Only if the round trip costs are significant; i.e. only if there is
significant impedance in the power and ground networks, outside of the
actual switching circuit that does the computation. Can you store the
charge usefully? To my naive way of thinking, that implies capacitors,
probably a bank on chip.

(It's interesting to think about why AC ultimately beat the DC systems
that Edison initially deployed. I don't think power efficiency in the
device was a big part; mainly it was the ability to transmit AC longer
distances, with the ease of transformers to change voltage level.
Dissipation in transmission was part of it. I don;t know if this has
any relevance to electronics.)


> I can't tell you. Andy apparently doesn't care much about the physics
> behind integrated circuits, his knowledge stops at the gate level.This
> is completely ok for digital design, but I wonder why he makes that
> sort of suggestions ;-).

Over the years I have learned that, if at a dead end, one must ask
stupid questions and come up with blue sky ideas, questioning the
implicit assumptions of the field, and/or starting from first
principles. Apparently we are approaching such a dead end wrt
circuits. I'm not *that* afraid to embarrass myself. My definition of
a computer architect is the guy who is able to go in and call "Bullshit"
on any specialist in the team who is sandbagging and/or reporting
obstacles. I've simplified logic by 100x. Dave Papworth, my former
boss on P6, once showed that the process technology guys were
sandbagging by more than 30% - he asked me for my CRC Handbook of
Chemistry and Physics, looked up the physical properties of aluminum,
and went and banged heads together.




> One interesting property of quantum mechanics is that for irreversible
> logic, there's a minimum amount of energy that is necessary to make it
> happen. Reversible logic does not have this drawback. Therefore,
> people investigate into reversible logic, even though the actual
> components to get that benefit are not in sigh (not even carbon nanotube
> switches have these properties, even though they are much closer to the
> physical limits for irreversible logic). Many people also forget that
> quantum mechanics does not properly take changes in the system into
> account, and that means that your reversible logic only works with the
> predicted low power when the inputs are not changing any more - and this
> is just the uninteresting case (the coherent one - changes in the system
> lead to decoherence, and thereby to classical physics).

I've been around long enough that I remember, during the ECL to CMOS
transition, when some people said that one of the advantages of CMOS was
that it might support reversible, adiabatic, or charge recovery logic.
Apparently that did not pan out, eh? Apparently reversible, adiabatic,
or charge recovery logic is always a few generations ahead of the
present technology, on materials a few generations ahead of the
silicon-hafnium-germanium whatever we are using now.

I can't help but wonder if this is part of the problem: the search for
the perfect, the lowest possible power, as opposed to what can be done now.

===

Anyway, this may be a moot point. IEEE Spectrum October 2000, p.40 has
an article by Pushkar Apte & George Scalise at the SIA, mentioning
graphene base BiSFET (Bilayer pseudo-Spin Field Effect Transistors) (not
to be confused with BISFET, the BI-Stable FET). They say the BiSFET
"needs only one-hundredth to one-thousandth the power of a standard MOSFET".
From: "Andy "Krazy" Glew" on
Bill Todd wrote:
> Gavin Scott wrote:
>> "Andy \"Krazy\" Glew" <ag-news(a)patten-glew.net> wrote:
>>> I am not aware of an Itanium shipped or proposed that had an "x86
>>> core on the side".
>>
>> I'm pretty certain that Merced had hard support for x86 and McKinley
>> may well have as well.
>
> Hardware support for x86 on Itanic existed on Merced, McKinley, Madison,
> and Madison II (the one with the 9 MB on-chip cache), finally
> disappearing in 2006 with Montecito.

Aarghh!!! Read what I said:

I'm not aware of any Itanium shipped or proposed that had "an x86 core
on the side".

I am aware of Itania that had hardware support for x86.

If you care enough, go look up the patents.

---

Perhaps the problem is terminology: to me, and to people at Intel, "x86
on the side" means that you have an x86 core right next to the Itanium
core, connected at the bus, or perhaps the L1 or L2 cache.

We similarly used the term "TCOTS", Tiny Core On The Side", to refer to
something like having a big x86 core, and a small x86 core for when in
power saving modes. As opposed to just powering up the big core for,
say, 1ms every second.

If you have an x86 decoder generating Itanium VLIW instructions, using
Itanium execution units, cache hierarchy, and TLBs, perhaps with a few
widget instructions and/or execution units, then that is not an x86 on
the side.
From: Bill Todd on
Andy "Krazy" Glew wrote:
> Bill Todd wrote:
>> Gavin Scott wrote:
>>> "Andy \"Krazy\" Glew" <ag-news(a)patten-glew.net> wrote:
>>>> I am not aware of an Itanium shipped or proposed that had an "x86
>>>> core on the side".
>>>
>>> I'm pretty certain that Merced had hard support for x86 and McKinley
>>> may well have as well.
>>
>> Hardware support for x86 on Itanic existed on Merced, McKinley,
>> Madison, and Madison II (the one with the 9 MB on-chip cache), finally
>> disappearing in 2006 with Montecito.
>
> Aarghh!!! Read what I said:

I did. However, I was not responding to you but to the statement that
Gavin made (which could be interpreted to suggest opposition to your
statement above or simply as clarification of the bounds within which
that statement was valid, if he was not certain whether you were denying
the existence of any hardware support for x86 on Itanic).

You will note that my own phrasing (like Gavin's) was fairly precise in
this area: I did not presume to know Gavin's intent, but clarifying how
long hardware support for x86 existed on Itanic seemed appropriate given
his uncertainty about that point.

- bill
From: Bernd Paysan on
Robert Myers wrote:
> Let's see. Quantum mechanics properly applied takes account of
> everything in the whole universe, which is, so far as I know, quantum
> mechanical and reversible in it's entirety.

Nope. That's just wishful thinking from people who do QM. The concepts
behind decoherence are partly understood, and some even quantified (e.g.
the critical temperature corresponds to what has been called "thermal
de-Broglie wavelength", which more or less describes the relation
between random changes in the conductor and a "volume of coherence"),
but overall, this is not part of QM. QM describes what happens within
the volume of coherence, classical physic describes what happens
outside. There is no accepted unified theory that gives a good reason
for this boundary and works equally well on both sides of the coherence
fence.

Note that when the "observer" is actually a quantum mechanical object,
it won't disturb the other parts of the system - it will be part of this
reversible dance.

> Thus, even though you can't do operations with *no* net cost in
> energy, we can still build and operate devices that act as quantum
> mechanical computers to an arbitrarily good approximation.

No, we can't. The quantum computer people are struggling with the same
thermal de-Broglie wavelength as the superconducting people. Just that
in quantum computing, the wavelength also depends on how many bits your
system has - but unfortunately, it goes exponentially with the number of
bits. So you can either make it arbitrary slow for many bits, or reduce
the number of bits to an almost useless amount, but not both at the same
time.

The people who work on these research projects are certainly more
optimistic than I am, but maybe Andy is right, and I've taken some
lessons from Nick ;-).

--
Bernd Paysan
"If you want it done right, you have to do it yourself"
http://www.jwdt.com/~paysan/