Bob posted:
1) Everything actually transmitted is analog.
Well, not quite. This touches on one of my personal hot buttons, which
is a widespread confusion about that the terms "analog" and "digital"
really mean.
"Analog" does NOT equate to either "continuous" or "linear", and in
fact there are examples of "analog" communications schemes which
are neither. "Analog" simply means that the information being carried
is encoded on to the signal in such a way that a given parameter of
the signal - the voltage, say - is varying in a manner that is
*analogous* to the original information. In an AM radio signal, for
instance, the amplitude of the carrier is changing in a manner analogous
to the variations in the original sound waves. In a standard analog video
signal, the voltage of the signal varies according to the luminance in the
original image. It is incorrect to say that "everything is analog" from this
perspective, just as it is incorrect to say that "everything is really digital"
(often heard from those who are invoking the discrete nature of things, per
quantum mechanics and such). "Digital" simply means that the information is
being encoded in such a way that the signal
states are to be interpreted as corresponding to numeric values (or
more broadly, some other such symbolic representation) rather than
being a "drawing" of the original information.
The reason we study "analog" communication has to do with an
incorrect assumption that is embedded in the original question.
"Digital" communication is NOT inherently superior in all aspects
as compared with "analog" communication; the superiority of one
over the other depends on the specific requirements of the application
being considered, and the specifics of the encoding schemes, analog
OR digital, being compared. What most people think of as
"digital" communications (straight binary encoding, transmitted
serially in a given channel) typically will have advantages in terms of
noise immunity, but it actually horribly inefficient in terms of its
efficiency - the amount of information that can be transmitted per second in a
given bandwidth. It also suffers from a sort of "cliff effect" vs. many common
"analog" schemes - meaning that this form of digital communications will
remain essentially "perfect" (lossless) right up to a certain limit in the
noise level - beyond which, EVERYTHING is lost (as opposed to common analog
systems, which degrade more gracefully in the presence of noise).
So - the bottom line here is that it just ain't that simple. We study both
analog AND digital communications because each still has its place.
There's really no such thing as an "analog" or "digital" SIGNAL, when
you get right down to it - it's all just electricity or EM waves. What
makes it "analog" or "digital" is the method used to encode the information
carried by that signal, and therefore how it is supposed to be interpreted by
the receiver.
I believe that a designer who understands electronics will not think in terms
of "what's analog and what's digital." She/he will use whatever design skills
will get the job done.
Speaking of hot buttons, you make a false statement in
"In an AM radio signal, for instance, the amplitude of the carrier
is changing in a manner analogous to the variations in the original
sound waves."
The Carrier does not vary - it is constant. The sidebands, which are generated
in the process of modulation, are analogous to the original audio signal.
Don