Jump to content

Talk:Nyquist–Shannon sampling theorem/Archive 2

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Archive 1Archive 2Archive 3

citation for theorem statement

Finell has requested a citation for the statement of the theorem. I agree that's a good idea, but the one we have stated now was not intended to be a quote, just a good statement of it. It may take a while to find a great quotable statement of the theorem, but I'll look for some. Here's one that's not too bad. Something you also find incorrect ones, which say that the sampling frequency above twice the highest frequency is necessary for exact reconstruction; that's true for the particular reconstruction formula normally used, but it not a part of what the sampling theorem says. That's why I'm trying to be careful about wording says necessary and/or sufficient in various places. Dicklyon 22:18, 29 October 2007 (UTC)

Nyquist-Shannon sampling theorem and quantum physics?

When I browsed through the article, I felt that there might be a connection to what is known as the "duality" of time and energy in quantum Physics. Partly because the interrelation of limiting frequency and time spacing of signals seems to originate in the properties of the Fourier transform, partly because from physics it is known that the longer you look, the more precise your measurement can be. Does anyone feel compentent to comment on this (maybe even in the article)? Peeceepeh (talk) 10:50, 22 May 2008 (UTC)

The Fourier transform pair (time and frequency) are indeed a Heisenberg dual, i.e. they satisfy the Heisenberg uncertainty relationship. I'm not sure if this is what you were alluding to.
I'm not sure I see a direct connection to the sampling theorem, though. Oli Filth(talk) 11:38, 22 May 2008 (UTC)

Sampling and Noisy Channels

At Bell Labs, I was given the impression that "Shannon's Theorem" was about more than just the "Nyquist rate". It was also about how much information per sample was available, for an imperfect communication channel with a given signal-to-noise ratio. Kotelnivkov should be mentioned here, because he anticipated this result. The primary aim of Kotelnikov and Shannon was to understand "transmission capacity".

The Nyquist rate was an old engineering rule of thumb, known long before Nyquist. The problem of sampling first occured in the realm of facsimile transmission of images over telegraph wire, which began in the 19th century. By the 1910s, people understood the theory of scanning -- scanning is "analog" in the horizontal direction, but it "samples" in the vertical direction. People designed shaped apetures, for example raised cosine, which years later was discovered again as a filter window by Hamming (the head of division 1135 where I worked at Bell Labs, but he left shortly before I arrived).

And of course mathematicians also knew about sampling rate of functions built up from bandlimited fourier series. But again, I do not believe Whittiker or Cauchey or Nyquist discovered what one woudl call the "sampling theorem", because they did not consider the issue of channel noise or signals or messages.

Also, it seems folks have invented the term "Nyquist-Shannon" for this article. It is sometimes called "Shannon-Kotelnikov" theorem. You could argue for "Kotelnikov-Shannon", but I believe Shannon developed the idea of digital information further than the esteemed Vladimir Alexandrovich. I hesitate to comment here, after seeing the pages of argument above, but I hope you will consider consulting a professional electrical engineer about this, because I believe the article has some problems. DonPMitchell (talk) 22:29, 9 September 2008 (UTC)

See channel capacity, Shannon–Hartley theorem, and noisy channel coding theorem to connect with what you're thinking of. As for the invention of the name Nyquist–Shannon, that and Shannon–Nyquist are not nearly as common as simply Nyquist sampling theorem, but somewhat more sensible, seems to me; check these books and others; let us know if you find another more common or more appropriate term. Dicklyon (talk) 01:53, 10 September 2008 (UTC)

Nyquist–Shannon sampling theorem is not correct?

Dear Sir/Madam,

Sorry, but I think that Nyquist–Shannon sampling theorem about the sampling rate is not correct.

Could you please be so kind to see the papers below?

http://www.ieindia.org/pdf/88/88ET104.pdf

http://www.ieindia.org/pdf/89/89CP109.pdf

http://www.pueron.org/pueron/nauchnakritika/Th_Re.pdf

Also I believe the following rule could be applied:

"If everything else is neglected you could divide the sampling rate Fd at factor of four (4) in order to find the guaranteed bandwidth (-3dB) from your ADC in the worst case sampling of a sine wave without direct current component (DC= 0)."

I hope that this is useful to clarify the subject.

The feedback is welcomed. Best and kind regards

Petre Petrov ppetre@caramail.com —Preceding unsigned comment added by 78.90.230.235 (talk) 21:30, 24 December 2008 (UTC)

I think most mathematicians are satisfied that the proof of the sampling theorem is sound. At any rate, article talk pages are for discussing the article itself, not the subject in general... Oli Filth(talk|contribs) 22:00, 24 December 2008 (UTC)
Incidentally, I've had a brief look at those papers. They are pretty incoherent, and seem mostly concerned with inventing new terminology, and getting confused in the process. Oli Filth(talk|contribs) 22:24, 24 December 2008 (UTC)
I believe that Mr. Petrov is very confused, yet does have a point. He's confused firstly by thinking that the sampling theorem is somehow associated with the converse, which is that if you sample at a rate less than twice the highest frequency, information about the signal will necessarily be lost. As we said in this talk page before, that converse is not what the sampling theorem says and is not generally true. I think the what Petrov has shown (confusingly) is a counter-example, dis-proving that converse. In particular, that if you know your signal is a sinusoid, you can reconstruct it with many few samples. This is not really a very interesting result and is not related to the sampling theorem, which, by the way, is true. Dicklyon (talk) 05:38, 25 December 2008 (UTC)
On second look, I think I misinterpreted. It seems to me now that Petrov is saying you need 4 samples per cycle (as opposed to 1/4, which I though at first), and that the sampling theorem itself is not true. Very bogus. Dicklyon (talk) 03:12, 26 December 2008 (UTC)

Dear All, Many thanks for your attention. May be I am confused but I would like to say that perhaps you did not have pay enough attention to the “Nyquist theorem” and the publication stated by above. I’m really sorry if my English is not enough comprehensible. I would like to ask the following questions:

  1. Do you think that H. Nyquist really formulated clearly “sampling theorem” applicable to real analog signal conversion and reconstruction?
  2. . What is the mathematical equation of the simplest real band limited signal (SBLS)?
  3. Do you know particular cases when the SBLS can be reconstructed with signal sampling factor (SSF) N= Fd/Fs <2?
  4. Do you know particular cases when the SBLS can not be reconstructed with SSF N= 2?
  5. Do you know something written by Nyquist, Shannon, Kotelnikov, etc. which gives you possibility to evaluate the maximal amplitude errors during the sampling the SBLS, SS or CS with N>2? (Emax, etc, Please see the formulas and the tables in the papers).
  6. What is the primary effect with sampling SS, CS and SBLS with SF N=2?
  7. Do not you think that clarifying the terminology is one possible way to clarify the subject and to advance in the good direction?
  8. If the “classical sampling theorem” is not applicable to the signal conversion and cannot pass the test of SBLS, SS and CS to what it is applicable and true?

I hope that you will help me to clarify the subject. BR P Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 09:22, 25 December 2008 (UTC)

Petrov, I don't think anyone ever claimed that Nyquist either stated or proved the sampling theorem. Shannon did, as did some of the other guys mentioned, however. I'm most familiar with Shannon's proof, and with decades of successful engineering applications of the principle. Using the constructive reconstruction technique mentioned, amplitude errors are always zero when the conditions of the theorem are satisfied. If you can rephrase some of your questions in more normal terms, I might attempt answers. Dicklyon (talk) 03:12, 26 December 2008 (UTC)
He should take it to comp.dsp. They'll set him straight. 71.254.7.35 (talk) 04:02, 26 December 2008 (UTC)

Rephrasing

Hello! Marry Christmas to all! If I understand clearly:

  1. Nyquist has never formulated or proved “sampling theorem” but there are “Nyqiust theorem/zone/frequency/criteria” etc? (PP: Usually the things are named after the author? Or this is a joke?)
  2. Shannon has proved “sampling theorem” applicable to real world signal conversion and reconstruction ? (PP: It is strange because I have read the papers of the “guys” (Kotelnikov included) and I have found nothing applicable to the real world! Just writings of theoreticians who do not understand the sampling and conversions processes?)
  3. Yes the engineering applications have done a lot to mask the failure of the theoreticians to explain and evaluate the signal conversion!
  4. The amplitude errors are zero?? (PP: This is false!. The errors are not zero and the signal cannot be reconstructed “exactly” or “completely! Try and you will see them!)
  5. Starting the rephrasing:
    • N< 2 is “under sampling”.
    • N=2 is “Shannon (?) sampling” or just “sampling”.
    • N>2 is “over sampling”.
    • SBLS is “the simplest band limited signal” or according to me “analog signal with only two lines into its spectrum which are a DC component and a sine or co-sine wave”.
  6. comp.dsp will set me straight? (PP: OK).

I hope the situation now is clearer. P.Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 06:40, 26 December 2008 (UTC)

A proof of the sampling theorem is included in one of (I don't remember which) "A Mathematical Theory of Communication" or "Communication in the presence of noise", both by Shannon.
The "amplitude errors" are zero, assuming we're using ideal converters (i.e. no quantisation errors, which the sampling theorem doesn't attempt to deal with), and ideal filters. In other words, the signal can be reconstructed perfectly; the mathematical proof is very simple.
I'm not sure you're going to get very far by introducing your own terminology and concepts ("SBLS", "sampling factor", etc.), because no-one will understand what you're talking about! Oli Filth(talk|contribs) 13:10, 26 December 2008 (UTC)
  1. ???A Mathematical Theory of Communication" or "Communication in the presence of noise", both by Shannon?? I have read them carefully. Nothing is applicable to the sampling and ADC. Please specify the page and line number. Please specify how this publications are related to the real conversion of an analog signal.
  2. Perhaps I will not advance with my terminology but at least I will not repeating unrelated to the signal conversion "proven" theory.
  3. Errors are inevitable. You will never reconstruct "exactly" an analog signal coveted into digital for. Try it and you will see!
  4. About the amplitude error. Could you please pay attention to the Figure 5 at page 55 at http://www.ieindia.org/pdf/89/89CP109.pdf. You will see clearly the difference between the amplitude of the signal and the maximal sample. OK?
BR P. Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 15:21, 26 December 2008 (UTC)
The sampling theorem doesn't attempt to deal with implementation limitations such as quantisation, non-linearities and non-ideal filters. No-one has claimed that it does.
You can reconstruct a bandlimited analogue signal to an arbitrary degree of accuracy. Just use tighter filters and higher-resolution converters.
What you've drawn there is the result of a "stair-case" reconstruction filter, i.e. a filter with impulse response . This is not the ideal reconstruction filter; it doesn't fully eliminate the images. In practice, a combination of oversampling and compensation filters can reduce the image power to a negligible level (for any definition of "negligible") and hence eliminate the "amplitude errors". None of this affects the sampling theorem!
In summary, no-one is disputing the fact that if you use sub-optimal/non-ideal converters and filters, you won't get the same result as the sampling theorem predicts. Oli Filth(talk|contribs) 15:33, 26 December 2008 (UTC)


Hello again!

I am really sorry but we are talking about different things.

I am not sure that you are understanding my questions. and answers.

I am not disputing any filters at the moment.

Only the differences between the amplitude of the samples and the amplitude of the converted signal.

Also I am not sure that you have read "the classics" in the sampling theory.

Also, please note that there is a difference between the "analog multiplexing" (analog telephony discussed by the "classics" during 1900-1950) and analog to digital conversion and reconstruction.

I wish you good luck with the "classics" in the sampling theory! BR P Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 15:46, 26 December 2008 (UTC)

You started this conversation with "I think that Nyquist–Shannon sampling theorem about the sampling rate is not correct", with links to papers that discussed "amplitude errors" as if there was some mistake in the sampling theorem. That is what I have been talking about! If you believe we're talking about different things, then yes, I must be misunderstanding your questions! Perhaps you'd like to re-state exactly what you see as the problem with the sampling theorem.
As for filters, as far as your paper is concerned, it's entirely about filters, although you may not realise it. In your diagram, you're using a sub-optimal filter, and that is the cause of your "amplitude errors". Oli Filth(talk|contribs) 15:59, 26 December 2008 (UTC)
Joke?

Petrov, you ask "Usually the things are named after the author? Or this is a joke?" This is clear evidence that you have not bothered to read the article that you are criticizing. Please consider doing so, or keeping quiet. Dicklyon (talk) 00:44, 27 December 2008 (UTC)


Hello!

Ok.

I will repeat some of the questions again in more simple and clear form:

  • Where H. Nyquist has formulated or proved clearly stated “sampling theorem” applicable in signal conversion theory? (paper, page, line number?)
  • Where is the original clear definition of Nyquist theorem mention in Wikipedia (? (paper, page, line number?)
  • Where Shannon has formulated or proved “sampling theorem” applicable in signal conversion theory with ADC? (paper, page, line number?)
  • What we will lose if we remove the papers of Nyquist and Shannon from the signal conversion theory and practice with ADC ?
  • What is your definition of “band limited” signal discussed by Shannon and Kotelnikov?
  • Is it possible to reconstruct an analog signal which in fact is with infinite accuracy if you cut in to finite number of bits and put into circuitry with finite precision and unpredictable accuracy (As you know there are no exact value in electronics)?
  • The number e =2.7... and pi=3.14... are included in most of the real signal. How you will reconstruct them “exactly” or “completely”?

I am waiting for the answers

Br

P.Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 10:40, 27 December 2008 (UTC)

I don't know why you keep requesting where Nyquist proved it; the article already summarises the history of the theorem. As we've already stated, Shannon presents a proof in "Communication in the presence of noise"; it is quoted directly in the article. As we've already stated, this is an idealised model. Just as in all aspects of engineering, practical considerations impose compromises; in this case it's bandwidth and non-linearities. As we've already stated, no-one is claiming that the original theorem attempts to deal with these imperfections. I don't know why you keep talking about practical imperfections as if they invalidate the theorem; they don't, because the theorem is based on an idealised model.
By your logic, we might as well say that, for instance, LTI theory and small-signal transistor models are invalid, because the real world isn't ideal! Oli Filth(talk|contribs) 11:57, 27 December 2008 (UTC)


"If a function x(t) contains no frequencies higher than B cps, it is completely determined by giving its ordinates at a series of points spaced 1/(2B) seconds apart."

PP: Imagine that you have a sum of a DC signal and a SS signal.

How you will completely determine them by giving only 2 or even 3 points?

OK? —Preceding unsigned comment added by 78.90.230.235 (talk) 10:49, 27 December 2008 (UTC)

The theorem and the article aren't talking about 2 or 3 points. They're talking about an infinite sequence of points.
However, as it happens, in the absence of noise, one can theoretically determine all the parameters of a sinusoid with just three samples (up to aliases). I imagine that if one had four samples, one could determine the DC offset as well. However, this is not what the theorem is talking about. Oli Filth(talk|contribs) 11:57, 27 December 2008 (UTC)


PP: "They're talking about an infinite sequence of points." Where did you find that? paper, page, line number?

"I imagine that if one had four samples, one could determine the DC offset as well" This is my paper. Normally should be covered by the "classical theorem". OK? —Preceding unsigned comment added by 78.90.230.235 (talk) 12:17, 27 December 2008 (UTC)

It's pretty clear that you haven't read the original papers very carefully (or misunderstood them)! In "Communication in the presence of noise", Theorem #1 states it. Yes, it's true that the word "infinite" is not used in the prose, but then look at limits of the summation in Eq.7.
As for your paper, it's already a known fact (in fact, it's obvious; four equations in four unknowns), and is not in the scope of the sampling theorem (although you can probably derive the same result from the theorem). Oli Filth(talk|contribs) 12:25, 27 December 2008 (UTC)

PP: H. Nyquist, "Certain topics in telegraph transmission theory", Trans. AIEE, vol. 47, pp. 617-644, Apr. 1928 Reprint as classic paper in: Proc. IEEE, Vol. 90, No. 2, Feb 2002.

Question: Where in that publication is the "Sampling theorem"?

"I don't know why you keep requesting where Nyquist proved it..." You are stating that there is "Nyquit theorem?" (please see the article in Wikipedia). There should be a statement and a proof. OK? Where they are? —Preceding unsigned comment added by 78.90.230.235 (talk) 12:43, 27 December 2008 (UTC)

There is no article on "Nyquist theorem", only a redirect to this article. Please stop asking the same question over and over again; both Dick and I have already answered it, and the article already explains it. Oli Filth(talk|contribs) 12:47, 27 December 2008 (UTC)


PP: http://www.stanford.edu/class/ee104/shannonpaper.pdf page 448, Theorem I:

1. First failure for SS, CS or SBLS sampled and zero crossings. (One failure is enough!)

What failure? The only point of contention is in the nature of the inequality (i.e. an open or closed bound). It is generally accepted today that it is true only for an open bound. The article discusses in the introduction and in the section "Critical frequency". Again, it is clear that you haven't actually read the article.

2. Second failure: "completely" is wrong word.

Please don't tell me you're talking about your "amplitude errors" again...

3. Third failure: It is about "function" not about " a signal". Every "signal" is a "function", but not every "function" is a "signal". OK?

How is this a failure?

4. Forth failure: "common knowledge"??? Is that a proof?

No. What follows is a proof.

5. Fifth failure: No phase in the Fourier series! The phase is inherent part of the signal!

isn't constrained to be real, and neither is (and hence neither is ). Oli Filth(talk|contribs) 13:11, 27 December 2008 (UTC)

Imagine same number of failures for another theorem, e.g. Pythagoras theorem! Will you defend it in that case? —Preceding unsigned comment added by 78.90.230.235 (talk) 12:56, 27 December 2008 (UTC)

PP: "...F(ω) isn't constrained to be real, and neither is f(t)...".

You could write any equation, but cannot produce any signal.OK?

Sorry, I am talking about real signals with real functions and I am forced to evaluate the errors. You can produce the signals and test the equipment. Please excuse me. May be my mistake to start that talk. —Preceding unsigned comment added by PetrePetrov (talkcontribs) 13:19, 27 December 2008 (UTC)

"Real" as opposed to "complex"... i.e. phase is included. Oli Filth(talk|contribs) 13:21, 27 December 2008 (UTC)


PP: Hello! Again, I have looked at the papers of the "classics" in the field. May be the following chronology of the events in the field of the “sampling” theorem is OK:

1. Before V. Kotelnikov: H. Nyquist did not formulated any “sampling theorem”. His analysis (?) even of the DC (!) is really strange for an engineer. (Please see the referenced papers). No sense to be mention in sampling, SH, ADC and DAC systems. In "analog multiplexing telephony" is OK.

2. V. Kotelnikov (1933) For the first time formulated theorems, but unfortunately incomplete because did not include the necessarily definitions and calculations. No ideas on errors! May be should be mention just to see the difference between the theory and the practice.

3. C. Shannon.(1949 ) In fact repetition of part of that given by V. Kotelnikov. There is no even clearly formulated proof of something utilizable in ADC. No excuse for 1949! The digital computers were created!

No understanding of the signals (even theoretical understanding) to test its “theorems”. the necessarily definitions and calculations. No ideas of errors! No idea of application of an oscilloscope and multimeter!


4. Situation now: No full theory describing completely the conversion of the signals from analog to digital form and reconstruction.

But there are several good definitions and verifiable in practice theorem to evaluate the errors of non sampling the SS and CS into their maximums. Verifiable even with an analogous oscilloscope and multimeter!


I hope that is good and acceptable. BR

P Petrov —Preceding unsigned comment added by 78.90.230.235 (talk) 08:41, 28 December 2008 (UTC)

I'm going to say this one last time. The sampling theorem doesn't attempt to deal with "errors", such as those caused by non-ideal filters. Please stop stating the same thing time and time again; everyone already knows that the theorem is based on an ideal case. It has nothing to do with "multimeters and oscilloscopes". The only theoretical difference between "analog multiplexing" and A-D conversion is the quantisation. To say that there is "no understanding of the signals..." is total nonsense. Please stop posting the same mis-informed points!
Incidentally, Nyquist uses the term "D.C." in the context of "DC-wave", as the opposite of "Carrier wave"; we would call these "baseband" and "passband" signalling today.
If you have something on "the conversion of the signals from analog to digital form and reconstruction" from a Reliable source, then please post it here, and we'll take a look. Your own papers aren't going to do it, I'm afraid. However, even if you do find something, it's unlikely to make it into the article, because the article is about the original theorem. Oli Filth(talk|contribs) 11:37, 28 December 2008 (UTC)

Hello!

1. No need to repeat more times. From my point of view the "Nyquist-Shannon theorem" does not exists, and what exists is not applicable fully (even largely)into practice. You are free to think that it exists and the people use it.

  • And you are free to not to accept it! (although saying "it doesn't exist" is meaningless...) Yes, of course people use it. It's been the basis of a large part of information theory, comms theory and signal-processing theory for the last 60 years or so.

2. Please note that there are "representative" (simplified but still utilizable) and "non-representative" ("oversimplified" and not usable ) models. The "original theorem" is based on the "oversimplified" model and is not representable.

  • You still haven't said why. Remember, one can approximate the ideal as closely as one desires.

3. I have seen the "DC" of Nyquist before your note and I am not accepting it.

  • I have no idea what you mean, I'm afraid.

4. Because I am not a "reliable source" I will not spam any more the talks here.

  • You're free to write what you like on the talk page (within reason - see WP:TALK). However, we can only put reliable material into the article itself.

5. If you insist on the "original theorem", please copy and paste "exactly" the texts of Nyquist, Shannon, Kotelnikov,etc. which you think are relevant to the subject and let the readers to put their own remarks and conclusions outside the "original" texts. You could put your own, of course. OK?

  • The article already has the exact text from Shannon's paper. I'm not sure what more you expect?

6. I have put here a lot of questions and texts without individual answers. If Wikipedia keep them someone will answer and comment them (may be).

  • I believe I've answered all the meaningful questions. But yes, this text will be kept.

7. I do not believe that my own papers will change something in the better direction, but someone will change it because the theory (with “representative” models) and the practice should go in the same direction and the errors (“differences”) should be evaluated.

  • The cause of your "errors" is already well understood. For instance, CD players since the late 1980s onwards use oversampling DACs and sinc-compensation filters to eliminate these "errors". That's not due to a limitation in the theory, it's due to hardware limitations. The solution can be explained with the sampling theorem. Oli Filth(talk|contribs) 15:09, 28 December 2008 (UTC)

Good luck again. I am not sure that I will answer promptly to any comment (if any) posted here.

BR Petre Petrov

Rapidly oscillating edits

I noticed some oscillation between 65.60.217.105 and Oli Filth about what to say about the conditions on x(t). I would suggest we remove the parenthetical comment

"(which exists if is square-integrable)"

For the following two reasons. First, it exists also in many other situations. Granted this is practically the most common. Second, it is not entirely clear the integral we then follow this statement with exists in if x(t) is square integrable. I do not think it detracts at all from the article to simply say that X(f) is the continuous Fourier transform of x(t). How do other people feel about this? Thenub314 (talk) 19:03, 3 January 2009 (UTC)

PS I think 65.60.217.105 thinks the phrase continuous Fourier transform is about the Fourier transform of x(t) being continuous, instead of being a synonym for "the Fourier transform on the real line." Thenub314 (talk) 19:14, 3 January 2009 (UTC)

I realise that I'm dangerously close to 3RR, so I won't touch this again today! The reason I've been reverting is that replacing "square-integrable" with "integrable" is incorrect (however, square-integrability is a sufficient condition for the existence of the FT; I can find refs if necessary). I'm not averse to removing the condition entirely; I'm not sure whether there was a reason for its inclusion earlier in the article's history. Oli Filth(talk|contribs) 19:10, 3 January 2009 (UTC)
I agree with your guess as to how 65.60.217.105 is interpreting "continuous"; see his comments on my talk page. Oli Filth(talk|contribs) 19:36, 3 January 2009 (UTC)
Yes, thanks for pointing me there. Hopefully my removal of "continuous" will satisfy him. I suppose I should put back "or square integrable". Dicklyon (talk) 20:08, 3 January 2009 (UTC)
Not a problem. I agree with you Oli that the Fourier transform exists, but the integral may diverge. I think it follows from Carleson's theorem about almost everywhere convergence of Fourier series that this happens at worst almost everywhere, but I don't off hand know of a reference that goes into this level of detail (and this would apply only to the 1-d transform).
Anyways I am definitely digressing. The conditions are discussed in some detail in the Fourier transform article, which we link to. So overall I would be slightly in favor of removing the condition entirely. But I think (Dicklyon)'s version works also. (Dicklyon), how do you feel about removing the parenthetical comment?
I wouldn't mind removing the parentheetical conditions. Dicklyon (talk) 22:06, 3 January 2009 (UTC)

Geometric interpretation of critical frequency

I'm not sure the new addition is correct. Specifically:

  • the parallel implied by "Just as the angles on a circle are parametrized by the half-open interval [0,2π) – the point 2π being omitted because it is already counted by 0 – the Nyquist frequency must be omitted from reconstruction" is invalid, not least because the Nyquist frequency is at π, not 2π.
  • the discussion of "half a point" is handwaving, which is only amplified by the use of scare quotes. And it's not clear how it makes sense in continuous frequency.
  • it's not made clear why the asymmetry disappears for complex signals.

Oli Filth(talk|contribs) 19:24, 14 April 2009 (UTC)

Critical frequency

This section is unnecessarily verbose. It is sufficient to point out that the samples of:

are identical to the samples of:

and yet the continuous functions are different (for sin(θ) ≠ 0).

--Bob K (talk) 19:29, 14 April 2009 (UTC)

higher dimensional nyquist theorem equivalent?

The Nyquist theorem applies to more than just time-series signals. The theorem also applies in 2-D (and higher) cases, such as in sampling terrain (for example), in defining the maximum reconstructable wavenumbers in the terrain. However, there is some debate as to whether the theorem applies directly, or whether it has subtle differences. Can anyone care to comment on that or derive it? I will attempt to do so following the derivations here, but I probably will lose interest before then.

It seems that it should apply directly given that the Fourier transform is a linear transform, but the debate has been presented so I thought it should go in discussion before the main page. Thanks.

Andykass (talk) 17:45, 12 August 2009 (UTC)

You need to ask for sources, not derivations. Dicklyon (talk) 02:03, 13 August 2009 (UTC)
Check the article on Poisson summation formula, and especially the cited paper Higgins: Five short stories... There is the foundation for sampling on rectangular and other lattices and an local abelian groups, connected with the name Kluvanek.--LutzL (talk) 08:24, 13 August 2009 (UTC)

this T factor issue is coming up again.

remember that "Note about scaling" that was taken out here ?

well, the difference of this article from the common (and flawed, from some of our perspectives) convention of sampling with the unnormalized Dirac comb and including a passband gain of T in the reconstruction filter is starting to have a consequence. i still think we should continue to do things they way we are (why repeat the mistake of convention?) but people have begun to object to this scaling (because it "not in the textbooks" even though it is in at least one).

anyway, Dick, BobK, anyone else want to mosey on over to Talk:Zero-order hold and take a look and perchance offer some comment? r b-j 21:01, 26 January 2007 (UTC)

OK, I gave it my best shot. Dicklyon 23:06, 26 January 2007 (UTC)
Hello again. I certainly can't match Doug's passion for this subject. And I can't improve on Rbj's arguments. I haven't given this as much thought as you guys, but at first glance, it seems to me that the root of the problem is our insistence that "sampling" is correctly modelled by the product of a signal with a Dirac comb. We only do that to "prove" the sampling theorem in a cool way that appeals to newbies. (It certainly sucked me in about 40 years ago.) But there is a reason why Shannon did it his way.
Where the comb really comes from is not the sampling process, but rather it is an artifact of the following bit of illogic: Suppose we have a bandlimited spectrum on interval -B < f < B, and we do a Fourier series expansion of it, as per Shannon. That produces a function, S(f), that only represents the original spectrum in the interval -B < f < B. Outside that interval, S(f) is periodic, which is physically meaningless. But if we ignore that detail, and perform an inverse Fourier transform of S(f), voilà... the Dirac comb emerges for the first time.
Then we compound our mistake by defining sampling to be the product of a signal with a Dirac comb that we created out of very thin air.  I'd say that puts us on very thin ice.
--Bob K 23:14, 26 January 2007 (UTC)
Thin ice is right. Taking transforms of things that aren't square integrable is asking for trouble. Doing anything with "signals" that aren't square integrable is asking for trouble. But as long as we're doing it, might as well not make matters worse by screwing it up with funny time units. There's good reason for this approach in analysing a ZOH, of course, but one still does want to remain cognizant of the thin ice. Dicklyon 23:40, 26 January 2007 (UTC)
I totally agree about the units. I'd just like to reiterate that even without the "square-integrable" issue, what justification do we have for treating S(f) as non-compact (if that's the right terminology)? I.e., what right do we have to assign any importance to its values outside the (-B, B) domain? Similarly, when we window a time-series of samples and do a DFT, the inverse of the DFT is magically periodic. But that is just an artifact of inversing the DFT instead of the DTFT. It means nothing. It is the time-domain manifestation of a frequency-domain approximation.
If this issue seems irrelevant to the discussion, I apologize. But my first reaction to the ZOH article was "the Dirac comb is not necessary here". One should be able to have a perfectly good article without it. But I need to go and really read what everybody has said there. Maybe I will be able to squeeze that in later today.
--Bob K 16:16, 27 January 2007 (UTC)
the liklihood of crashing through the ice is no greater than that of crashing in Richard Hamming's airplane designed using Riemann instead of Lebesgue integration. why would nearly all of these texts including O&S (which i have always considered kinda a formal reference book, not so much for describing cool DSP tricks, but more as a rigorous description of simply what is going on) have no problem with using the Dirac comb? They instead like to convolve with the F.T. of the Dirac comb (which is, itself, a Dirac comb) which is more complicated than just using the shifting theorem caused by the sinusoids in the Fourier series of the Dirac comb. wouldn't that have to be even thinner ice, yet these textbooks do it anyway. their only problem is the misplaced T factor.
BTW, Dick, i agree with you that
is more compact and nicer than
but also less recognizable. it's just like
instead of
except it is harder to see the scaling of time in the infinitely thin delta. r b-j 08:02, 27 January 2007 (UTC)
r b-j 08:02, 27 January 2007 (UTC)
I'm not up on the history of this thread, but FWIW I like     better than   .   And I like   best,   because it's easiest to see that its integral is T.
--Bob K 16:31, 27 January 2007 (UTC)
Bob, good point, and that's why we stuck with that form. Dicklyon 17:11, 27 January 2007 (UTC)
I just read your response at ZOH, and the point about scaling the width instead of the amplitude is compelling. That elevates     up a notch in my estimation. --Bob K 16:44, 29 January 2007 (UTC)
Robert, re the thin ice in textbooks like O&S, it's OK, but it's too bad they don't put the necessary disclaimers, references, or whatever to allow a mathematician to come in and understand the conditions under which the things they derive make sense. It's about enough for engineers, because they're all too willing to let the mathematical niceties slide, but then that makes it tricky when people try to use and extend the ideas or try to make them rigorous. So we end up arguing... not that we have any real disagreement at this point, but so often I see things where a Fourier transform is assumed to exist even when there is no way within delta functions and such even. Dicklyon 17:11, 27 January 2007 (UTC)

I find confusing the traditional textbook discussion of using of a Dirac comb to represent discrete sampling. I am also not sure that I agree with the assertions made here that it all wrong ('mistake'). As I understand it, delta functions only having meaning with multiplication AND integration over infinity. So, simply multiplying a Dirac comb by a function does not, on its own, represent discrete sampling. One must also perform the integration. Doesn't this correct the dimensional issues ('T factor')? —Preceding unsigned comment added by 168.103.74.126 (talk) 17:43, 27 March 2010 (UTC)

Simplifications?

Bob K, can you explain your major rewrite of the "Mathematical basis for the theorem" section? I'm not a huge fan of how this section was done before, but I think we had it at least correct. Now I think I have to start over and check your version, some of what I'm not so sure I understand. Dicklyon (talk) 21:10, 12 September 2008 (UTC)

Hi Dick,
I thought it was obvious (I'm assuming you noticed Nyquist–Shannon_sampling_theorem#math_Eq.1), but I'm happy to explain. I wasn't aware of the elegant Poisson summation formula back when we were creating the "bloated proof" that I just replaced. Without bothering with Dirac comb functions and their transforms, it simply says that a uniformly sampled function in one domain can be used to construct a periodically extended version of the continuous function's transform in the other domain. The proof is quite easy and does not involve continuous Fourier transforms of periodic functions (frowned on by the mathematicians). And best of all, it's an internal link... no need to repeat it here. Or I could put it in a footnote, if you like that better.
Given that starting point, it is obvious that can be recovered from under the conditions assumed by Shannon. All that's left is the math to derive the reconstruction formula.
Is that what you wanted to know?
--Bob K (talk) 22:28, 12 September 2008 (UTC)

As I see it, the main problem with this new version of the proof is that it doesn't appeal to most people's way of thinking about sampling ... many will think about picking off measurements in the time domain. Furthermore, there really should be two versions of the proof, one that works in the time domain and one that works in the frequency domain. Although Bob K might disagree, I think the time domain proof (that was once on this page) is fine, and it should use the Dirac comb. But the application of the Dirac comb involves more than just multiplication of the Dirac comb by the function being sampled. Also needed is integration over the entire time domain. Oddly, I don't see this seemingly important step in textbooks. —Preceding unsigned comment added by 136.177.20.13 (talk) 18:50, 27 March 2010 (UTC)

I'm in agreement that the proof that existed earlier was far more clear than what we see now, but 136, could you be more specific about what you mean by your last three sentences? How is multiplication of the function being sampled by a Dirac comb inadequate to fully performing the sampling operation? What goes in is the function being sampled, and what comes out is a sequence of Dirac impulses weighted by the sample values; the sample values fully defines the Dirac comb sampled signal in the time domain. 70.109.175.221 (talk) 20:15, 27 March 2010 (UTC)

This is '136' again. I'm still working this out myself, and I'm probably wrong on a few details (I'm not a mathematician, but a scientist!) but think first of just one delta function d(t-t0) and how it is applied to a function f(t), we multiply the function by the delta function and then integrate over all space (t in this case). So, Int[f(t) . delta(t-t0)] dt = f(t0). This, in effect, samples the time series f(t) at the point t0. And, if you like, the 'units' on the delta function are the inverse of its arguement, so integrating over all space doesn't change the dimensional value of the sample. Now, the comb function is the sum of delta functions. To sample the time series with the comb function we have a sum of integrated applications of each individual delta function. So, Sum_k Int[f(t) . delta(t - k.t0)] dt and this will equal a bunch of discrete samples. What I'm still figuring out is how this is normalized. Recall that int[d(t)]dt = 1. For the comb function this normalizing integral is infinite, but I think you can get around this by first considering n delta functions, then take the limit as n goes to infinity. You'd need to multiply some of the results by 1/n. —Preceding unsigned comment added by 168.103.74.126 (talk) 20:36, 27 March 2010 (UTC) A related issue is how it is we should treat convolution of the comb function. Following on from my discussion of how discrete sampling might be better expressed (right above this paragraph), it appears to me that convolution will involve an integral for the convolution itself, an infinite sum over all delta functions in the comb, and another infinite integration to handle the actual delta-function sampling of the time series. —Preceding unsigned comment added by 168.103.74.126 (talk) 22:00, 27 March 2010 (UTC)

You might want to take this up on USENET at comp.dsp. Essentially, the sampling operation, the multiplication of the dirac deltas are what samples f(t). To rigorously determine the weight of each impulse, mathematically, we don't need to integrate from -inf to +inf, but only need to integrate from sometime before t0 to sometime after t0. But multiplication by the dirac comb keeps the f(t) information at the discrete sample times and throws away all of the other information about f(t). You don't integrate over all t for the whole comb. For any sample instance, you integrate from, say, 1/2 sample time before to 1/2 sample after the sample instance. 70.109.175.221 (talk) 05:59, 28 March 2010 (UTC)

This is '136' again. I'd also like to say that the formula 3, which is supposed to show the time domain version of the sampling theorem results, kind of makes a mess of needed obvious symmetry between multiplication and convolution in the time and frequency domains. So, the multiplication of the rectangle function in the frequency domain (to band limit results) should obviously be seen as convolution of the sinc function in the time domain (which amounts to interpolation). What we have right now does not make any of this clear (and, at least at first glance, seems wrong). Compare the mathematical development with the main formula under 'interpolation as convolution' on the Whittaker-Shannon page. This formula should be popping out here on the sampling page as well. So, I'm afraid what we have on this page is not really a 'simplification'. Instead, it is really just a mess. —Preceding unsigned comment added by 75.149.43.78 (talk) 18:04, 4 April 2010 (UTC)

Reconstructability not a real word?

I can't find reconstructability in any dictionary. What I do find are the following terms:

  1. Reconstruction (noun)
  2. Reconstructible (adjective)
  3. Reconstruct (verb)
  4. Reconstructive (adjective)
  5. Reconstructively (adverb)
  6. Constructiveness (noun)

This would point to reconstructable not being a real word, but reconstructible is. Reconstructiveness and reconstructibility might be. --209.113.148.82 (talk) 13:16, 5 April 2010 (UTC)

max data rate = (2H)(log_2_(V)) bps

Quoting from a lecture slide:

In 1924, Henry Nyquist derived an equation expressing the maximum rate for a finite-bandwidth noiseless channel.
H is the maximum frequency
V is the number of levels used in each sample
max data rate = (2H)(log_2_(V)) bps
Example
A noiseless 3000Hz channel cannot transmit binary signals at a rate exceeding 6000bps (this would mean there are 2 "levels")

I can't relate that very well to this article. I recognize the 2H parameter, but the "levels" referred to here I'm not sure where they come from.

Then it says Shannon extended Nyquist's work:

The amount of thermal noise ( in a noisy channel) can be measured by a ratio of the signal power to the noise power ( aka signal-to-noise ratio). The quantity (10)log_10_(S/N) is called decibels.
H is the bandwidth of the channel
max data rate = (H)log_2_(1+S/N) bps
Example
A channel of 3000Hz bandwidth and a signal-to-noise ratio of 30dB cannot transmit binary signals at a rate exceeding 30,000bps.

Just bringing this up because people looking for clarification from computer communication lectures might find the presentation a bit odd, take it or leave it. kestasjk (talk) 06:47, 26 April 2010 (UTC)

The first example is misleading. It should state "a noiseless 3000Hz channel cannot transmit signals at a rate exceeding 6000 baud." Nyquist says nothing of the bit rate. Oli Filth(talk|contribs) 07:34, 26 April 2010 (UTC)
Oli, you need to qualify your statement to say binary signals. a noiseless channel of any finite and non-zero bandwidth can conduct a signal of any information rate. but if you're limited to binary signals, what you say is true. 70.109.185.199 (talk) 16:01, 26 April 2010 (UTC)
No, there is no "need". Baud is symbols per second, see the given link.--LutzL (talk) 16:28, 26 April 2010 (UTC)
The point is, Nyquist said nothing about the information rate, and Shannon said nothing about the alphabet size, so the comparison is an "apples vs oranges" one. Oli Filth(talk|contribs) 16:32, 26 April 2010 (UTC)


A few clarifications and suggestions:
1. Many of the issues in your lecture notes are discussed in the bit rate article.
2. The term "data rate" in the lecture notes should be replaced by gross bit rate in the Nyquist formula, and net bit rate (or information rate) in the Shannon-Hartley formula. The formulas are not about the same data rate. Many computer networking textbooks confuse this. I tried to clarify that in this article once, but it was reverted.
3. Almost all computer networking textbooks credit Nyquist for calculating gross bit rate over noiseless channels, while telecom/digital transmission literature typically call this Hartley's law. At Wikipedia, datatransmission is discussed in the Nyquist rate article. I agree with that it should also be discussed in the Nyquist theorem article, because so many students are checking it. Hartley's law is so important that it deserves its own Wikipedia article, and not only a section in the Shannon-Hartly article.
4. When applied to data transmission, the bandwidth in the Nyquist formula refers to passband bandwidth=upper minus lower cut-off frequency (especially if passband transmission=carrier modulated transmission). In signal processing, it refers to baseband bandwidth (also in so called over-sampling, which is said to exceed the Nyquist rate).
5. The Nyquist formula is valid to baseband transmission (i.e. line coding), but in practice when it comes to passband transmission (digital modulation), most modulation schemes only offer less than half the Nyquist rate. I have only heard about the vestigial sideband modulation (VSB) digital modulation scheme, that may offer near the Nyquist rate.
6. Many of the articles in information theory, for example Nyquist–Shannon sampling theorem, channel capacity, etc, can only be understood by people with a signal processing/electrical engineering background. The article lead should written in a way that can be understood by computer science students without university level math knowledge. I tried a couple of years ago to address this issue, but someone reverted most of my changes instead of further improving them, so I gave up.Mange01 (talk) 18:55, 26 April 2010 (UTC)

Misinterpretation?

I've reverted this edit, because I believe you're misinterpreting what Shannon was saying. He was not saying that the time window T was some integral part of the sampling theorem, merely that in a bandwidth W and a time window T, there are a total of 2TW dimensions. To start talking about longest wavelengths and so on without an explicit source is original research, I'm afraid. Oli Filth(talk|contribs) 09:15, 19 August 2009 (UTC)

Following this edit, I'll add the following points:
  • T doesn't imply a "lower frequency bound". If you believe it does, then please provide a source that says this explicitly.
  • Your second quote isn't from Shannon. Again, please provide a source.

Oli Filth(talk|contribs) 14:59, 19 August 2009 (UTC)

Are you reading the same Shannon quote that I am? Shannon wrote exactly (bold emphasis is mine):

...and that we are allowed to use this channel for a certain period of time T. Without any further restrictions this would mean that we can use as signal functions any functions of time whose spectra lie entirely within the band W, and whose time functions lie within the interval T.

Proper reading of the quote relative to the lower bound is:

...we can use as signal functions any functions of time ... whose time functions lie within the interval T.

He is saying "we can use as signals" any "spectra" (i.e. input signals), "whose time functions lie within the interval T". In other words, the input signals have to have time functions which lie within the sampling duration. Shannon tells us that time functions are signals, "we can use as signal functions any functions of time". So Shannon is saying that the signals must lie within the sampling duration T. And that is common sense. How can you sample a signal which does not lie within the sampling duration? I can't imagine why you've wasted so much time on something so obvious and fully specified by the historic Shannon quote cite. The sampling duration interval T dictates the longest wave period that can be sampled. How can you sample a signal with a longer period than sampling duration T? So if the duration of T bounds the period of the wave of input signals, then it thus means T is a low frequency bound. That is not interpretation, rather is a direct mainstream accepted mathematical relationship. T is period, which dictates frequency by the relationship 1/T. When you bound largest period on the upper end to T, then frequency is bounded by 1/T on the lower value. That is a simple mathematical identity. There is no research nor interpretation involved. I am just amazed at the slowness of your mind. You are apparently not qualified to be an editor of this page, especially if you still do not understand after this multiple redundant explanation. You are apparently confusing wavelength (space domain) with wave period (time domain).
As for the 2nd blockquote in the edit I provided (i.e. the 3rd blockquote in the introduction section), it was my attempt to show what Shannon's original quote would be, if the emphasis was shared with the obvious sampling duration lower frequency bound-- feel free to edit it and make that more clear or delete that 2nd blockquote. Shannon's focus is obviously on the upper bound, because obviously most communication work is focused on high frequency issues, but Shannon did fully qualify the lower bound in his opening paragraph as I have explained above. I admire Shannon for his completeness, even he had no reason to be interested in Fat tail signals. Thus, I don't need to cite any other source but Shannon, as I am not interpreting anything, merely quoting what Shannon wrote. Feel free to edit my contribution to remove any portion that you can justify is "interpretation" or original research, but please keep the quote of what Shannon wrote on the matter and keep a coherent reason for including the quote. The lower bound is becoming more important lately, as the world is seeing the error of ignoring fat tail signals (e.g. fiat money systems, systemic risk, etc). I was predicting this 20 years ago. As for your threat to ban me for undoing your incorrect reverts, if ignorance is power, then go ahead. --Shelbymoore3 (talk) 22:49, 19 August 2009 (UTC)
If Shannon had meant to state what you have written (the second "quote" in your edits to the article), then he would have written it that way himself. But he didn't. It's not up to us to extrapolate from the sources beyond what they support. Again, as I said, please provide a source that explicitly supports your interpretation or I will remove it again, as everything you've currently written is original research.
A couple of points in response to your arguments above:
  • Of course you can sample a sinusoid whose period is longer than your observation window; what matters is the bandwidth of the signal, not its frequency.
  • It's obvious that any signal that has finite support in the frequency domain has infinite support in the time domain. What Shannon was getting at was in the context of signals of infinite time-domain support (the context was communication signals), where the dimensionality is exactly 2TW in any interval of length T. There is no lower bound implied by the sampling theorem; if there were such a bound, then obviously 2TW would no longer hold.
  • Furthermore, Shannon goes on to explain what he means by "a time function with the interval T".
  • I'm well aware that you meant "wave period" and not "wavelength" (although actually the difference is irrelevant in the context of sampling). Don't use my quoting of your original mistake as an excuse to guess at what I am "qualified" to do.
Incidentally, I have no power to ban you (I'm not an administrator), but I am able to report you for WP:incivility, disruptive editing and not providing sources. Before you make any further edits (whether to the article or to the talk page), I suggest you read the guideline on incivility very carefully. Oli Filth(talk|contribs) 23:49, 19 August 2009 (UTC)
The error in your understanding is based on your two statements, the first being correct in isolation and the second being false for "any" but true if you wrote "any sinusoid":
  • Of course you can sample a sinusoid whose period is longer than your observation window; what matters is the bandwidth of the signal, not its frequency.

  • It's obvious that any signal that has finite support in the frequency domain has infinite support in the time domain.
You are assuming that the composite signal being sampled is a sinusoid (sine wave), so that you can model the un-sampled portions of the wave in the time domain. Shannon did not make that less general assumption (you won't find the word sinusoid in the cited section of his paper). Rather he made a more general statement about "time function". For example with a Fat tail event, e.g. the collapse of the fiat economy or an earthquake or the observation of black swans, the event itself may be a very high frequency impulse wave at it's start and trailing ends, but have long period between the impulses. This is why the general way that Shannon stated it is so important to understand. Bandwidth alone won't capture the non-sinusoid composite signals, and that is why Shannon added the requirement "and whose time functions lie within the interval T". Shannon refers to the composite nature of the signals where he wrote "spectra".
Yes in fact Shannon does further define "time functions lie within the interval T" and it states exactly what I have stated in prior paragraph:

To be more precise, we can define a function to be limited to the time interval T if, and only if, all the samples outside this interval are exactly zero. Then we can say that any function limited to the bandwidth W and the time interval T can be specified by giving numbers 2TW.

So Shannon has clearly stated that bandwidth alone is not sufficient, and that you also have to bound on the period T. The sinusoid function can completely sampled by 2 times the maximum cycles in time interval T, but other composite signal functions will not. Thus Shannon stated the theorem in a more general mathematical way where he elaborates in the "III. GEOMETRICAL REPRESENTATION OF THE SIGNALS" that 2TW is a N-dimensional space.
Incidentally this is why bandwidth (e.g. more bits of storage) alone is not going to increase information, and can actually hide information (the disordered/rare signal).
Regarding your veiled threat to report me to an administrator (the impulsive attacks on my talk page have not been conducive to fostering an amicable debate), I really don't care what you do, because I just wanted to have this debate which I have already archived, will link to, and publish widely on the internet, to show how Wikipedia and knowledge in general is declining due to centralization of control, i.e. an increase in Entropy. If experts don't even understand the most basic theorem of sampling theory, then all statistics in world is flawed. I went through this debate over at NIST working group on anti-spam several years ago and got the same type of still unresolved misunderstanding. It amazes me that people can't read clearly what Shannon wrote. Besides I can link to the historical page, when ever I want to refer to the correct introduction of the Shannon-Nyquist sampling theorem. Isn't it in-civil of you to revert my edit three times before this debate on discussion page has been resolved? You accuse me on my talk page of reverting your reverts on my talk page-- circular straw-man. Why keep building your political case to ban me, and just focus on resolving the content debate with less threats on my talk page and thus more efficiency? You can not force me to defend myself here in this limited political jurisdiction, when I can simply supercede your power on the wider internet.
Note the discussion of the sparse signals above is related to the Nonuniform Sampling (see section) which Shannon's paper states uniquely identifies each signal in the 2-dimensional TW space, which is employed by compressed sensing discussed in Beyond Nyquist section. —Preceding unsigned comment added by Shelbymoore3 (talkcontribs) 01:30, 20 August 2009 (UTC)
You can complain all you like, but Wikipedia articles require sources for contestable claims, that's non-negotiable! I'm happy to be proven wrong (anything that enhances or corrects my understanding of signal theory is a good thing), but you're not going to be able to do that unless you provide a source that corroborates your interpretation. If this really is such a fundamental misunderstanding (as you say) then it should be easy to point at an authoritative source that discusses the idea of a lower frequency bound. Oli Filth(talk|contribs) 08:08, 20 August 2009 (UTC)
I am not complaining, I am explaining. See my reply below to Dicklyon about sources. We can simply quote Shannon, no need to find another source. We can remove all my words from the edit. Even if I do find a source which is supporting my understanding, you are still not likely to understand it, if you can't understand what Shannon wrote about T, the "time function" and the 2 dimensional space part of his theorem. Others will be adding comlexity on top of that foundation, and the issue I raised won't come up except for those doing work on non-sinusoid (non-deterministic) signals. Do you have any expertise in that direction? Can you explain to me how your interpretation of Shannon's theory applies to non-deterministic signals? Surely you understand your own interpretation well enough to explain it.
Shannon is saying that the "time functions lie within the interval T". Oli you are correct that for a fully constrained, periodic deterministic "time function" (e.g. sinusoid), then the number of samples is infinite support in the T domain. And thus there is no low frequency bound. These signals sit on a single line in the TW 2D space. But Shannon is also allowing in his theorem for signals which are not deterministically bound to that W domain and can be a 2D point any where in that space. He is saying that the "time function" must be deterministic within the sampling period. That is what he means when he says the 0 outside. --Shelbymoore3 (talk) 13:20, 20 August 2009 (UTC)
I don't know whether the sampling theorem fundamentally changes in the case of random processes; my intuition would be "no"! Again, I'm always happy to learn when I'm mistaken, so if you have a source that describes otherwise, then please present it.
I'm afraid you're not making yourself very clear in your second paragraph:
  • What do you mean by "fully constrained", and how do you relate it to the concepts of "periodic" and "deterministic"?
  • The "number of samples is infinite support" doesn't make sense; Shannon's definition of "a signal that lies in the interval T" is one whose samples outside T will be zero, that's not the same as saying it has finite support.
  • "deterministically bound to that W domain" again doesn't make a lot of sense; a signal can be anywhere in that 2TW space and still be deterministic.
  • I'm not sure Shannon was saying anything regarding deterministic vs. random, so I'm pretty sure that's not what is meant by "0 outside". Oli Filth(talk|contribs) 14:06, 20 August 2009 (UTC)
See the reply I gave to LutzL below. Note that 1 day periodic signal with a bandwidth of 0.1Hz will appear to be a random signal if sampled at an interval T of less than 1 day. Thus even if sampled at 0.05Hz for 1 hour, the sufficient finite support for the 0.1 Hz bandwidth would not provide infinite support for the sampling period interval. This is because the signal's time function is not continuous, so it is not deterministic a priori for any interval less than 1 day (rather it will appear to be random due to aliasing until T is 1 day or greater). By "fully constrained", I mean sampling rate of 0.5Hz and interval of 1 day, per the example I just provided. Obviously if we know the signal is continuous (how do we ever know that in real world?), then 2TW samples fully constrains the signal. Dis-continuous is more salient term here.

I agree with Oli Filth on the inappropriateness of this new interpretation and edits in the lead of this article. Statements like "this is ignored by most texts," without a cited source, are WP:OR and therefore inappropriate. There may be something to Shelbymoore3's point, though I must say I don't see it, but whatever it is, it's not part of the sampling theorem. Furthermore, he's clearly not hearing what Oli said above (e.g. above saying "You are assuming that the composite signal being sampled is a sinusoid" is an absurd reading of what Oli actually wrote); and he's ignoring the polite warnings and attempts to counsel, coming back with personal attacks against a long-time constructive editor. Shelbymoore3, it would be better to slow down, listen better, and learn how wikipedia works, than to bang your head against what is likely to be a pretty hard wall, since I'm not going to tolerate such edits any more than Oli did; the only reason he reverted you was that he got there first. Here's a suggestion: start on the talk page, pointing out what sources say; you've done a bit of that with Shannon, but going beyond what he said and putting in your own nonstandard interpretation of it is not going to work; find a source that supports your interpretation, or give it up. Dicklyon (talk) 06:45, 20 August 2009 (UTC)

  • Your personal opinion of what is appropriate is irrelevant, because you have not quoted Shannon's theorem to support your opinion. What I care is what Shannon wrote, since it is his theorem we are supposed documenting. I have already suggested on this discussion page, that I would accept an edit of my contribution, to discard what I wrote and retain only the exact quotes of what Shannon wrote regarding the period T and the specific definition of "time function", as I quoted Shannon above. That would remove all interpretation and leave it to the reader to decide what Shannon meant. Sorry to disappoint your failed attempt to be condescending, but there is no banging against wall, I have already published a link to and archive capture of (in case you delete my comments) this discussion page to a few million visitor websites, and your censorship will soon be widely known and also widely subverted. You can be proud of having your names on it and be exposed for it.
  • The issue of whether the sampled signal has a deterministic periodic "time function" equation such as a sinusoid is critical, because as Shannon has stated (as quoted below) in his definition of the "time function", we must be able to know the behavior of the signal outside the sampling window T:

To be more precise, we can define a function to be limited to the time interval T if, and only if, all the samples outside this interval are exactly zero. Then we can say that any function limited to the bandwidth W and the time interval T can be specified by giving numbers 2TW.

  • Your demand that I find sources who explain what Shannon meant is really absurd, as we can simply quote Shannon. Shannon is the source for his theorem. I doubt people who are intelligent enough to fully understand what Shannon wrote about the time period T and the definition of the "time function", have bothered to re-explain it, since it is blatantly obvious that Shannon has already explained it. Why would such very busy experts waste their time publishing redundant information, they simply cite Shannon. Perhaps we can find a source with experts who are into sampling Fat tail signals, as maybe they have had to explain their sampling window T in terms of basic sampling theory. Maybe some other reader of this page will be able to help us in that regard. If you are sincere at finding the truth, I suggest that you take the quotes I cited from Shannon, and you explain to me an alternative meaning other that what I have explained? And why don't you quote some sources for your alternative interpretation? Your non-compliance with this request is admission of defeat.
  • I am agreeable if we simply quoted Shannon about the time period requirement and definition of the "time function". So what do you claim is the standard interpretation of those quotes? Oli gave his interpretation at very top which was meaningless. Yes Shannon has stated that T and W' form N dimension space. So what? Why did Shannon mention this? I have told you why. What is your reason for Shannon devoting a whole section III to the geometry of signals in that N dimensional space? Is it because he is defining the limitations of the theory with respect to "time functions" that are not deterministic, employing a very general construct of a N dimension space. If the "time function" is deterministic, then Oli is correct that finite frequency support provides infinite support in the time domain. But my point is that Shannon was more general and is accomodating "time functions" that could be chaotic, e.g. Fat tail. I hope you understand that a "time function" is not required to be periodic and deterministic-- maybe that is the little epipheny you are missing? --Shelbymoore3 (talk) 13:05, 20 August 2009 (UTC)
Why should the time period T be mentioned in this article? This article is about the first three pages of the Shannon paper. The later ones, esp. section III is about what is today known as the Shannon-Hartley theorem on signal transmission rates over noisy channels. And functions that have zero samples outside some bounded intervall don't have anything like fat tails, see Whittaker-Shannon interpolation formula on how such a function looks like. Since the Fourier transform of such a function inside the frequency band is given by a trigonometric polynomial, there is no gap in the spectrum around zero.
The only point of yours that is remotely sensible is that from the samples inside a time interval with length T alone one cannot conclude anything about the function outside this interval, even under the assumption that the function is bandlimited. See again the interpolation formula. Especially there is nothing certain to be said about the Fourier spectrum, not below wavelength T (however, the band limit imposes a lower bound on the wavelengths) and moreso above. (Which by the way is the reason that the so called Huang-Hilbert transform is absurd.) If one includes additional assumptions, fast decay outside the intervall, or zero samples etc. then statements about the spectrum can be more securely quantified.--LutzL (talk) 14:00, 20 August 2009 (UTC)
T is on page 2 where the theorem is. Shannon-Hartley theorem applies to continuous-time channel, so it not applicable to sampling over finite period T, besides it has nothing to do with Fat tail signals. I do not understand the remainder of your point about zero samples.
Yes that my entire point, that "..and whose time functions lie within interval T". Suppose you are sampling an event (e.g. garage door opening) that occurs once a day (don't know that a priori), so if your time interval is less than a day, then you can not be assured to capture one period of that signal. Also note that the signal has a bandwidth W which is perhaps reciprocal of the 10 seconds to open the garage door. So the "time function" of the signal does not lie in the interval less than 1 day. There is a lower bound and an upper bound on the frequency. For intervals less than 1 day, the signal appears to be random, but this is aliasing. Can you deny this simple example of the dual bound? Shannon mentions both of these bounds in the first paragraph of the theorem:

...and that we are allowed to use this channel for a certain period of time T. Without any further restrictions this would mean that we can use as signal functions any functions of time whose spectra lie entirely within the band W, and whose time functions lie within the interval T.

--Shelbymoore3 (talk) 16:05, 20 August 2009 (UTC)
It's not clear what your example has to do with the sampling theorem, which presumes an infinite time and infinite set of samples. If there's a refinement of the theorem to finite T, where can we find that in a source? Shannon is talking about something completely different at that point (channels, not signal sampling). Dicklyon (talk) 16:44, 20 August 2009 (UTC)


(edit conflict) But you should be aware that any signal with limited support can not have its spectrum confined within a frequency band W. Shannon knew this, do you? You should, since this is the next sentence after your quote. And this article as well as the interpolation formula article are concerned with the functions that Shannon goes on to describe: functions that are exactly contained in the frequency band and are small outside the interval, that is falling reciprocal with the distance to the interval. No fat tails, and no gaps in the spectrum. As I said, this article is not concerned with section III, this is Shannon-Hartley or the noisy channel coding theorem. Here we deal with section II, which is complicated enough since some people still think that sine functions are admissible as signals in this context.--LutzL (talk) 16:52, 20 August 2009 (UTC)
The aliases occur due to high-frequency components, not low ones. Your garage door is a square wave (or something similar), with harmonics to infinity, and is therefore not properly bandlimited. If your "signal" were correctly bandlimited, then there wouldn't be a problem (theoretically).
Incidentally, you seem to be blurring the distinction between "random" and "deterministic, but unknown". If the signal to be sampled is random, then the samples will always be random. Similarly, if the signal is deterministic, the samples will always be deterministic. (I'm sure you're aware of that, it's just that you seem to be conflating the two in your recent posts.) Oli Filth(talk|contribs) 18:44, 20 August 2009 (UTC)
Both of you are going off on irrelevant, circular logic straw-men. It is quite simple to understand. First, note that for the example signal I provided above (period 1 day and perfect sine wave pulse width 10 seconds), if T is 1 day and W is 0.1 Hz, then the signal can be reconstructed with no aliasing, if the number of equally spaced samples is 2TW (i.e. 2 x 1 day x 0.1 Hz = 17,280 samples). Do you disagree with the prior sentence?
I assume you agree, thus we need only quote Shannon to show that his theorem applies to the above example:

If the function is limited to the time interval T and the samples are spaced 1/2W seconds apart, there will be a total of 2TW samples in the interval. All samples outside will be substantially zero. To be more precise, we can define a function to be limited to the time interval T if, and only if, all the samples outside this interval are exactly zero. Then we can say that any function limited to the bandwidth W and the time interval T can be specified by giving 2TW numbers.

For the sake of conceptual understanding, ignore any complication of real world impossibility of sampling a sine wave pulse with bandwidth W, we can simply increase W and that does not affect my conceptual point about T. Shannon mentions we should ignore this aliasing issue (we can smooth when we reconstruct with low pass filter):

Although it is not possible to fulfill both of these conditions exactly, it is possible to keep the spectrum within the band W, and to have the time function very small outside the interval T.

The above is a slam-dunk logic. There can not be any possible retort, except to introduce irrelevant straw-men. I have provided an example and quoted the best possible source Shannon, that clearly shows that the sampling interval T has a lower bound for discontinuous signals. For continuous signals the lower bound of T is 1/W only applies if the samples are equally spaced. For discontinuous signals, the lower bound of T is the period of the maximum discontinuity, i.e. of the Fat tail. Will you continue to withhold this important aspect of the theorem from the main Wiki page?
I propose the following be added to the introduction of the main Wiki page, then provide a citation to Shannon's quote above:

For continuous-time signals, the sampling interval T must be at least 1/W for equally spaced samples, otherwise the lower bound is unlimited except by factors such as the resolution of the sampling device. For discontinuous pulse, e.g. Fat tail, signals then the lower bound of the sampling interval T is the period between pulses, and the number of equally spaced samples is 2TW.

However, I will acknowledge that perhaps Shannon was only concerned with the continuous portions of a discontinuous signal, because of what he wrote near the end of section II, but it is my understanding that this was mentioned last because his prior discussion of theorem is fully generalized to discontinuous signals when he mentioned 2TW equally spaced samples ("...spaced 1/2W seconds apart"...), and the following is obviously only applicable to the continuous signals:

The numbers used to specify the function need not be the equally spaced samples used above. For example, the samples can be unevenly spaced, although, if there is considerable bunching, the samples must be known very accurately to give a good reconstruction of the function. The reconstruction process is also more involved with unequal spacing. One can further show that the value of the function and its derivative at every other sample point are sufficient. The value and first and second derivatives at every third sample point give a still different set of parameters which uniquely determine the function. Generally speaking, any set of independent numbers associated with the function can be used to describe it.

The above Shannon quote actually states implicitly that equal spaced samples must be used for discontinuous signal functions, because obviously a discontinuous function has dependent values (= 0) in the discontinuity:

...Generally speaking, any set of independent numbers associated with the function can be used to describe it.

—Preceding unsigned comment added by Shelbymoore3 (talkcontribs) 03:24, 21 August 2009 (UTC)
Further evidence that Shannon was aware that his sampling theorem is applicable to discontinuous signals is contained in section XII. CONTINUOUS SOURCES of his paper. He explained that continuous signals may be broken into discontinuous ones and that the aliasing error (due to high frequencies at the discontinuity) could be quantified and tolerated (e.g. check sums for digital data sent over analog channel):

If the source is producing a continuous function of time, then without further data we must ascribe it an infinite rate of generating information. In fact, merely to specify exactly one quantity which has a continuous range of possibilities requires an infinite number of binary digits. We cannot send continuous information exactly over a channel of finite capacity. Fortunately, we do not need to send continuous messages exactly. A certain amount of discrepancy between the original and the recovered messages can always be tolerated. If a certain tolerance is allowed, then a definite finite rate in binary digits per second can be assigned to a continuous source. It must be remembered that this rate depends on the nature and magnitude of the allowed error between original and final messages.

Dicklyon your point about the signal being infinite in time and T only applying to the channel is irrelevant. Oli your point that my example signal is not contained with the parameters given by Shannon, is not true as my explanation above shows. I think the problem you are having is you have been accustomed to applying what Shannon wrote only to the continuous time domain signals, and as I said in my original edit, that discontinuous signals such as Fat tail are also handled by Shannon's theorem-- the proof is explained above. You don't need to go introducing other theorems about channel noise, as that is irrelevant. Shannon's sampling theorem is applicable to any idealized signal, whether it be continuous or discontinuous during it's period. Shannon is obviously aware of that, by the very general way he wrote his theorem. If you can't bring yourself to understand such a simple concept, then there will be some limit to how many of your misunderstandings I can continue to reply to. I do not say this to be disrespectful, but rather because my time for this is limited. Thank you.
--Shelbymoore3 (talk) 02:25, 21 August 2009 (UTC)
Oli you erroneously claimed that the aliasing was due to not meeting the bandwidth requirement of my example signal, but I can prove you are wrong in 2 ways. First, it is obviously that if the signal is not sampled for at least 1 day in duration, then the pulses will sometimes not even appear in the samples. That is not high frequency aliasing, but aliasing due to insufficient sampling interval. Second, if the bandwidth of the pulse is taken to be W, then the even if I sample at 2TW, then if T is less than 1 day, I will still get the type of aliasing where sometimes the pulse is never showing up at all in my samples-- that is clearly not aliasing due to insufficient W support. --Shelbymoore3 (talk) 04:40, 21 August 2009 (UTC)
If you believe that the pulses not appearing in the sample stream is not due to high-frequency content aliasing, then you sorely misunderstand basic Fourier analysis and the sampling theorem. This is trivially discounted by first running your pulse train through a low-pass filter (which acts as anti-aliasing filter in this case), and then sampling. It may also be discounted by changing your sample interval to either extreme e.g. 1.1 minutes or 1 year + 1 minute; still some of your pulses will not appear in the sample output. Oli Filth(talk|contribs) 10:06, 21 August 2009 (UTC)
Incorrect. The low-pass pre-filter has the same problem as the the sampling device would, in that it can't see any amplitude from the coming pulse until up to T <= 1 day has elapsed. --121.97.54.2 (talk) 17:23, 21 August 2009 (UTC)
What you are describing is time windowing --> filtering --> sampling, which is not the same as filtering --> windowing --> sampling. Oli Filth(talk|contribs) 17:12, 22 August 2009 (UTC)
Transposing "windowing --> filtering" does not change the duration of time your system is going to need to wait before it can output any reconstructed signal. The filter is still going to need 1 day of input on that sparse signal example I had provided. --Shelbymoore3 (talk) 11:14, 23 August 2009 (UTC)
By definition, the filter will have the input from + to - infinity. Oli Filth(talk|contribs) 11:24, 23 August 2009 (UTC)
Infinity is a looong time to wait in real world. That is why my ENTIRE point of the debate still stands that in real-world, the maximum period between sparse events determines our sampling interval (i.e. I think of this as a low frequency bound). And the larger implication of this, is that by definition we do not know the maximum period of Fat tail events. So this means that for Fat tail phenomena, sampling theory tells us that we can not get a clue about the future from sampling an independent Fat tail channel-- possibly only Mutual information can help us. My goal was to get across to the student of Nyquist-Shannon that assuming infinite models apply to real world can be very dangerous, which we will soon all see in our lives:
http://www.professorfekete.com/articles.asp --Shelbymoore3 (talk) 01:33, 24 August 2009 (UTC)
Infinity is fine in my original hypothetical counterexample! You can think of it as a lower frequency bound if you like, but you'd be doing your understanding a disservice, as it's really nothing to do with frequency (at least in the Fourier-analysis sense). Oli Filth(talk|contribs) 19:25, 24 August 2009 (UTC)
(friendly tone again) My point remains that for a sparse event signal where the time-limit is known a priori, your infinite time pre-filter can not convert the time-limit bound into a bandlimited bound. Instead must consider the time-limited bound aliasing in the non-infinite time pre-filter and in the thus non-perfectly bandlimited sampling. Thus time-limit bound is a low frequency bound in the broadest definition of the word "frequency"-- agreed not in the fourier sense but that is irrelevant in my point. Additionally my point immediately below is that for Fat tail signals, by definition the time-limit is not known a priori, thus sampling/measuring itself can be entirely counter productive, unless you have Mutual information.
"Sigh"! That's why I stated that the filter should be placed before the time windowing. Then its output will be truly bandlimited. Doing it in this order will give you a totally different result to doing it the other way round (or not at all). No events will be "lost".
I'm glad we agree that it's not frequency in the Fourier sense; a lot of this discussion thread could've been avoided if you'd stated in that in the first place! Regards, Oli Filth(talk|contribs) 08:39, 25 August 2009 (UTC)
(friendly tone) Some think that measuring is better than not measuring at all, but because they may be under the illusion from an infinite sampling model, they will get just less precision, but in fact the result can be completely opposite than the target signal, i.e. Fat tail in prior paragraph. Students of science are trained to develop a blind faith that infinity can hide in elegant closed analytical form the Second law of thermodynamics trend towards maximum disorder, i.e. maximum information or capacity to do work. In short, science is a faith in a the stability of a shared order-- one that can not be permanent. This is why any universal theory that is not founded upon a world of maximum disorder will never be the final one. This is IMHO why space-time is not the totality of the universe, why Big Band and infinite time are nonsense, as neither describe what is at the "infinite" edge that can never be reached by our perception of order-- disorder. --Shelbymoore3 (talk) 04:01, 24 August 2009 (UTC)
Go and learn some math. Bandlimited signals, about which the theorem only speaks, are infinitely smooth, even analytic and have finite energy=L2-norm and never a finite support. No band-limited signal in that sense is periodic (apart from the zero signal). There is no discontinuity. The aspect of discontinuity in real world signals should be discussed in the more general sampling article. And of course, since the sinc-system is orthogonal, the error in the reconstructed signal is (proportional to) the sum of squares of the samples you leave out. So if anything interesting happens outside the sampling interval, the error will be big. However, if something not to drastic happens far away from the interval, then the error inside the interval will still be small. It's an interpolation formula, after all. And where Shannon speaks of samples outside the interval, he means the points of the equally spaced sequence. Up to today, there is very little certain about unequally spaced (that is, without a periodic pattern) samples. In that regard, Shannon's sentence is like Fermat's last theorem. (See Higgins: "Five stories...")--LutzL (talk) 06:27, 21 August 2009 (UTC)
Correct is no such thing as a perfectly band-limited signal in the real world, and Shannon admits that, as I already quoted and will repeat his quote again:

Although it is not possible to fulfill both of these conditions exactly, it is possible to keep the spectrum within the band W, and to have the time function very small outside the interval T.

What we do is trade a hopefully little bit aliasing smoothed in the reconstruction for the fact that all signals in the real world are somewhat discontinuous. So in my example, just ignore the higher harmonics from the edge of the pulse, as we just smooth those away on the reconstruction. And in fact, we do this for every real signal in the world. So please stop the nonsense about Shannon's theorem not applying to signals that have a weird shape, such as the Fat tail example I provided. I do understand your point that if we modeled my example signal analytically such that it had a period of 1 day and sine wave pulse of 10 seconds with some extremely steep falloff, then it would most definitely be subject to Shannon's theorem and no one here would deny that, then you would have a high band W in order to capture that falloff, but still T would need to be 1 day, else would need a sampling device with near infinite accuracy. In other words, it may be that we could measure an earthquake ahead of time with infinitely accurate (noise free) sampling device, but in practice we can't do it. For Fat tail you will need to lengthen the sampling interval instead. I am sympathetic to your point as whether this discussion applies more generally to sampling and not specifically to the Shannon-Nyquist theorem, but let me ask you to explain how we eliminate Shannon's statement "and whose time functions lie within interval T"? That statement is going to need finite support in the real world, and I don't think we want to say that the most fundamental theorem doesn't apply to real world signals. And I want to point out that Shannon-Nyquist is telling us what are the bounds of our sampling criteria. It is important that people understand that in the real world, the bound is both frequency and some tradeoff of interval, power, accuracy (did I miss any factors)? If people were more aware of this, there would be a lot of less nonsense statistics out there. I am pretty confident quantum theory is a mess because the measuring devices are aliasing. Yeah we never will know what to do about randomly unequal samples on real world signals, unless we have initial data (mutual information). The problem I have in general with your line of argument, is that we live in the real world and Shannon's paper was about a real world system. In this world we live in, nothing is an absolute. Everything (mass, energy, space-time, thoughts, etc) these are just perceptions (some shared through resonance). I know what our space-time is contained in-- disorder, but that is off topic except to make you think a little bit about the pity of the absolute and you pushing away Shannon's theorem to the perfect world that does not exist. --Shelbymoore3 (talk) 08:21, 21 August 2009 (UTC)
In case I wasn't sufficiently clear, I think you are wrong Lutz to claim inapplicability of Shannon-Nyquist to the problem of choosing a suitable interval T. The theorem is all about that. Shannon-Nyquist gives us the initial fundamental understanding of the relationship between W and T. Specifically, the theorem explains that for an idealized signal (perfectly continuous aka analytical aka deterministic aka fully constrained), then the choice of interval T is irrelevant, and that the only requirement for 0 aliasing is we need 2TW samples, where W is the idealized band for our signal. Shannon also explains that real world signals will require us to choose a suitable trade-off between W and T such that the aliasing is minimized:

Although it is not possible to fulfill both of these conditions exactly, it is possible to keep the spectrum within the band W, and to have the time function very small outside the interval T.

So stop telling me that Shannon-Nyquist does not apply to real world signals and stop telling me that it doesn't give us the initial concept of the relationship between the W and T. All I am asking we do, is make this clear to the readers, so they understand that Shannon-Nyquist sets up the initial framework for all the rest of the work on those tradeoffs, e.g. the other theorems you all mentioned about noise, etc.. I do admit that I goofed on my proposed edits, because I tried to frame this tradeoff as a dual bound, whereas it is really a tradeoff. The quantitative tradeoff choice will be affected by factors outside the scope of Shannon-Nyquist, but the initial concept of the existence of such a tradeoff is laid out by Shannon-Nyquist. Again my case is that if we only mention the 2TW and do not mention this tradeoff, then we are leaving out a huge conceptual part of the theorem, because the theorem is for real world signals, as the quote from Shannon above attests. I should hopefully be able to rest my case now, unless you retort with something significantly new or false. --Shelbymoore3 (talk) 09:04, 21 August 2009 (UTC)
I'm not sure if we are getting somewhere with this discussion. In some way I feel that this discussion misses the target Shannon was aiming at. Shannon was not concerned with sampling for reconstruction. His concern was: Given an ideal channel with bandwidth W, that is, any signal that gets into this channel is cut off at this bandwidth by an ideal bandpass filter, that is, if one wants an unperturbed signal, one has to put a bandwith constrained signal function in. Then how many different data points can one pass through this channel inside a time interval T, so that they are exactly recoverable at the other end of the channel. His proposed and proven answer is 2WT data points. And this follows nicely from the properties of the sinc-interpolation formula, which can be proven in different ways. Since one would have to start a bandlimited signal not only at the Big Bang, but at time minus infinity, this idealized modell is not true in practice. So practically one gets less than those 2WT data points. But this is not the concern in section II. The tradeoff mentioned in the quote is to restrict oneselves to exactly bandlimited functions, which are then necessarily not exactly zero outside the interval, but can be assumed to be zero at the sampling points outside the interval. You see, there is no connection to strawmen like "garage doors" or "sine functions" or the new one, "earth quakes", because they have nothing to do with signal transmission.--LutzL (talk) 10:17, 21 August 2009 (UTC)
The choice of a low-pass pre-filter governs the tradeoff between W and T, then refer to what I wrote previously, because thus you did not refute any of my points. Sigh. Btw, the Big Bang is a nonsense, as is the concept of infinite time because order can't be infinite without violating the 2nd law of thermodynamics stating that universe trends to maximum disorder, but that is (not entirely) off topic and I have a paper coming out on that this month. --121.97.54.2 (talk) 17:06, 21 August 2009 (UTC)
Well we are getting somewhere perhaps, if I then ask you what will be the sampling interval for the low-pass pre-filter? You see it is still the same problem, that you in practice need a longer sampling interval for sparse signals. Sampling theory for the real world will not allow you to sample sparse signals with arbitrarily small T. Agree or disagree? Then we have to debate if this off topic of this theorem of sampling theory.
The point from the very start is that 2TW is not a sufficient constraint for sparse or Fat tail signals. There is a practical constraint of the minimum sampling interval, i.e. minimum value for T. Shannon mentions this constraint, because the time-function for the signal must lie within T. If the band W of our sparse signal is 0.1Hz, but the sparseness ranges up to 1 day, then the sparse time signal is not going to lie within T = 1/W. Thus Shannon's theorem is saying we will need to choose a larger duration for T. If we pre-filter the sparse signals to band 1/60*60*24 Hz, then we will loose our signal to aliasing. Excuse me but I am so sleepy my eyes won't stay open, so I am not sure if this will make sense when I read it next. I reserve the right to edit this response in future. --121.97.54.2 (talk) 18:01, 21 August 2009 (UTC)
PROOF: If as Lutz and Oli both suggest, we do low-pass pre-filter at 1/60*60*24 Hz (i.e. 1/day), then 2TW samples means T must be no less than one day, otherwise we get less than 2 samples! A signal can not be reconstructed with less than 2 samples! So there is the UNDENIABLE proof right there! Thus T >= 1/W. If we low-pass pre-filter at some higher band W, then T decreases, but the larger sampling interval was just passed to the low-pass pre-filter, which is simply another sampling device, i.e. you kicked the can down the road but didn't avoid the requirement that "the time function lies within the sampling interval T" as the theorem states. So I think we can conclude now that I was correct, that for sparse signals (Fat tail) the theorem requires T which is no less than the largest period between sparse events. I realize that for an idealized signal (or after you have passed it through your ideal output pre-filter), then 2TW just tells us how many samples we need, not whether they need to spaced evenly over the entire T, but in the real world we can't kick the can down the road, because the pre-filter is subject to the sampling theory also, so no matter how we slice it, then we need T evenly spaced samples (some where in our pre-filter chain). --Shelbymoore3 (talk) 12:28, 22 August 2009 (UTC)

(outdent) I must say that I've lost track at what your argument is any more (maybe you could re-state it concisely?). No-one is suggesting that one is able to reconstruct a signal outside the observation window; that's obviously impossible (in the general case). This is still not the same as saying that there's a "lower frequency bound", which was your original argument. Incidentally, a filter is not a "sampling device". Oli Filth(talk|contribs) 17:19, 22 August 2009 (UTC)

Shelbymoore3, I strongly advise not restating the argument. What's needed is a secondary source from which we can work. It is not OK to restate the sampling theorem based on an idiosyncratic interpretation of Shannon's original work. We quote what he said about sampling, which is the bit that hundreds of secondary sources quote, prove, and expound upon. That's what this article is about. If you want to go beyond that, bring a source that goes that way. Otherwise, please give it up, because more talking here is not going to be productive in the way you like. Dicklyon (talk) 19:47, 22 August 2009 (UTC)
Dicklyon how can you know what I like? I am very very happy with the result of this discussion, irregardless of whether you censor and refuse to quote what Shannon wrote. I note that you guys removed the change to the title that I made from "Misinterpretation?" to "Arbitrarily small T?". You removed the more accurate title which summarized the specific misinterpretation that is discussed in this section. What good does it do to pack every possible misinterpretation into one section? Please change the title to something mutually agreeable and more specific. Your efforts to top-down censor (even on the discussion page) are obvious for all readers to see. Dicklyon as you read below, I simply do not understand the benefit to withholding the full specification of the requirements on the signal, as per an exact quote of what Shannon wrote? Just because the implications on sparse signals does not interest you, you withholding the full theorem meaning from the reader, is a disservice to humanity. I must say you seem to be a grumpy person.
Indeed, the fact that wikipedia arguments make me grumpy is one of my many failings. I'm looking at Shannon's paper here (took me a while to find it since you stuck an inappropriate extra word "sampling" into the quote), and I don't think that sentence has anything to do with the theorem that follows it, which is self-contained and is proved without reference to any interval T. He later counts the samples in the interval T as 2*T*W. I don't see how you get from that to the stuff you've been proposing to add. Dicklyon (talk) 00:37, 23 August 2009 (UTC)
As I explained below, the Shannon quote on the main page about "x(t)" fully specifies the theorem, because the T requirement is implicit if there exists an x(t) that can be sampled. However, as I point out below, the 1st sentence of the 3rd paragraph of the main article has an erroneous summary of the "x(t)" Shannon quote, in essense it removes the requirement of T. You see that requirement on T and time function is implicit in the quote containing "x(t)". If you want to explain that "x(t)" quote, then you must mention that the time function exists within the sampling interval T. --Shelbymoore3 (talk) 01:40, 23 August 2009 (UTC)
I don't get that interpretation, nor have I seen anything like it in a source. In the Shannon paper itself, as well as in sources that talk about this theorem, the reconstruction is generally shown as a sum over an infinite number of samples, which implies the complete lack of the concept of a sampling interval T. See for example [1]. If the interval is finite, perfect reconstruction is not possible; that's not what the theorem is about. Dicklyon (talk) 01:49, 23 August 2009 (UTC)
See next (outdent) below...
As for Dickly's repeated assertion that I have not quoted sources, I will repeat again, that I have only asked that we quote Shannon explicitly. We only need to quote his statement that in addition to the 2TW, Shannon also requires that "the time function lie within the sampling interval T". I do not need another source to quote Shannon. I am saying we don't have to make any interpretation at all, we can simply add this to the bottom of the Introduction section:

In addition to the sampling rate requirement 2TW, Shannon also stipulated that "the time function must lie within the sampling interval T".

Oli I appreciate your tone and your good faith request for me to re-summarize the issues. I think to keep it simple, we could just leave it as the above change required to the main page.
Oli notwithstanding that we keep it simple and just quote Shannon in a short one sentence addition to the end of the Introduction section, I will re-summarize for you. Yes I only propose to say that we can not reconstruct a signal outside the sampling interval, well that is also not exactly correct, so more precisely let's just quote Shannon as per above. There is some misundertanding between you and I on syntax, but we apparently agree on the semantics. The point is that the signal's time function must be deterministic within the sampling interval. The way Shannon stated it, is best. My use of other words, e.g. "deterministic", "fully constrained", "continuous", etc.. lead to misunderstandings on syntax. Let's stick with Shannon's exact quote instead. What I originally mean by "low frequency bound" is that if the signal has a long period of events (e.g. sparse signals, Fat tail), then I view that has a low frequency requirement, i.e. the signal lies outside the sampling interval. I think we got ourselves into disagreement based on syntax, not based on semantics, where we apparently agree. —Preceding unsigned comment added by Shelbymoore3 (talkcontribs) 23:50, 22 August 2009 (UTC)
Oli an analog filter (e.g. capacitance, impedance, and inductance, or their analogs in mechanics) is apparently not a sampling device, but in the case of filtering a sparse signal, the analog filter can not sense the sparse event until it has occurred. So in that respect it has the same sampling interval requirement. And I say "apparently" because actually in the real world a filter does not have infinite resolution, so therefor it is a sampling device-- just the aliasing is usually not significant enough to consider. But for sparse signals, the sampling theory may apply to the filter. My point is made already. You can can stop reading here. Now let me ramble off topic of this page a bit just for entertainment value (well actually it is relevant but in original research only). See one of the problems people have is they think in terms of absoluteness of perception, but perception is only relative to the universe's trend to maximum disorder (2nd law of thermo). Time-space is contained within disorder, it is not an absolute, just look at the complex plane of the Lorentz equations. —Preceding unsigned comment added by Shelbymoore3 (talkcontribs) 00:07, 23 August 2009 (UTC)
I believe Dicklyon's concern was that article talk pages are for discussing improvements to the article, whereas this thread has now diverged into arguing about the subject matter, which is not really the purpose of the talk page. Providing a source would bring a swift end to the matter.
You're right, there is certainly some confusion on the terminology you're using. I think the notion that we all apparently agree is: sampling a time-windowed signal preserves no information about the signal outside the window. Whilst we could add your suggested prose to the article, I believe it would be superfluous, as the article doesn't attempt to address the notion of signals with finite time-domain support, nor even the notion of the dimensionality of the signal space. Oli Filth(talk|contribs) 00:10, 23 August 2009 (UTC)
And no, an LTI filter is unequivocally not a sampling device, has infinite resolution, and never introduces any aliases! Oli Filth(talk|contribs) 00:14, 23 August 2009 (UTC)
Incorrect, an LTI filter only has an infinite resolution given infinite time (continuous-time) sampling interval, and the resolution degrades as the sampling interval approaches the lower limit of the interval within which lies the time function of the signal. Wait I will be reading the main page again to see if your assertion is correct that quoting Shannon is superfluous. --Shelbymoore3 (talk) 00:37, 23 August 2009 (UTC)
What do you understand "resolution" to mean? A hypothetical ideal LTI filter, by definition, operates from + to - infinity. I assumed such a filter when I originally brought it up. Oli Filth(talk|contribs) 00:45, 23 August 2009 (UTC)
I just want to understand how these infinite time models quantitatively interact with the real world, otherwise they are useless to me except as thought experiments towards the useful goal. See below (my prior edit) the Band-limited vs. Time-limited quantitative loss of infinite support (resolution) in the time domain. Infinite time for me is a fairytale that doesn't exist, because the only that is infinite in my model of the world is the trend to maximum disorder. The models that try to hide disorder in infinite time are strawmen, that have to be broken down over time by new science. --Shelbymoore3 (talk) 06:33, 23 August 2009 (UTC)
I have re-read the main page, and I agree that the first paragraph and the Shannon quote fully specifies the theorem, because by definition x(t) can not be defined if it doesn't lie within the sampling interval. I agree there is a need to explain that Shannon quote for the reader. But the problem is the third paragraph is removing the requirement that the "time function lie within the sampling interval T", so we need to fix this sentence on the main page:

In essence the theorem shows that an analog signal that has been sampled can be perfectly reconstructed from the samples if the sampling rate exceeds 2B samples per second, where B is the highest frequency in the original signal.

--Shelbymoore3 (talk) 01:06, 23 August 2009 (UTC)
See next (outdent) below...
Exactly so, because in all sources we know of, that's not what the sampling theorem is about. I'm open to improving the article by the addition of such stuff, but only if we find sources that connect it to the topic of the sampling theorem. Here are places to look. Dicklyon (talk) 00:22, 23 August 2009 (UTC)
Incorrect, according to Wikipedia's policy, we do not need more than one source if that one source is canonical. Shannon is the source for what he wrote. Everyone using his theorem is implicitly using the requirements Shannon gave in the theorem. If everyone was incorrectly using his theorem (i.e. ignoring the requirement that "the time function must lie within the sampling interval T", then we would still have an obligation to point out the part of Shannon's theorem that the mainstream does not use. We are documenting the theorem itself-- try to remember that. The theorem is an orthogonal topic here on Wikipedia. --Shelbymoore3 (talk) 00:37, 23 August 2009 (UTC)
We already quote the theorem in its entirety, and its proof does not require this extra condition, nor is there anything in the theorem or its proof abokut a finite interval T. The exact reconstruction depends on the interval being infinite. With respect to the T and W limitations, Shannon says that "it is not possible to fulfill both of these conditions exactly" and then goes on to write a theorem involving only the W condition. Live with it. Dicklyon (talk) 01:52, 23 August 2009 (UTC)
See next (outdent) below... --Shelbymoore3 (talk) 02:12, 23 August 2009 (UTC)

(outdent) Dicklyon in reply to your claim above that the Shannon quote in first paragraph of the theorem, "time function must lie within interval T", does not apply to theorem and it is not used by anyone, I want you to note that the concise statement of the theorem involves a time function "x(t)". It is obvious to everyone that you can not sample your time function if it does not exist inside of your sampling interval, which is what Shannon wrote "time function must lie within interval T". Nobody on this earth is sampling in infinite time (and besides Shannon's paper is not about sampling in infinite time, it is about a communications system in the real world). So it incredibly obvious that is why Shannon mentioned the requirement, "time function must lie within interval T". In his concise statement of the theorem "x(t)", this requirement is implicit. The 1st sentence of the 3rd paragraph of main article does not say "sampled for infinite time", therefor that sentence is in error and disagreement with the theorem:

In essence the theorem shows that an analog signal that has been sampled can be perfectly reconstructed from the samples if the sampling rate exceeds 2B samples per second, where B is the highest frequency in the original signal.

--Shelbymoore3

--Shelbymoore3 (talk) 02:08, 23 August 2009 (UTC)

The above sentence from the main article is in error, because it states that merely sampling at 2W rate will reconstruct a signal of bandwidth W, which is false if the signal lies outside the sampling interval. So either the sentence has to state that the sampling interval is infinite, or it has to qualify that the signal lies within the sampling interval. --Shelbymoore3 (talk) 03:20, 23 August 2009 (UTC)

Here's what he said:
This "one answer" is a theorem that punts on the finite time interval, since that requirement "is not possible to fulfill", in order to get a "more useful way" of describing what's possible. Of course I agree that nobody in the world samples the infinite past and future. That's no barrier to to a mathematical theorem, though. On the other hand, if you have a source that interprets it differently, I'm all ears. Dicklyon (talk) 02:25, 23 August 2009 (UTC)
I am not disagreeing with the quote of that theorem-- it is complete because it not only works in the infinite case but it is also general enough to imply that if f(t) exists in your sampling window, then the requirement of T has been implicitly fulfilled. Thus Shannon never punted, he took the opening paragraph and put that T requirement in the more concise statement of the theorem. I am disagreeing with the 1st sentence of the 3rd paragraph of main article, which is in error (see my reasons above for why). --Shelbymoore3 (talk) 03:20, 23 August 2009 (UTC)
Let me expound on my reason why that sentence in the main article is inconsistent with the theorem as quoted. I wrote above, "So either the sentence has to state that the sampling interval is infinite, or it has to qualify that the signal lies within the sampling interval". The problem is one of syntax. "Signal" can mean many different things in the context of the 1st sentence of the 3rd paragraph. Shannon was clear (both in his concise statement, and in the paragraph that precedes it) that we must be sampling a signal that has a time function that lies within the interval. In other words, we must have infinite (or near infinite) support in the time domain, i.e. the time function must be deterministic given 2W samples taken any where within the interval. The 1st sentence of the 3rd paragraph removes that requirement, and it thus erroneous. We can fix that sentence very easily as I have suggested above, then we are done here. How hard can that be for you? --Shelbymoore3 (talk) 03:37, 23 August 2009 (UTC)
OK, good point. I just added "(infinite sequence of)". Feel free to remove the parens if you think that's not clear enough or strong enough. Dicklyon (talk) 03:33, 23 August 2009 (UTC)
Thank you very much. I express my humble appreciation. IMHO, your edit is quite sufficient for consistency with the infinite case.
However, we still have the problem that Shannon spoke about how the theorem applies to the real world in his opening paragraph of section II of his paper (as you quoted in above discussion). What do you think about adding another sentence after the one you edited as follows? "If the sequence of samples are not infinite, it is possible (to band-pass pre-filter) to have a bandwidth B for a chosen sampling interval T such that the time function x(t) of the signal is very small outside of T, then 2BT samples suffice.". The point is that in real world we can choose a suitable sampling interval, apply a band-pass filter to constrict the signal to the finite sampling interval, then Theorem applies because x(t) becomes very small outside the sampling interval. Although this is implied by the infinite case, I think we can be more explicit so the reader is not having to be genuis to understand how this is applied to the real world. And our canonical source is Shannon. He told us how to apply the theorem to the real world. Let's tell our readers what he said. --Shelbymoore3 (talk) 04:10, 23 August 2009 (UTC)
Also note Lutz was correct before to write "bandlimited", and I was incorrect to write "low-pass", pre-filter. I have now written "band-pass" above. We have to remove the low frequencies outside of the interval T also. So in end, I was correct, there is a low-frequency requirement in Shannon's theorem when applied to non-infinite signals. Also I admit I have learned the theorem better from this discussion, as evident by what I wrote just above and can now see clearly (analytically) how the opening paragraph is Shannon explaining theorem for non-infinite intervals. Shannon merely states that we can use a finite T by band-pass pre-filtering-- that does not require changing the concise statement of the theorem, it is just a little extra point to help the reader apply the theorem in real world. I hope it is more clear to you now also? Glad we can help each other. That is the spirit of Wikipedia. --Shelbymoore3 (talk) 04:25, 23 August 2009 (UTC)
Yes, that's all good, but let's also stick to the letter of wikipedia, as in WP:V, WP:RS, WP:NOR. Find a source that tells it the way you see it, and then we can consider it. Dicklyon (talk) 04:43, 23 August 2009 (UTC)
Here is a source to explain the granularity (spacing of samples) limit to infinite support in time-domain for time-limited, band-limted real waveforms (at least in quantum realm) is 2TW >= 1/Pi, thus T >= 1/2PiW:
http://en.wikipedia.org/wiki/Bandlimited#Bandlimited_versus_timelimited
I was incorrect to write to "band-pass" pre-filter a few minutes ago above. When sampling in limited time, we need only a low-pass filter to remove the discontinuity (high frequencies) at our sampling interval ends. Thus Oli is correct, there is no low frequency bound. The bound is only that the signal must appear in the time-limited observation window, which for sparse signals means the sampling interval must be greater than the longest period of sparse events (which is what was originally in my mind when I was thinking low-frequency requirement). My point was to make sure the reader understands that simply sampling at 2W without regard to time interval, is not sufficient for real-world signals, because they can not be ideally band-limited and secondly their periodic nature may be more Fat tail than the sampling interval chosen. The sampling interval T has to be chosen in a tradeoff to minimize pre-filter aliasing and to contain the largest period of signals of interest.
The current article already mentions that time-limited signals can not be ideally band-limited, and I would like to suggest we add a link in it to the aforementationed Bandlimited#Bandlimited_versus_timelimited Wiki section:
http://en.wikipedia.org/wiki/Nyquist%E2%80%93Shannon_sampling_theorem#Practical_considerations

a "time-limited" signal can never be bandlimited. This means that even if an ideal reconstruction could be made, the reconstructed signal would not be exactly the original signal. The error that corresponds to the failure of bandlimitation is referred to as aliasing

The weakness with the current page appears to be the lack of coherent connection between the introduction and the 1 sentence about time-limited buried in Practical Considerations sub-section. Also the word "band-limited" isn't mentioned in the intro, so the reader may not make the connection from "highest frequency of B" to band-limited. I think this can be fixed by further improving the 1st sentence in the 3rd paragraph of the introduction to make it more consistent with the theorem, specifically to make it clear that the analog signal is infinite (not just the samples being infinite), so that the reader will be forced to proceed to the Practical Considerations section if they are sampling a real world time-limited signal, and to insert the word "bandlimited", as follows (bold below is only to show you the 2 words I have proposed to be added). This will be my last proposed edit if we can agree:

In essence the theorem shows that a continuous-time analog signal that has been sampled can be perfectly reconstructed from the (infinite sequence of) samples if the sampling rate exceeds 2B samples per second, where the band-limit B is the highest frequency in the original signal.

--Shelbymoore3 (talk) 06:19, 23 August 2009 (UTC)
As I am re-reading the Introduction, it could be rewritten, perhaps even made more concise, to explain that the theorem applies to non-existent Bandlimited signals, which can never be time-limited, thus don't exist (except in mind of a mathematician). This very fundamental point is just not clear in the way the intro is currently written. Imagine the high school student coming to read this most fundamental sampling theorem, and getting extremely confused. I think from the very start, you need to make it clear that this theorem is for imaginary world. Then the Practical Considerations section explain that the theorem can be applied to the real world, by accounting for the aliasing that occurs when approximating a Bandlimited signal in a timelimited interval. The Introduction says "signal (for example, a function of continuous...", but that is really vague for a newbie. It would be must better to be explicit. --Shelbymoore3 (talk) 07:24, 23 August 2009 (UTC)
I provided a suggested edit. Feel free to revert it, but I think you will find it is more concise and coherent. I think the section is now shorter and more explicit:
http://en.wikipedia.org/w/index.php?title=Nyquist%E2%80%93Shannon_sampling_theorem&oldid=309564659 —Preceding unsigned comment added by Shelbymoore3 (talkcontribs) 07:51, 23 August 2009 (UTC)

(outdent)Thanks Dicklyon, Oli Filth, and Lutz. I am done, some edits have apparently stuck (for now) with Oli's refinement. I didn't entirely achieve my objective of making it clear on main page, that we can not predict Fat tail events from time-limted sampling intervals (out-of-scope I guess), but making "infinite" explicitly clear on the main page should make the reader think carefully about how time-limiting signals changes the conclusion of the theorem. I wish we could further qualify "albeit in practice often a very good one" on the main page, but perhaps it is out-of-scope of discussion of the theorem? Just give it some thought, remember a student needs to start on this theorem first, so we don't want them to have any false concepts about time-limted signals being nicely wrapped as a close approximation in all cases by this theorem. I realize it says "some" but you know once the camel gets his nose under the tent... --Shelbymoore3 (talk) 10:32, 23 August 2009 (UTC)

Revisited

(outdent) Please add the following to this discussion, which are my final conclusions: http://goldwetrust.up-with.com/knowledge-f9/book-ultimate-truth-t148.htm#3159 Thank you again very much for all your help and discussion. Shelbymoore3 (talk) 00:14, 8 June 2010 (UTC)

It did very well in explaining your conclusions, Shelby. You should write that book and explain it to everyone. Oli, Dick, sorry I wasn't around to help with this. 71.169.184.208 (talk) 03:48, 8 June 2010 (UTC)
FYI, the examples of error you give in that link are nothing to do with not being sampled "infinitely" in time; they're just examples of sampling non-bandlimited signals without applying an anti-aliasing prefilter. Also, it doesn't make sense when you say "Shannon-Nyquist applies to these filters too" as the filter is still operating in the continuous-time domain. What matters is the spectrum of the output of the filter, i.e. is it bandlimited? Oli Filth(talk|contribs) 07:07, 8 June 2010 (UTC)
Mr "71.169.184.208", at least Oli has provided his name and stated his errors. Oli there does not exist on this planet a bandwidth-limited signal, nor continuous-time domain filter-- these are mathematical abstractions only. In the real world, all filters have aliasing error. For example, certainly the eyes are operating in the analog domain, but yet (per the example I provided) their (time and spatial) resolution is not infinite, so they are not sampling in the continuous domain. Shelbymoore3 (talk) 16:34, 8 June 2010 (UTC)
"Filters have aliasing error" is meaningless unless you're talking about quantum effects or Planck time. Oli Filth(talk|contribs) 16:48, 8 June 2010 (UTC)
Filters do indeed have aliasing error, because no filter has infinite resolution. For the laymen who might read this, continuity means infinite resolution. Whether we apply infinite samples in the discrete phase or in the so called analog (or continuous time domain) pre-filter, there is still a requirement for infinite resolution. That is why it wasn't necessary for me to separate the pre-filter conceptually from the examples I provided in my link. For example, whether the cells of the retina are required to have a response rate that can sample signals of frequency higher than 1/1000 sec, or whether some filter is affixed to the front of the eye that has this response rate, it is still required that we know a priori the Nyquist limit of the signals we want to sample with the eyes or the pre-filter. Any signal outside of the bandlimited resolution of the sampling+filter system, will result in aliasing error. Since we can't know a priori what the signal is, then the conclusion of my link is correct, that science is never certain. Shelbymoore3 (talk) 16:52, 8 June 2010 (UTC)
You're still throwing terminology around carelessly, which makes it hard to determine what you mean. I have no idea what you're referring to when you state that "no filter has infinite resolution", unless you really are talking about quantum effects. A filter doesn't have a response rate in the sense of sampling. It is not meaningful to talk of the "Nyquist limit" of a signal. Equally, "bandlimited resolution" is not a particularly meaningful term. Oli Filth(talk|contribs) 17:01, 8 June 2010 (UTC)
My statement "no filter has infinite resolution" is equivalent to saying that no Ideal filter exists in the real world. In other words, no real world filter has infinite impulse response. Notice that Aliasing and Anti-Aliasing are both linked from the Ideal filter wikipedia article, http://en.wikipedia.org/wiki/Ideal_filter#See_also . The imperfect impulse response rate of a real world filter does cause aliasing, and specifically one needs to be aware of the Nyquist limit in the input signal and whether the impulse response rate of the filter will be sufficient. The wikipedia article for Ideal filter, specifically mentions the Shannon-Nyquist theorem relies on an Ideal filter. Shelbymoore3 (talk) 17:24, 8 June 2010 (UTC)

This isn't what's meant by "resolution"; if you don't use the correct terminology then discussions such as these become impossible. Incidentally, the ideal filter article does not say that they are "subject to Shannon-Nyquist formula", or any words to that effect. Oli Filth(talk|contribs) 17:30, 8 June 2010 (UTC)

Do you have any statement to refute that real world filters do not have infinite impulse response rate and thus do cause aliasing errors? Otherwise I will consider that I have won the debate and we are done. Thanks. You may want to re-read my prior edit. Shelbymoore3 (talk) 17:34, 8 June 2010 (UTC)
Yes, real-world filters can have infinite impulse responses; all linear analog networks do, for instance. What you mean is, no real-world filters have brickwall frequency responses. Yes, no brickwall filters exist, but this neither disproves the Shannon-Nyquist theorem, nor is it a practical problem, because the resulting errors can be made arbitrarily small.
Again, this is sloppy use of terminology and concepts which lead me to believe that you don't really have a solid grounding in this material to have a meaningful discussion, so quite what you mean by "winning the debate" is unclear. Furthermore, this talk page is meant for discussing potential changes to the article, not for arbitrary "debates"! Oli Filth(talk|contribs) 17:43, 8 June 2010 (UTC)
Yes I mean that no real-world filters have brickwall frequency response and thus they create aliasing error if the input signal is not bandlimited a priori. The resulting errors can not be made arbitrary small, because throwing away data does not generate certainty. You can decide that the pre-filter should throw away high-frequency data, in which case the space rocket never existed as it flew past my eyes too fast, or you can decide that the pre-filter should replicate all high-frequency data into the pass band, in which case original signals in the pass band are obfuscated (aliased). My conclusions have nothing to do with disproving or proving the Shannon-Nyquist theorem. The theorem is a mathematical abstraction of the ideal world which does not exist in the real world, and it provides insight into the aliasing error that thus occurs in the real world. Hey I just came to post my final conclusions and to give thanks. You decided to debate my conclusions. This discussion could be about the statement "albeit a a good one in many cases", which I assert in my conclusions is misleading and disingenuous (although true), but I wasn't pushing that case here. Btw, why do you insist on outdenting, please keep the entire discussion about my final conclusions, under the original outdent, so it is clear where it began and ended. It makes me wonder if you are trying obfuscate. And yes, I clearly won the debate, even though you apparently try to frame the debate in terms of terminology instead of conceptual understanding. It is quite clear from the beginning that I mean filters throw away data and that is aliasing. 1+1 = 2, but 2+0 = 2 also. The number 2 is not a reversible operation to the 1+1. Filters are also not reversible. They thus discard information. I am writing for the layman in order they may understand that basic concept, not trying to make formal terminology proof for you. I assume you are smart enough to get to the point of the simple concept. Shelbymoore3 (talk) 18:11, 8 June 2010 (UTC)
If you feel that there's a debate to be "won" here, then that's your prerogative. You continue to misuse terminology (for instance, you still refuse to accept that this is not what is meant by "aliasing"; this term has a well-defined meaning) and it's clear that you don't really understand the basics of signal theory or Fourier analysis, so meaningful discussion becomes a tortuous affair. This entire thread has not provided any of the "incumbents" with any further insight, because the whole thing seems to have been about correcting your misconceptions.
To summarise:
  • "throwing away high-frequency data" is not aliasing, it's a fundamental requirement of an anti-aliasing filter. This is true of both an ideal brickwall filter and a non-ideal version.
  • The anti-aliasing filter does not "replicate all high-frequency data into the pass band"; this is what occurs if you don't use the filter.
  • If the amount of retained information is not sufficient for your application, then you're sampling too slowly. This is true of both a hypothetical ideal system and of a non-ideal version.
  • Your rocket scenario is an example of what happens when you don't use an anti-aliasing filter.
Incidentally, outdenting is commonly used when the indenting gets unwieldy. Continuity is inferred from the use of "(outdent)". Oli Filth(talk|contribs) 18:37, 8 June 2010 (UTC)
Oli I could debate you on the correct meaning of terms ("aliasing", etc), actually I will below in a roundabout way of proof, but such verbose debate would obfuscate the key point I made in my conclusions. Perhaps you didn't realize that you just confirmed one of my original conclusions. One conclusion I made was that the use of a pre-filter (or reconstruction filter), does not provide any escape from the requirements of infinite sampling if the desired band-limit of the original signal is not known a priori. You have just confirmed that. Thank you. Thus, contrary to your assertion, the rocket scenario could not be solved with a pre-filter unless you know a priori exactly the scenario you are trying to capture, and willing to discard other unexpected signals as a trade-off. For example, the pre-filter would need to freeze-frame the rocket ship for a long enough duration for the eye to register the image. Freeze-framing the rocket ship, would mean that if we had 1 million rocket ships sequentially within that same 1/1000 sec, then the pre-filter would only show me one of them (or a blurred composite). Blurred or only 1 ship-- that sort of arbitrary result, sure sounds like an "aliasing" error to me. Okay maybe the blurred pre-filter would be "anti-aliased" in your definition, but in my understanding data has been discarded and any time data is discarded (sampling below the Nyquist limit), aliasing error can result. If you don't want to call this aliasing error, you will still agree it is an error, if the desired result was to be certain of the count the number objects passing in front of my face. Irregardless of the debate on the definition of aliasing, this example provides sufficient refute of your statement that the pass-band is not corrupted by the filter, such that there is no certainty of the count of objects passing in front of my face.
In any case, you have now agreed with me that science is never certain, because we can't know the bandlimit of the signal we are measuring without committing to lose data. And you are now agreeing with me that the pre-filter can do nothing to fix this-- only increasing the sampling rate to infinity can.
So can you still assert that my conclusions were wrong? Can you still assert that science can be made certain by using anti-aliasing filters? Do you now understand why it isn't correct to say that my examples can be solved for all possible input signals by using an anti-aliasing filter (as you originally asserted erroneously)?
Btw, the indenting wasn't unwieldy for me yet (couldn't we have waited until we at least got indented past the center of the screen?), perhaps you are not using a widescreen? EDIT: I see you just added a continuation graphic for the outdent to make it more clear, and thus I thank you and that is sufficient, and I withdraw this line of complaint. Shelbymoore3 (talk) 19:15, 8 June 2010 (UTC)
I do understand your implied point, which post anti-aliasing filter is that (for an Ideal filter) there is no longer in any aliasing error in the pass-band with respect to the information capacity of its bandlimit. The point I am making is that with respect to the information capacity of the input signal, there is still aliasing error. Before you can declare that I am ignorant of sampling theory, you better at least understand that I did not qualify my use of the term "aliasing" as you so erroneously assumed. My use of the term was the broader and more correct use. Shelbymoore3 (talk) 19:33, 8 June 2010 (UTC)
Rather than debating the meanings, I'd rather you just used terminology correctly so that we may both understand one another without needing a meta-debate! It is not correct to state that "aliasing" refers to arbitrary information loss. In fact, the correct use of the term is very apt; it describes the situation (roughly speaking) where high-frequency components become indistinguishable from low-frequency ones, i.e. one is the "alias" of the other. I implore that you don't continue to cobble together technical-sounding phrases (every component of "aliasing error in the pass-band with respect to the information capacity of the input signal" is utterly meaningless).
To the topic in hand, you are still confused; let me attempt to explain. There are two options in your rocket scenario:
  1. Sample the "waveform" directly, with no anti-aliasing filter
  2. Sample the waveform after having passed through an appropriate anti-aliasing filter
In scenario #1, it is quite possible that your discrete-time sample sequence will not register the rocket at all, as you identified long ago. This is precisely due to aliasing, as the "waveform" is not bandlimited. In scenario #2, your sample sequence will represent the low-frequency content in the waveform. This will of course be "smeared", that's the nature of low-frequency content. But it will be correct, and it will guarantee to register the rocket. And this is absolutely *not* aliasing, in any sense of the word. The fact that you've had to remove high-frequency content first is precisely what the sampling theorem states and requires. What you call the "pass band" (I assume you mean high-frequency content) has not been "corrupted", it has been removed prior to sampling.
"Freeze-framing" is a very coarse approximation to what's happening with an anti-aliasing filter; you'd do best not to think of it in those terms, as it only leads to misconceptions.
In your multi-rocket example, the problem (if there is one) is simply that you are not sampling fast enough.
To address far as your rather nebulous "science is never certain" comment; obviously if we have no idea what the characteristics of the thing we're trying to sample are (namely the spectral content) then of course attempting to sample it is foolhardy. No-one with any sense would attempt to do so, in the same way that they wouldn't attempt to the measure the temperature of a furnace with a medical thermometer. Use the right tool for the job, If this is the crux of your argument, then there is no argument, because I agree with you. However it's a very obvious point which is made clear at the outset of the article. Oli Filth(talk|contribs) 19:42, 8 June 2010 (UTC)


I am not confused. And now we are getting to the end of this debate. So you now agree with me that an anti-aliasing filter can not solve the deficiency of sampling rate. With all your tangential verbosity ("meta-debate") about terminology, it boils down to that simple sentence, which is what my conclusion stated.
The use of an anti-aliasing pre-filter can not remove the requirement for infinite sampling, if the desire is to capture all possible input signals. That is what my conclusions stated. I said the scientific method is very useful, because we are able to make some reasonable assumptions about repeatability and expected possibilities in our limited space-time slice of the universe. And I stated that on the scale of universe of possibilities, that science is random, because precisely we can not make those assumptions that apply in our limited domain. In fact, science up to now says the universe is trending to maximum disorder, as is stated in the 1856 law of thermo as documented on Wikipedia.
Why you do feel a need to attack my use of layman's terminology. I am writing this for a wider audience. I understand of course that "freeze-framing" is a simplistic and incomplete summary of a band-pass filter, but it is sufficient to make the point I made and be comprehensible to a wider audience. No you made the wrong assumption as to the meaning of the "pass band", as it is the signal that is output by the band-pass anti-aliasing pre-filter. The output is corrupted relative to the information capacity of the input signal. In other words, the count of the # of rockets passing in front of my face is lost, because the sampling rate of the eye is not sufficient. No amount of band-pass, anti-aliasing, pre-filtering can fix this. I am not confused about this. Maybe you are, but seems you are agreeing with me, so I don't know why you continue to try to say I don't understand.
You are still confused about what "aliasing" means in the broadest definition. The fact that the count of the # of rocket ships has been aliased by the insufficient sampling rate into an incorrect count, is aliasing error. No anti-aliasing filter ("smearing") can help, because the overall sampling rate is below the Nyquist limit of the input signal. This is what the theory says. You are the one is confused. I had this very clear in my conclusions in footnote 1, and then you started an erroneous debate. I am sorry to be so frank, but seriously you can't erase aliasing with an anti-aliasing filter of insufficient bandlimit, you only move the aliasing error around to a more aesthetic ("smeared") condition. In short, "anti-aliasing" is a misnomer. The only way to truely anti-alias, is to increase the sampling rate to above the Nyquist limit.
The next line of argument against my conclusions might be that given I know the bandlimit of my sampling rate, then I can know the uncertainty of my sample. I also refuted this is in my conclusions, because we may not (and will not if we are talking universe scale of possibilities) know a priori in which dimensions (facets) of sampling we are deficient. Shelbymoore3 (talk) 20:17, 8 June 2010 (UTC)

Look, if you aren't prepared to use terminology like the rest of the world does then this endeavour is pointless, and it's not unreasonable to reach the conclusion that you have no idea what you're talking about, or quite possibly that you are simply trolling. You aren't simply "using layman's terminology", you're constructing sentences that make literally no sense if interpreted according to any standard definition of "aliasing" or "information capacity" or "Nyquist limit" or "frequency content" or "pass band" or "deterministic" or "infinite impulse response", etc. There is no "broader" definition of aliasing. So if I don't understand what you're trying to say, then it's your fault for completely abusing terminology!

From the phrasing of your post above, the only reasonable interpretation is that you are still completely confused about the difference between actual aliasing (spectral folding) with standard low-pass filtering.

It's blatantly obvious that no sample rate will be sufficient if the bandwidth of the signal is completely unknown. It's also blatantly obvious that an anti-aliasing filter that doesn't remove aliasing frequency content will be useless. This was never questioned by anyone. But this was never your argument, it was all that nonsense about "infinite time", "time limiting", "lower frequency bounds", "sparse signals" and "fat tail signals". Oli Filth(talk|contribs) 20:38, 8 June 2010 (UTC)

I was writing the following while you were writing the above, so let me post this first.
Further on the debate about the meaning of "aliasing", if we re-sample a 1000x1000 pixel image with a single pixel wide vertical line to a 1x1000 image using the best anti-aliasing filter, the line is still going to disappear (1/1000 grayscale gradients can not be discerned by the eye). I expect your point would be that at least with the pre-filter, then 1/1000 of time we don't get a solid vertical line as the sampling result (the high-frequency will not be aliased into the low-frequency). You would be correct and I agree. What I am saying is that the high-frequency count of lines is still being aliased into the low-frequency, assuming that our sampling system can not discern a 1/1000 count. The anti-aliasing filter accomplished nothing then. My point is that the anti-aliasing filter can not overcome the limitations of the final sampling. If you take my comments in that context, then I think you can understand I was not mis-applying the term. Shelbymoore3 (talk) 20:46, 8 June 2010 (UTC)
Oli, now I will respond to your latest post above. The above explains the broader definition of aliasing. There is still aliasing even after the use of an bandlimiting anti-aliasing filter. The sampling resolution of the amplitude also has to be considered, etc.. If you would prefer to call this error something else, then what term would you like to use? I always understood that any error generated by insufficient sampling (below the Nyquist limit) was aliasing error.
You keep accusing me of abusing terminology and saying I have no idea what I am talking about (an ad hominem attack that I don't agree with), yet you don't disagree with the conclusions I made. How can I be so ignorant, yet you do not disagree with my conclusions. Why should I have to go an explain all your misunderstandings of my use of terminology, when we already agreed that my conclusions are not incorrect. Shelbymoore3 (talk) 21:06, 8 June 2010 (UTC)
No, I would disagree with you on two counts. Firstly, with an anti-aliasing filter the line will never disappear; the result will always be the same. Secondly, it is not the number of lines that may cause them to become indistinguishable; instead, it is how close they are to one another compared to the reciprocal of the sample rate. This is a very important difference (but one which obviously boils down to the same thing if your lines are completely periodically spaced). You've chosen the degenerate limit case as your example, where no matter how far apart the lines are, it's still too close. A single sample can only ever represent a DC component. Oli Filth(talk|contribs) 21:03, 8 June 2010 (UTC)
It's not an ad hominem attack; it's a comment on the content and presentation of your comments here. It's quite possible that you do in fact know what you're talking about, but it's impossible to glean that from your choice of words and arguments.
Of course I don't disagree with what I've referred to as "blatantly obvious" just above, because it is indeed blatantly obvious; you're simply restating the axioms of sampling theory. Everything else you've been saying (both this week and a year ago) is, in all honesty, nonsense. Oli Filth(talk|contribs) 21:12, 8 June 2010 (UTC)
The line will disappear if the sampling system can not register a 1/1000 gradient. You are considering only the output of the filter, not the output of the sampling system (i.e. the gradient my eye can register). On 2nd point, agreed it is the (Nyquist) frequency of the lines, not the quantity, but that is irrelevant to my point that the anti-alias filter will not help you count them if the sampling rate is insufficient. And so I ask you again, what term do you use for this error generated by sampling below the Nyquist limit (even with the use of pre-filter)? Shelbymoore3 (talk) 21:18, 8 June 2010 (UTC)
Well considering the Wikipedia edits that were caused by this discussion have stuck for nearly a year, and considering that the article was highly erroneous (failing to qualify that the theorem requires infinite sampling, as you now say is "blatantly obvious") before adding my edits regarding infinite time and infinite samples, I would say it is more likely I know what I am talking about. Shelbymoore3 (talk) 21:21, 8 June 2010 (UTC)
I am not saying anything else but what you've agreed is blatantly obvious. You are trying to divert the discussion to a debate of terminology, which is unnecessary, since you agreed with my edits to the Wikipedia article and you agreed with my balantly obvious conclusions. Then you went off on this tangent trying to refute my footnote 1 in my conclusions but now we come full circle and you admit that my footnote 1 is blatantly obvious. Who is confused? Shelbymoore3 (talk) 21:35, 8 June 2010 (UTC)
Well considering the Wikipedia edits that were caused by this discussion have stuck for nearly a year, and considering that the article was highly erroneous (failing to qualify that the theorem requires infinite sampling, as you now say is "blatantly obvious") before adding my edits regarding infinite time and infinite samples, I would say it is more likely I know what I am talking about. Shelbymoore3 (talk) 21:21, 8 June 2010 (UTC)
It's now apparent that you've shifted the goalposts to talking about quantisation, which is a completely unrelated concept with completely different underlying maths (although the two necessarily appear together in any real-world implementation). This is absolutely nothing to do with the ins and outs of the sampling theorem or aliasing. Oli Filth(talk|contribs) 21:22, 8 June 2010 (UTC)
If you remember, no-one agreed with you that the article was in any way "erroneous", and certainly not "highly erroneous"! The "infinite sampling" you've been talking about now is not the "infinite sampling" you were referring to a year ago. And what you are referring to as "infinite sampling" today is not "required", nor is it blatantly obvious (because it is wrong!). Oli Filth(talk|contribs) 21:50, 8 June 2010 (UTC)
The 2nd point had nothing to do with quantization. Would please answer the question as to what term you will use for the error generated by sampling below the Nyquist limit, even when the pre-filter is employed? You claim it is not "aliasing". So what is it? This is the 3rd time I have asked you.
Yes the 1st point considers the error of the whole system including the sampling resolution of quantization, which yes is a different form of error called quantization error, but in this case it is being caused by sampling below Nyquist. You see what trouble you get yourself in by compartmentalizing yourself inside myopic definitions? Definition man can never beat conceptual man. Einstein I think was evidence of that. Shelbymoore3 (talk) 21:35, 8 June 2010 (UTC)
It's very difficult to keep track of which two things you're currently confusing at any one time. Error due to insufficient bandlimiting is "aliasing". Error due to insufficient amplitude resolution is "quantisation". Error due to not having a high-enough sample rate doesn't really have a name, as far as I know.
Again, phrases like "the sampling resolution of quantisation" are utter gibberish. And no, the quantisation is nothing to do with "sampling below Nyquist". Oli Filth(talk|contribs) 21:50, 8 June 2010 (UTC)
Okay so you do not know what term to use for error due to not having a high-enough sample rate? That is a critical question, because your only possible criticism of the footnote 1 in my conclusion (as linked above), is that I am using the word "aliasing error". So I need you to tell me what term to use there? What is the correct term? If you don't know the term, then I am telling you that my understanding and what I was taught and read what that any error caused by sampling below Nyquist, is "aliasing error". I understand that you think "aliasing" error is only due to insufficient bandlimiting. Well if that is the case, then there must be another name for insufficient sampling error? I was always taught it was aliasing. And that is why I say there is broader definition. So since you know everything about definitions of technical terms, please tell me the term? If you don't know a term, then I am going to use aliasing error and you have no right to say that what I was taught is wrong, if you do not know what the error is called. Sorry for being so verbose, but I want to make sure you understand me this time.
The fact that in my example we get a 1/1000 quantization level is because the image was sampled below Nyquist. If the image was sampled at 100/1000, then the quantization would be within the quantization resolution of the eye. I inserted the word "sampling" because the process of quantization by the eye, is actually a form of sampling (measurement). I am sorry if the fact that most words in the English language have multiple definitions confuses you. Complain to Websters? Okay I understand your compartmentalization. The anti-aliasing filter outputs 1/1000, and this is accurate. The quantization of the eye causes an error. You are not incorrect. My mind just doesn't bother with the orthogonality, as I see in this case that the problem is insufficient sampling. Raise the sampling rate, then the quantization error is eliminated. Because the reality is I can't alter the quantization capability of the eye, the only thing I can alter is perhaps the spatial resolution of the eye by using a magnifying lens. It is not that I don't understand there are two concepts there, it is just I my mind has eliminated the dependent variables. You know I like to get to the point faster. You have a way of dragging me into meta-debates which are inconsequential to the final conclusions (edits and my footnote 1).
The "infinite sampling" I am talking about now is the same as when the edits were made to the Wikipedia page. And the theorem does require infinite sampling if one is to perfectly reconstruct a signal with infinite bandlimit. The article did not state that before my edits were accepted. That is all I was ever saying, before and now. Some distributions are Fat Tail, meaning they need to be sampled for a long time and some distributions are very high frequency, so they need very high sampling rates. If we do not know the expected distribution a priori, then we do not have certainty, until we have sampled infinite samples (both infinitely small spacing and infinite time, which is the two limits I was referring to that got you so confused last year apparently). It was never nonsense, and it is most definitely not wrong. There were many misunderstandings on terminology (which may lead to think "nonsense"), apparently because your mind operates in a much more compartmentalized fashion than mine does. If no one agreed that the article was erroneous, then why were the edits accepted. Please stop trying to re-spin history in attempt discredit me. The prior discussion is there above for readers to form their own summary. Shelbymoore3 (talk) 22:25, 8 June 2010 (UTC)
Reading back over the recent discussion (since "Revisited" heading), it is clear that I have been very consistently correct in arguing that errors will occur when there is insufficient sampling rate, irregardless of any pre-filter. The apparent source of all this meta-debate is that I used "aliasing error" to mean any such error. This threw you off into an unnecessary meta-debate (which then had confusion over the term "pass band" meaning the signal passed through to the output by the band-pass filter...geez that isn't obvious?). You only needed to clarify for yourself what type of errors I am referring to. Since you don't know any other name for them, I think it encumbant on you to try harder to clarify first before attacking. Shelbymoore3 (talk) 22:44, 8 June 2010 (UTC)
Let's review your first criticism of my recent final conclusions:
Oli wrote "FYI, the examples of error you give in that link are nothing to do with not being sampled "infinitely" in time; they're just examples of sampling non-bandlimited signals without applying an anti-aliasing prefilter."
Okay the above is erroneous statement. Don't you see that now? You've now agreed with me that it is blatantly obvious that applying an anti-aliasing filter will not eliminate the errors due to insufficient sampling. So your assertion that the errors have nothing to do with sampling rate and are caused by not applying a pre-filter, is flat out wrong. It is blatantly obvious, you said so yourself later in the discussion. I think you confused yourself, because you were so focus on "aliasing" can only be caused by lack of a pre-filter. But I wasn't referring only to your "aliasing errors". I was referring to any kind of error due to insufficient sampling.
Oli wrote "Also, it doesn't make sense when you say "Shannon-Nyquist applies to these filters too" as the filter is still operating in the continuous-time domain. What matters is the spectrum of the output of the filter, i.e. is it bandlimited?""
What I wrote is that these filters can not eliminate the errors caused by insufficient sampling rate, that the pre-filters can not get around the requirements for sampling given by Shannon-Nyquist. Again who is confusing the issue? I think it is you. Maybe you can help me clarify a few points, so people don't get confused. But certainly I have only been trying to state what is "blatantly obvious" you claim. Shelbymoore3 (talk) 22:58, 8 June 2010 (UTC)

Brevity for sanity:

  • The fact that there is no particular term for "insufficient sampling" does not mean that it's meaningful to use a term for a superficially similar, but very different, effect. All that will happen is that people will assume you're talking about something other than what you mean, and come to the conclusion that you have no idea what you're talking about (as this thread has demonstrated).
  • That's not my only criticism by a long stretch. I disagree with your use of terminology; I also disagree with what you mean. "Shannon-Nyquist applies to those filters too" is not meaningful, for instance.
  • Again, quantisation is not caused or affected by sampling rate. It's entirely down to the accuracy of the ADC.
  • Again, it may be OK to use terminology loosely in non-technical setting, but you won't be understood here if you continue to misuse terms such as "sampling" to mean "quantisation" and so on. If you don't want the meta-debates, then say what you mean, and use the right terms.
  • No, today you're talking about "infinite sampling" in the sense of "infinite resolution". Last year, you were talking about the need to sample to ±∞. What you don't appear to have realised is that the reason the article was edited was not to address your perceived issues with "fat tail signals".
  • I've already addressed the notion that if you don't know the properties of the signal you're measuring, sampling is a fool's errand. This is very much not a revolutionary concept.
  • "Spin" the history however you like, it doesn't alter the fact that 99% of what you've written to date demonstrates a serious misunderstanding of basic signal theory.

Oli Filth(talk|contribs) 22:49, 8 June 2010 (UTC)

This is tedious to the point that I'm stopping now. You simply won't admit that everything you've said is either confused or a restatement of the obvious, and that you're way out of your depth, so discussion becomes impossible. For that reason, I can only conclude that you're trolling.

In response to your last edit, no the statement is not erroneous. The example(s) on your webpage simply demonstrate insufficient anti-aliasing, not necessarily insufficient sample rate. So I don't believe you now that you've backtracked and said that you were really talking about some other meaning of aliasing (nor do I really care...).

And finally, if all you've been trying to state is the blatantly obvious; (a) why bother? (b) why dress it up with all these red herrings and silly examples of something different? Oli Filth(talk|contribs) 23:09, 8 June 2010 (UTC)

In respect order to yours above:
  • There is a term of "errors due to sampling below Nyquist" and it is called "aliasing error". Since you know of no other term, you will be unable to cite a source to prove me wrong. In fact, I can cite a source to prove you are wrong: http://dictionary.reference.com/browse/aliasing, "The static distortion in digital sound caused by a low sampling rate". I am so tired of your slander, and I demand an apology. I have now cited from the dictionary! Next time learn to use a dictionary when you don't know the multiple definitions of words. It is not my job to teach you vocabulary!
  • Please re-read my linked conclusion and please read what I added above while you were writing yours. Shannon-Nyquist applies to the pre-filters in the sense that the pre-filter can not supercede the need for sufficient sampling.
  • I did not misuse "sampling" and "quantization". I wrote "sampling resolution of quantization". The ADC has to sample the amplitude and quantize it. My gosh how far you go to twist things to try to slander me.
  • I have always been talking (last year and now) about the need to sample at both limits (spacing, i.e. resolution and duration). Do not tell me what I have been saying. I know better than you what I am writing about. Given you can't even bother to open a dictionary, I think it time you stop blaming me for your lack of comprehension skills. I am more than happy to accept constructive feedback to improve understanding, but the slander has to stop. And again the addition of the words "infinite" to the Wikipedia article were necessary to address the possibility of non-bandlimited distributions. Your spin nothwithstanding.
  • Good I am glad you finally admit it. Only took a zillion pages of nonsense to get you to admit that all science makes assumptions about the limits of the input signal. That was what all the discussion about Fat Tails was about. You know Fat Tails are events that no one is expecting (not in their assumptions for limits).
  • Slander, slander, slander. The facts speak for themselves. Shelbymoore3 (talk) 23:18, 8 June 2010 (UTC)
Seriously, give it up. You're just going on the record as a fool who can't admit that they're wrong. Oli Filth(talk|contribs) 23:24, 8 June 2010 (UTC)
There you are trying to spin again. How about deciding, "Here he says it is blatantly obvious that if I don't have sufficient sampling rate, then I won't be able to detect the rocket scenario counts, then he turns around above and says the examples (including the rocket scenario I assume?) only suffer from lack of pre-filter. Which is it Oli? Please choose one.".
The answer is abundantly clear to anyone who has a clue about the sampling theorem. It is apparent that you don't know the answer. Oli Filth(talk|contribs) 23:38, 8 June 2010 (UTC)
It is abundantly clear to me how you act when you've been cornered and have no way out of admitting you lost the debate, other than to use diversionary tactics or quit. That is fine. I will make sure the misunderstandings you had are corrected in the conclusions. I will make it clear that I am talking about errors due to insufficient sampling rate, which dictionary.com says is "aliasing" error. Thanks for sharing your misunderstanding, even if you won't admit it. Shelbymoore3 (talk) 23:43, 8 June 2010 (UTC)
There is no debate to be won or lost here; this could go on indefinitely and you still wouldn't admit that you don't know the difference between sampling and quantisation, or between aliasing and bandlimiting, or between IIR and brickwall, or any other number of fundamental misunderstandings that you're on record as holding. Your arguments aren't even wrong. There is no possibility of forward progress here. Oli Filth(talk|contribs) 23:49, 8 June 2010 (UTC)
None of your assertions above have any bearing on the conclusions I made, and you still refuse to admit that if I am talking about "errors due to insufficient sampling" in my conclusions, then you are agreeing that the pre-filter is irrelevant. You need to choose, is the pre-filter relevant to "errors due to insufficient sampling" or not? As for the terminology side circus you keep wanting to redirect to, it is not relevant to the conclusion in the prior 2 sentences. And I am only concerned about the conclusion, as that is what I posted here today. As for your claims about my conflation of terms and concepts, and how hung up you are on misunderstandings which can be corrected (instead of being constructive, you prefer to slander me), I have already explained myself on that, and I will let readers form their own opinions. Now if you could simply choose above, then we can both agree to end this. I will go and make some improvements to my conclusions paper to make sure the use of the term "aliasing error" does not confuse people (such as yourself) who do not use the layman's definition in dictionary.com. Shelbymoore3 (talk) 00:02, 9 June 2010 (UTC)
Do you seriously believe I do not know the difference between quantization and sampling? My gosh it is only like something I learned as a sophmore in Electrical Engineering. It is simply incredulous the way you focus in on little tidbits and loose sight of the main conclusion. Geez I don't sit here in write in Wikipedia for living like you. I am trying to get a main point across, and you are focused on side circus which has nothing to do with my point. Shelbymoore3 (talk) 00:10, 9 June 2010 (UTC)
I've been as constructive as I can be, but you clearly don't want to learn. Go and read a good book about signal theory. Oli Filth(talk|contribs) 00:06, 9 June 2010 (UTC)
You still refuse to choose and admit you were arguing both sides of the fence. Sure going to look to readers as though you are running away from the simple choice and question. I have read several books on signal theory, although I admit that was 24 years ago.
I am also surprised that you apparently don't know that ADC (analog-to-digital converters) have a form of sampling resolution called quantization. The samples are taken in the amplitude domain, instead of the time domain. Maybe because you apparently have not thought about the way an ADC is actually designed internally. It can be a ladder of resistors and comparators, well each of these is a sample. The sampling is again not in the time domain, but in the amplitude (voltage or current) domain, but nevertheless they are samples that summed together to get the quantization. So it was entirely correct for me to write "sampling resolution of quantitization". For you to assume I am idiot, when I writing something that is based on a deeper understanding of ADC internals than you apparently have. You are incredulous in your slander! Shelbymoore3 (talk) 00:28, 9 June 2010 (UTC)
Assuming there are no more posts from others in this thread, this will probably be my final submission to this thread. I have as promised improved my conclusions document to take into consideration Oli's feedback about potential misunderstandings generated from my use of the word "aliasing": http://goldwetrust.up-with.com/knowledge-f9/book-ultimate-truth-t148.htm#3159 I don't think there are any red-herrings, merely a desire to educate people about the limitations of the scientific method, due to the impossibility of absolute certainty in sampling theory. Many (most?) people subscribe to the religion that science is absolutely certain and absolutely true. Shelbymoore3 (talk) 01:14, 9 June 2010 (UTC)

For what it's worth, I agree with Oli, and find Shelbymoore3's arguments somewhere between incomprehensible and irrelevant. It's possible that with work I could follow them, but I haven't put in much work, because the whole thing is unsourced, and clearly not relevant to the topic of this article, which is the sampling theorem in the sense usually understood and described in sources. Furthermore, the link he provided ([2]) is taking this argument off-wiki, which is highly irregular, and in my opinion reprehensible, behavior of a wikipedia editor. He refers to us there as "experts at wikipedia", which misses the point, since editors are not supposed to be writing based on their own expertise, but on rather on what's found in sources. Dicklyon (talk) 23:15, 8 June 2010 (UTC)

For what it is worth, I feel Oli is trolling. Here he says it is blatantly obvious that if I don't have sufficient sampling rate, then I won't be able to detect the rocket scenario counts, then he turns around above and says the examples (including the rocket scenario I assume?) only suffer from lack of pre-filter. Which is it Oli? Please choose one. Shelbymoore3 (talk) 23:23, 8 June 2010 (UTC)
If it is so blatantly obvious that universe scale spacetime introduces all possibilities and thus science can not be certain about theories that consider the universe scale (the signal can not be bandlimited in universe scope), then what is the harm of stating that implication of the theorem? I only came here to post my final conclusions which I wish were blatantly obvious everyone in the world, but unfortunately are not. And even a few notches below universe scale, we have long tail distributions and very high frequency quantum phenomenon. Shelbymoore3 (talk) 23:30, 8 June 2010 (UTC)
Shelby, do you know what a crank or crackpot is? Do you understand what schizophrenic delusions of grandeur are? Do you have any idea what you, your "Ultimate Truths", and your arguments here look like to other people? You look like a nice guy, but Shelby, you are not well. Now Oli and Dick are nicer than me (sometimes, I can find something to complain about, too, but for the most part, they're nicer than me). They have stuck pretty closely to the issue and refrained from the ad hominem. But, Shelby, really, if you truly had it all together, you would be embarrassed. I'll leave the personal commentary at that point.
Shelby, the sampling theorem is not about all that stuff. Shannon has some more to offer about the broader topic of information theory, but the sampling theorem is pretty much about the fact that the continuous Fourier transform is bijective and that uniform sampling (in, say, the "time domain") causes the spectrum of the continuous-time function that is sampled to be repeated and shifted at regular intervals spaced apart by 1/T. Then if what was sampled was bandlimited to B (meaning that the spectrum is zero outside of -Bf ≤ +B) then if the sampling rate 1/T > 2B, the shifted spectra do not overlap. If they don't overlap (and add), then it's mathematically possible to separate the original spectrum from the shifted images. Since it's bijective, if the original spectrum can be recovered, so can the original continuous-time signal because no other continuous-time signal can be mapped to that original spectrum that we recovered by simply lobbing off the non-overlapping images with a hypothetical brickwall filter.
Now it's all theoretical. No time-limited signal can be perfectly bandlimited and no bandlimited signal can be completely time limited. So, theoretically, we can't perfectly sample anything that didn't exist forever from the past into eternity. But how close to zero does one need to get before one can say "it doesn't matter"? There are a class of functions called prolate spheroidal functions that specialize in being virtually bandlimited and virtually time-limited. But I don't know too much about them, so I'll refer instead to the simple Gaussian pulse which has a self-similar Fourier transform. Now how large (and negative) does the exponent need to be before you'll concede that the function, if measurable at its peak, is too small to measure? If it (and its spectrum) are too small to measure at their tails, you can hack them off (to zero) which makes them sufficiently practically bandlimited and time limited that either can be sampled and the numerical error would be smaller than any finite precision instrumentation (with finite word width) could detect. But all of that are the practical considerations, not what the naked sampling theorem is about. What the naked sampling theorem is about is what's in the previous paragraph. It's nothing more than that.
To try to make it into more than that is indicative of what my first paragraph alluded to. And it's not Oli nor Dick who are intellectually challenged about it. They're just calling your bluff, and they're a lot more patient about it than I would be. 72.95.92.97 (talk) 01:31, 9 June 2010 (UTC)
I love it. I have been awake for about 20 hours, so give me a little time to compose my response to your challenge. It will be delightful, because unlike my quick "off the cuff" responses I gave to Oli, I am going to take some time to formally dig your grave. That will take some time, I hope you can be patient. 121.97.54.2 (talk) 02:03, 9 June 2010 (UTC)
You didn't read what I said. I am not patient.
From Crank (person):
Cranks overestimate their own knowledge and ability, and underestimate that of acknowledged experts.
Cranks insist that their alleged discoveries are urgently important.
Cranks rarely, if ever, acknowledge any error, no matter how trivial.
... [Cranks] misunderstand or fail to use standard notation and terminology,
Also take a look at John Baez's crackpot page.
You can type all you want to yourself. Hope you get off on it. Have a good time, I won't be engaging you on it. 72.95.92.97 (talk) 02:25, 9 June 2010 (UTC)
Actually even half-conscious due to sleep deprivation, it doesn't take much brain power to refute your points. As you state, the theorem is only about the requirement that the sampling rate be sufficient for the desired bandlimit. And my only conclusion is that in the cases where the bandlimit sampled exceeds the expected bandlimit, then the sampling rate will be insufficient and thus errors will result. And these errors can be catastrophic. I provided an example. If there are N rocket ships flying sequentially 2 meters in front of my face at equally spaced periods, such that all N rockets have flown through by my field of vision in less time than my vision system can register one image frame, then there is no bandlimiting filter than can cause my eyes to register the correct count of N rockets. This is true even for some M < N period of image frames. There is no amount of techno-babble that can get around the fact that there will be certain errors that are unavoidable if the sampling rate is insufficient for the bandlimit desired. Now I don't think you will argue against this assertion. Rather you will perhaps raise the point that it doesn't matter, because we know the limitation of the bandlimit that our sampling rate will support. I would agree, except that the point of my conclusions document is that it is the bandlimits that we don't expect, which cause the scientific method to have unknown errors. If we knew the expected spectral bandlimit of every signal we want to measure, then it means we are not making new science, but rather we are just regurgitating what we already know to be true. Obviously to make new science, we have to make theories about what we expect and go test those theories with experiments that employing sampling systems. We arrive at a result that agrees with our theory, then we are able to repeat it, then we conclude that we have certainty. But the reality is we don't have absolute certainty, because we don't know about the errors due to bandlimits that we did not expect, which even our pre-filter can not remove, as per the rocket example. I am eager to read your line of retort, so I can then better focus my analysis where is the main beef of your logic. So far, it seems all you have stated that is relevant, is that the sampling rate must be sufficient for the expected bandlimit. Did I miss a key point of yours? Btw, do you not have a name? It is cowardice to slander someone anonymously. Shelbymoore3 (talk) 02:35, 9 June 2010 (UTC)
As I expected, your non-reply shows you have no point. Come on guys? Don't you have any more bullets to shoot at me from anonymous IP addresses? Can't you find someone of sufficient IQ to give me a real challenge? No! Because anyone with sufficient IQ will realize I am correct. Shelbymoore3 (talk) 02:55, 9 June 2010 (UTC)
After getting a few hours sleep, I re-read the relevant portion of your statement that spans the ellipses in the following quote, "Shelby, the sampling theorem is not about all that stuff ... But how close to zero does one need to get before one can say "it doesn't matter"?". I like your description of the sampling theory, and understand the terminology, in fact I used the term bijective in a recent research document of mine, http://copute.com/dev/docs/Copute/ref/Functional_Programming_Essence.html#Monads, which by the way pertains to project I am working with some lofty real world goals, http://copute.com/dev/docs/Copute/ref/function.html . And I can still assert that you have no point against my "blatantly obvious" conclusions. I explained it sufficiently already in my retort several hours ago, but let me expound a bit now. The fact that the band-limited signal can be reconstructed from the 1/T samples, is in fact in support of my "blatantly obvious" conclusion that 1/T samples are required and that one can not know in the case of in all possibilities (for the input signal) what T should be. Shannon's theorem states (or implies) that for non-band-limited signals, then B will need to be infinite, so that the sampling spacing 1/2B will be infinitesimally small. Even if we arrive at a T that gives us predictable results in support of a theory or objective, this is no absolute guarantee that there does not exist (even repeatable) errors in the result of our sampling, due to assumptions about the band-limit of our input signal which are not true (in cases where we have no way to know our assumptions are not true, unless we increase T sufficiently to some unknown arbitrary level and accidentally discover the error). Oli explained this succinctly, "if you don't know the properties of the signal you're measuring, sampling is a fool's errand". Actually Oli was a little bit too harsh on the scientific method. We can make reasonable assumptions about the band-limit of the information we desire in our sampling system result, and we can obtain reasonable levels of repeatability (i.e. certainty), but Oli is correct in the "blatantly obvious" strict sense (and which is also my conclusion), that if we don't already know the band-limit a priori, then we are a fool to assume absolute certainty in the result. Oli's initial criticism of my "blatantly obvious" conclusion (that he now agrees with, as quoted), is that "aliasing errors" can be eliminated by pre-filtering to the band-limit of the sampling system. This was a misunderstanding, because I was using the definition of "aliasing" as generally used by layman's and as cited by dictionary.com, wherein "aliasing" can refer to any error that results from insufficient T (sampling rate), not just to errors from insufficient band-limiting of the pre-filter. I have now adopted Oli's preferred narrow definition of "aliasing" in my conclusions document, so as to not cause such confusion and discord.
Thus unless someone has more to add, I consider that my conclusions are correct and "blatantly obvious". All the circus about trying to discredit me based on topics which are not central to this conclusion, is just noise for those who enjoy politicizing science, and that includes your link to an asinine points based checklist for quantifying "crack-potted-ness". I always get a good laugh at ignoramuses who try to compile point systems (as if they never grew up and learned that democracy is a lie that is necessarily enslaved to the least common denominator of the bell curve/Gaussian distribution) to quantify the sanctity of vetting and peer review (or just plain sharing and desire to help mankind be humble about science versus God/universe scale, which I know is a topic that incites utter rage amongst some). Shelbymoore3 (talk) 08:54, 9 June 2010 (UTC)
You are probably correct that "there is no bandlimiting filter than can cause my eyes to register the correct count of N rockets." But since that's irrelevant to the sampling theorem, perhaps we should just leave it alone. Dicklyon (talk) 08:10, 9 June 2010 (UTC)
Thank you, and I agree to stop if that is the conclusion. Note I inserted some additional comments above yours, because I was writing them while you were posting, but I don't think they violate or interfere with your statement. It would be nice to close this discussion amicably. Shelbymoore3 (talk) 08:54, 9 June 2010 (UTC)
Indeed; consider it closed. Dicklyon (talk) 09:19, 9 June 2010 (UTC)
Apologies, but in my haste of trying to merge my prior comments and respond to your interim comment simultaneously, I failed to read carefully your assertion that my conclusion is irrelevant to the sampling theorem. I do thank you for admitting that the bandlimiting pre-filter can not eliminate errors due to insufficient T. That is all I need to support my published conclusions. However to your irrelevance point, the sampling theorem states the requirement that the input signal must be bandlimited according to the relationship 1/2B. Thus the sampling theorem is implying that the pre-filter can not overcome errors due to insufficient T. For me, that is enough to say it is relevant to theorem, but I understand the point of orthogonality that you are making. I will thus add that the constructive proof of the theorem characterizes the aliasing error that can result from a non-ideal bandlimiting, and this aliasing error is loosely analogous to the error (e.g. N rocket example) that results from insufficient T, irrespective of whether the signal was bandlimited filtered or not. I will take some time to work through the math again, and see if I can cite any source which would tie the different forms of error together more formally. So agreed, consider it closed, until or unless I could provide a formal source as stated above. Shelbymoore3 (talk) 09:36, 9 June 2010 (UTC)
The following quote from the current Wikipedia article is particularly relevant to the agreement upon which we have closed the discussion and I think also in support of my assertion that the revelevance to my conclusions is implied by the theorem, "The sampling theorem does not say what happens when the conditions and procedures are not exactly met, but its proof suggests an analytical framework in which the non-ideality can be studied. A designer of a system that deals with sampling and reconstruction processes needs a thorough understanding of the signal to be sampled, in particular its frequency content, the sampling frequency, how the signal is reconstructed in terms of interpolation, and the requirement for the total reconstruction error, including aliasing, sampling, interpolation and other errors. These properties and parameters may need to be carefully tuned in order to obtain a useful system.". I will continue to work through the analytical framework, to try to find a more formal classification of the relationship of the different forms of error in the total sampling system. Shelbymoore3 (talk) 10:16, 9 June 2010 (UTC)
Ah ha! The current Aliasing section of the current Wikipedia article, http://en.wikipedia.org/wiki/Nyquist%E2%80%93Shannon_sampling_theorem#Aliasing, actually states that aliasing error can be caused by insufficient sampling rate, "To prevent or reduce aliasing, two things can be done:
1. Increase the sampling rate, to above twice some or all of the frequencies that are aliasing.
2. Introduce an anti-aliasing filter or make the anti-aliasing filter more stringent.". So that voids Oli's assertion that there isn't a broader definition of aliasing. And it voids your assertion that my conclusions about error due to insufficient T, are not relevant to the theorem. I will still continue to search for a more formal mathematical proof of the analytical relationship between the two ways of causing "aliasing" errors. Shelbymoore3 (talk) 10:27, 9 June 2010 (UTC)

No it doesn't "void my assertion". You just haven't understood the above quote correctly. Oli Filth(talk|contribs) 12:33, 9 June 2010 (UTC)

Hahaha, enlighten us rather than hide behind a vacuous slander of what I think (are you inside my head)? State what you think it means or hide. It is your choice. Shelbymoore3 (talk) 12:43, 9 June 2010 (UTC)
How do I request a ruling to my assertion that Oli is trolling and refusing to cite sources or provide sufficient supporting logic for his slanderous allegations of my person? Oli I would appreciate it if you would just state your intepretation and stay away from the trolling type of behavior. You are really degenerating this process. Shelbymoore3 (talk) 12:47, 9 June 2010 (UTC)
If you had understood the quote, you would not have concluded that what I said is incorrect. I'm not going to "enlighten" you, because it's clear that you'll still continue to argue the toss. Suffice to say, is there only one definition of "aliasing" that any technical person would use, and the aliasing article corroborates this. Oli Filth(talk|contribs) 12:51, 9 June 2010 (UTC)
Refute the source I cited, or hide. You are trolling. Shelbymoore3 (talk) 12:54, 9 June 2010 (UTC)
If you have nothing useful to contribute, then please go away. Oli Filth(talk|contribs) 12:56, 9 June 2010 (UTC)
Ditto that. You are unable to refute the source I cited apparently. Checkmate. Shelbymoore3 (talk) 12:59, 9 June 2010 (UTC)
Furthermore, I continue to find sources to refute Oli's prior accusations of my misuse of terms. The Shannon paper itself affirms my correct use of the term "pass band" which Oli slandered earlier, http://www.stanford.edu/class/ee104/shannonpaper.pdf, "In fact, in the frequency-coordinate system those components lying in the pass band of the filter are retained and those outside are eliminated". Let this be a lesson for you Oli about the risk of sticking your head in a noose by making uncited (vacuous) slander attacks. Shelbymoore3 (talk) 13:36, 9 June 2010 (UTC)
My gosh Oli, even Wikipedia knows what Passband means. Why don't you use Wikipedia for terms you do not understand. Shelbymoore3 (talk) 14:27, 9 June 2010 (UTC)
Oli, yeah let's take a little read of the main Aliasing page, "aliasing refers to an effect that causes different signals to become indistinguishable (or aliases of one another) when sampled. It also refers to the distortion or artifact that results when the signal reconstructed from samples is different from the original continuous signal.". So when insufficient sampling rate T is used, and the reconstructed signal is different from the original continuous signal, then that is aliasing. Oli, you wasted a lot of my time with your false accusation that I was misusing the aliasing term. You even caused me to go make unnecessary edits to my conclusion page to remove the word "aliasing" and needlessly separate the errors due to the pre-filter from the errors due to sampling rate. As I told you from the very beginning, the distinction is unnecessary. They are both aliasing errors. Aliasing is any error where the reconstructed signal does not match the input signal. Checkmate dude! Shelbymoore3 (talk) 14:37, 9 June 2010 (UTC)