I wonder how long I can keep this up. All right. So the first thing -- any questions? Any comments? All right. The first thing I want to do is fix a little mistake I made at the end of last time. I was upset about it for days afterwards, so I just want to correct it. This was [inaudible] central limit theorem, which I hope you will agree was a real treat. I mean, itfs just an amazing fact, and the way that it brings in convolution, the [inaudible] transform of the galaxy and so on is really, I think, a real highlight of the course to see those things emerge and to -- the whole spooky nature of repeated convolutions approaching the galaxy, I think, is just an amazing thing. But I did screw it up at the end, so I just wanted to correct that. Everything was fine except for one final formula that I wrote, so Ifm not going to redo it. I just want to point out where I wrote the wrong thing. Itfs correct in the notes. I just wrote the wrong thing on the board. This was the setup -- so X1 through XN are, as they say in the biz, independent and identically distributed random variables. You donft have to write this down. Ifm just recalling what the notation was. And P of little X is the distribution that goes along with all of them because they have the same distribution. Therefs a single function, P of X, which describes how each one of them is distributed. So itfs a distribution for each. And then I formed P of N of X was the distribution for the sum scaled by square root of N. So itfs the average -- excuse me. There was some assumption we made on the Xs on normalation, that is. We assume they had mean zero and we assume they had standard deviation or variance one, and then if you form the sum, the mean of the sum is zero but the standard deviation or the variant center deviation of the sum is the square root of N, so itfs scaled by the square root of N. You divide by square root of N to have this sum, SN, I called it, have mean zero and standard deviation one, and P of N of X was the distribution for this. What we found was that the Fourier transform -- here we go. The Fourier Transform of P of N of S was the Fourier transform of P and S over the square root of N to the Nth power. And then the trick of the analysis -- the way the proof worked was to compute the Fourier transform of P and S over the square root of N just using the definition of Fourier transform that uses a very sneaky thing of looking at the Taylor series of expansion with the complex exponential integrating the terms and so on, and it was really quite clever. What I found was that the Fourier transform of P at S over the square root of N was approximately if N is large was one minus two Pi squared S squared over N. That was fine. That was fine. And then for some reason, I raised this to the Nth power, so the Fourier transform P at S over the square root of N to the N was then approximately this thing raised to the Nth one minus two Pi squared S squared over N to the N and then inexplicably I had the wrong approximation for this in terms of an exponential in terms of the power of E. That is this one minus two Pi squared S squared over N is approximately E to the minus two Pi squared S squared. Ifm going to look that up again to make sure I have it right this time. Sure enough. All right. Then from there, and for some reason I wrote two Pi squared over S squared last time, but itfs two Pi squared times S squared. Thatfs an elementary to the N. Thatfs an elementary fact from calculus that you learned probably a long time ago when you were first learning about properties of the exponential function. And then from there you take the inverse Fourier transform and you get the result. FT of this inverse Fourier transform of this galaxy and E to the minus two Pi squared S squared to get the result, which I am afraid to write down because Ifm afraid Ifm just going to make one more mistake, so Ifm not gonna do it to get the final result of the central limit theorem. There. I feel better now that Ifve corrected it. Ifm not sure everybody else here feels better, but I feel better. Okay. Now to get on to the dayfs events. Brothers and sisters, today we have to come to a reckoning with our failings. Today we have to confront things we have not confronted but must. Now first, the good news, and there is always good news. The good news is the formulas we have derived, the applications we have done are all correct. So relax. Nothing that we have done, strictly speaking, is wrong. However, much has been left unsaid and today, we have to say it. We have to confront the convergence of intervals. It is a sin to consider integrals that do not convert, and Ifm afraid there are times when we have sinned. But in the end, nothing that we have done is wrong, so relax. But much has been left unsaid. So it is time now. Are you ready? Are you ready? Okay. Now we need to have a better, more robust definition of the Fourier transform in order to consider -- I think I gotta stop this. Therefs only so much of this I can do. We need a more robust understanding of the Fourier transform. We need a more robust definition of the Fourier transform that will allow us, verily, to work with the signals that society needs to function -- sines, cosines, even the function one -- constant functions. And the Fourier transform as we have defined it will not do the trick. More robust definition, and thatfs really what it amounts to -- definition of the Fourier transform -- to deal with common signals, ones for which the classical definition does not work or will not -- is not adequate. The issue is exactly the convergence of the integral or if not the convergence of the integral for the function, applying Fourier inversion. The classical definition, let me just say, will not do. There are two issues, really. One is the definition of the Fourier transform itself -- the convergence of the integral defining the Fourier transform -- when I say convergence of the integral, I mean just some way of evaluating, some way of making the integral make sense -- convergence of the integral defining the Fourier transform. And the second and just as important, certainly for applications -- it doesnft do you any good to take the Fourier transform if you canft undo what youfve done. That is to say, you want Fourier inversion to work, also. You want Fourier inversion to apply. It shouldnft be surprising that if therefs a problem with the definition of the Fourier transform, there may be a problem also with Fourier inversion because the formulas are very similar. One involves the integral of either the minus two Pi ISX the interval of the integral of either the plus two Pi ISX but really, the issues of convergence as far as that goes are pretty much the same. Now there are two ways to approach this problem. This is a genuine problem that really has to be dealt with, and it has been dealt with in various ways over various periods of time. Fourier analysis is a highly developed subject, and it has gone through various stages of development. There are two ways of doing this. In the early stages, which are probably a fair way of characterizing it, it was on sort of an ad hoc basis. That is, special techniques were developed for special problems that would explain away a special function or a special signal. Thatfs one way. Two ways of dealing with these problems. That is one is so to speak ad hoc special techniques. Got a particular problematic integral? Okay, let me just work with that integral. Having trouble with convergence of a reticular signal? Having trouble applying Fourier inversion in a particular case? Okay, let me deal with that by some trick, by some method that applies not generally but to that one problem that youfre having difficult with and then you can advance until the next problem comes up. This was okay. This was all that could be done because it was all the understanding at the time led to. The second approach, and thatfs the approach that wefre gonna take, actually, is to rework the foundations. Thatfs the second approach. You have to rework the foundations to give, really, a different definition of the Fourier transform, one that applies more robustly, more equally, somehow, to all signals that come up all at once. You have to rework the foundations and the definition. This came much later. This came probably in the 1940s, and by that time, as I say, Fourier analysis was highly developed, both in its theory and its applications. It was a highly developed subject. This gets into mathematically deep waters, which we will not completely tread, but I do want to give you the confidence enough to swim in them a little bit, just so you know what the issues are and know how they were dealt with. Furthermore, the other thing that I want you to get out of this is a certain confidence that you can compute. The method wouldnft be any good, really, outside of mathematical departments and circles if it didnft offer a way of computing with confidence when that couldnft be done so much before. You can actually use this method to find specific Fourier transforms in a way thatfs airtight and is actually quite nice. Wefll do that. Donft fret over mathematical details. What Ifm going to do and I hope to do is take you to the brink of the abyss and then back off. Say what the issues are, say what the problems are but donft deal with the mathematical proofs and donft deal with all the fundamentals that have to be analyzed, because itfs quite daunting. Therefs a huge amount that has to be done. We donft need that. But I think what you can see is the general structure, the general overall direction to the argument. Let me give you an example to show where the issues are. Letfs go back to the very first example we considered, which already shows the problem. I may have mentioned this before, actually, but now I want to say it a little bit more sternly somehow. The problem is evident in the very first example. The very first example we had was the rectangle function in this Fourier transform. So F of T is the rectangle function, Pi of T. There is no problem at all taking the Fourier transform of that. Thatfs not the issue. The integral converges. Itfs a very simple function. The Fourier transform as classically defined is fine. The Fourier transform of the rectangle function is the sinc function. Sinc of S is sine of Pi S divided by Pi S. thatfs fine. The problem is inversion. The problem is getting the other way around. We did this by duality. There are really two things -- therefs either Fourier inversion or finding the Fourier transform of the sinc function. The problem is concluding that the inverse Fourier transform of the sinc function is the rectangle function. We use duality or more or less equivalently via what we call duality, and that was all fine. Nothing we did was wrong. It wasnft justified. The duality -- what is the same thing, the Fourier transform and the sinc function, is the rectangle function. Same kind of issue. The same issue was involved here in showing this as in showing this. The problem is writing down that integral. The Fourier transform of the sinc function of the inverse Fourier transform of the sinc function -- let me just stick with the problem of Fourier inversion. The inverse Fourier transform of the sinc function is the integral from minus infinity to infinity of E to the minus two Pi IST. Let me write out the sinc in terms of sine and cosine. Sine of Pi S divided by Pi S, and Ifm integrating with the respect to S, so what results is supposed to be a function of T. As a matter of fact, what results is supposed to be the rectangle function. Now I said at the time you want to try to evaluate that integral? Good luck, because itfs not so easy to do. So again, according to what we expect to be true, this integral turns out to be one when S is less than a half and it turns out to be zero when S is greater than or equal to a half. It turns out to be the rec function. Now even this actually is not quite correct because -- well, one thing at a time, here. In fact, by ad hoc techniques, special techniques and there are actually a variety of techniques that can be brought to bear on this, some more complicated than others, you can actually evaluate this integral. But itfs not obvious. No simple integration by par, no simple integration by substitution is going to get you this sort of result. This result is almost okay but requires special techniques. In fact, therefs a problem at the end points. Therefs a problem at plus and minus one with that integral. Itfs equal to one half, I think, when S is equal to plus or minus one half, so thatfs another reason for sort of defining the rectangle function that way so it doesnft jump from zero to one but has a value of one half at [inaudible] continuities. Never mind that. Special technique. So in fact, special problem at the end points -- end points S equals plus or minus one-half. But anyway, the point is that it can be dealt with. Itfs a little bit disconcerting that the simplest example in the entire subject already poses this problem. The most basic function already requires you to do special arguments just so you know the Fourier inversion works. Thatfs a little bit irritating. Oh, Jesus. Yes. There. Good. Donft do that again. Same thing because itfs an even function. There. Now, let me take a second example. Therefs a problem where itfs fine to compute the Fourier transform, but the problem is with Fourier inversion. Computing the Fourier transform of the rectangle function gives you a sinc function. Thatfs fine. But Fourier inversion does not apply directly without some extra work. As a second example of troubles, there are very simple functions for which you canft even get started. I believe I may have mentioned these before, but let me say them again now in this context of repentance and trying to lead a more virtuous life. A second example -- consider F of T equals one, the constant function. Then there is no way to make sense of in a classical way of the integral from minus infinity of its Fourier transform. The integral from minus infinity, either minus two Pi IST times one DT. Therefs no way in which this integral can be made to make sense, period. It just wonft work. I know you know that you write down this integral and you write, oh, thatfs just the delta function. That comes later, and that comes later really only as a convention because to write down this integral is to sin in the sense of expecting this to do anything. This integral does not make sense -- canft be made to make sense. And slightly more generally but equally importantly, other signals do not have classical Fourier transforms in the sense that the integral defines something -- the integral is well defined. For instance, sine and cosine have the same sort of problem. Likewise for F of T equals the sine of two Pi T or G of T equals the cosine of two Pi T -- therefs no way to make sense. You canft make sense of the integral from minus infinity to infinity of the classical Fourier transform. You have the minus two Pi IST, say, sine of two Pi T DT. The integral oscillates. The thing doesnft die out as youfre tending to infinity. Nothing good is gonna happen here. No secret combination of cancellations, no conspiracy of plusses and minuses is going to make this thing work. It just wonft work. These are the signals that society needs. What is to be done? You remember that the situation is somewhat analogous, philosophically, spiritually, heavenly. The situation is somewhat similar to what we confronted with Fourier series. That is, Fourier series was supposed to be a method that applied in great generality. You wanted to be able to write down a very general periodic function in terms of its Fourier series and then compute with it. The problem was those series didnft converge in any sense. You couldnft make those series converge, or maybe you had to make a very special argument to show that they converge and so on. What was finally -- and then finally the subject not so much collapsed of its own weight but a fundamental reworking had to be done. In order to make Fourier series apply as generally as you would like it to apply, you had to really change the notion of convergence. You had to rework the foundations. So itfs analogous. I donft want to make the analogy too sharp or too precise, but in some sense it is analogous, the Fourier series. In the foundations, we needed a new definition, new conception of convergence. Therefs a lot that went on with that. You had to sort of abandon -- if you wanted to advance the subject and if you wanted to talk about things in greatest generality, you couldnft talk about point wise convergence. You had to talk about convergence in the mean and so on. You had to talk about the square integral functions and all the rest of that jazz, quite far removed from the initial development of the subject. Itfs what had to be done, and ultimately led to a very fine theory and a very robust and applicable theory. Now wefre gonna have to do a similar thing here. Again, Fourier series -- the Fourier analysis, Fourier series and then the work on the Fourier transform is a very highly developed subject. You ask yourself if Ifm forced, somehow, because the thing is collapsing. Itfs too many ad hoc arguments, too many one off examples that have to be explained, so I really have to rework the foundations. But itfs such a highly developed subject that you are confronting the problem that any scientist or engineer or mathematician confronts in a situation like that is that nature provides you with so many phenomena that have been so thoroughly studied -- how do you choose the rights one upon which to base your theory? How do you choose a relatively small number of phenomena upon which or based on which the others can be explained? So much has been done in Fourier analysis and applications of Fourier analysis and everything. Where do you start? How to choose the basic phenomena which explains the others which can be used to explain others -- all others. Thatfs the problem. Itfs a problem that confronts anybody doing experiments. Think of this as viewing a series of experiments. What are you measuring? Is that significant? Is it not significant? What do I pay attention to? What do I ignore? The results are there in front of me. What do I do? Which ones do I choose? I want to explain how this was resolved. It was resolved quite brilliantly and quite decisively in the 40s. The first step, and Ifll say a little bit more about the history of this as we go along. The first step in understanding this is to somehow back away from the troubles and concentrate on what works well. Thatfs the first step. So back away from the problems. What is the best situation? Identify the best situation or what characterizes the best situation. In this case, what that means is you want to identify a class of signals for which everything you want is true. You want to identify -- when I say what is the best situation, I mean identify the best signals, the best class of functions for Fourier analysis -- let me just say Fourier transforms. Now the requirements are these. What we want is two properties. We want one. Letfs call this class of signals S. So think of S standing for signals but actually S is gonna stand for Schwartz, the person who isolated these particular class of functions as the best ones upon which to build a theory. So then we want to -- one. If F of T is in S, then the Fourier transform of T is defined and itfs also in S. Fourier transform of F is defined classically by the integral. The Fourier transform of F is also of the same class. Now that already rules out, for example -- that already tells us -- well, I havenft defined S yet. You can imagine that it tells us that the rectangle function shouldnft be in that class and the sine and the cosine shouldnft be in that class or the constant one shouldnft be in that class because for the rectangle function, therefs obviously a problem -- the Fourier transform of the rectangle function is the sinc function. The sinc function is not as good as the rectangle function in some sense because I canft plug it back into the formula for the Fourier transform or the inverse Fourier transform. Likewise with sine and cosine, theyfre not good, because the Fourier transform of the sine and cosine isnft defined or the Fourier transform and the constant function one isnft defined. This already rules out the exact signals that I want to deal with. It doesnft seem like the best class to me, but in fact, it is. Itfs gonna be a restricted class but itfs gonna serve as the foundation for a broader definition. Thatfs the first property. So the function -- if a function is in this class S, then so is its Fourier transform. The Fourier transform [inaudible] so is the Fourier transform. The second property is the Fourier inversion works. Two. Fourier inversion is defined by the classical by the formula, and that only makes sense even to talk about two in light of one because one says if you take the Fourier transform, you wind up back in the class, so you can take the Fourier transform of this function again or the inverse Fourier transform of this function because you havenft left the class of functions for which you can define the integral. So that is to say the inverse Fourier transform or the Fourier transform of F is equal to F and same way going the other way, the Fourier transform of the inverse Fourier transform of F is [inaudible]. Thatfs what we want. This is the best class -- if we can find a class of functions S that satisfies these two properties, then thatfs the best class of signals that we can hope for for the classical theory. For the classical theory, this is the best thing you can hope for. Therefs one other property that comes up that always comes up in the discussion, and I want to just state it now. Itfs not a requirement of the theory, but it comes up as sort of part of the theory and is actually extremely useful for the further developments. Therefs a further property -- let me just say that -- a further property -- an identity that I havenft written down yet. Ifm going to write it down now and wefll make a certain amount of use of it as we go along is whatfs sometimes called Parsevalfs identity for Fourier transforms. We have Parsevalfs identity, also called Raileyfs identity for Fourier series, and therefs a corresponding one for the Fourier transform. Parsevalfs identity is also sometimes called a [inaudible] formula. Itfs even sometimes called Raileyfs identity for Fourier transforms, and it says this. It says that the integral for minus infinity to infinity of if I integrate the square of the Fourier transform, itfs the same thing as the integrating from minus infinity to infinity integrating the square of the function. This is not a requirement of the theory. This is not a requirement of the best class of functions to consider, but it tags along with it. Itfs somehow almost inseparable from it. Ifm going to derive this for you a little later on today. This says this has physical interpretation in various fields depending on where itfs applied. This sort of identity comes up a lot in distinct but related fields, and it says something like you can calculate the energy. You think of the integral of the square of the functions as total energy, so you can calculate the energy in the frequency domain or you can calculate the energy in the time domain. Or sometimes you think of this as a spectral power, total spectral power. It can be computed in the spectrum or in frequency or considered in time and so on. Therefs various interpretations. Therefs various terms that go along with this depending on who youfre talking to. Mathematically, itfs always known as either Parsevalfs identity or Raileyfs identity -- something like that. Itfs analogous of what we had for Fourier series. For Fourier series, we had the sum of the squares of the Fourier coefficients. Thatfs what we get on the left hand side. Thatfs the spectral information and then on the right hand side we had the integral of the square of the function, but we only had it over one period. Here, itfs integrated from minus infinity to infinity because the signals are supposed to live on minus infinity to infinity. Ifll come back to that a little later on. As I said, thatfs not a requirement of defining the best class of signals, but it always comes up. Now, herefs the question -- itfs not such a stretch to say this is what we want. These are the functions that we want. These are the two properties that we really want, our best functions to have. The question is what is such a class and is there a simple enough way of describing it thatfs gonna be useful that I can actually work with and how do you find it? How do you find such a class? Again, imagine the difficulties. Imagine you are a scientist or a mathematician or an engineer confronting this problem and as I say, the field is highly developed. Nature has provided you with so many examples, so many different kinds of signals and so on. How do you choose? What properties of the Fourier transform, what properties of the signals do you choose upon which to base your theory? Itfs hard. Thatfs a hard intellectual challenge. It was met by Laurent Schwartz, who was a very famous French mathematician who taught for many years at the Ecole Polytechnique in Paris. So how to choose, how to define S? Thatfs the question. How do you isolate a small set of phenomena along all possible phenomena that nature has provided in order to build a reasonable theory? This is solved by an absolutely brilliant and inspired work. Genius does what genius does. Who can account for it? This is solved by Laurent Schwartz back in the 40s, and I think itfs significant, actually, that Schwartz taught at Ecole Polytechnique. Ecole Polytechnique is the famous technical school of engineering in France. It was founded in the Napoleonic era to train technically competent Frenchmen -- probably not so many French women at the time -- to become engineers. And Laurent Schwartz taught there, and Laurent Schwartz was although mathematician by training and mathematician by published work was thoroughly grounded and thoroughly familiar with the applications. And as a matter of fact, in his monumental book on the subject which is called the theory of distribution -- these generalized functions that wefre gonna talk about are often called distributions, not to be confused with probability distributions. Anyway, in his book on this, he -- I think in the preface, the first sentence of the preface or something like that, he credits not mathematicians, not even scientists but he credits Oliver Heaviside with really laying the basis for the whole theory. And what hefs thinking of there is delta functions that we now call delta functions and things like that, sort of operational -- what then was called operational calculus. So it was in a very applied context that Schwartz was thoroughly familiar with from which he extracted his theory of what was a more modern foundation of the Fourier transform. Itfs absolutely brilliant and inspired. Itfs very complicated. If you go through all the details, itfs very complicated. Wefre not gonna do that. But if you allow yourself to sort of go along for the ride, itfs actually not so complicated and itfs a pleasure to see it work, and itfs something that you can compute with. I keep coming back to that because I want you to believe me and I want you to realize that Ifm not just doing this for the hell of it. That is, we can actually compute in a very convincing way the Fourier transforms that you really want -- sines and cosines, the delta function, constant function, all that stuff. It really comes out of the theory in a beautiful way, but you do have to accept a certain amount going into it. You have to accept a certain change of point of view and change in perspective. Ifm going to give you the property right now, because it really is fairly simple to state. Ifll tell you sort of where it comes from and why it comes from, but Ifm not gonna go through the details even at that, although more details are provided in the notes. Herefs the definition. S, actually -- so S is now not just standing for signals but standing for Schwartz, is the class of the best class of functions for Fourier analysis. It is the so called class of rapidly decreasing functions. S is class of rapidly decreasing functions. And itfs defined by two properties -- two relatively straightforward properties. First of all, any function of the class is infinitely differentiable. You donft have to worry about smoothness. It has arbitrarily high smoothness. You want to take 1,000 derivatives? Fine. You want to take a million derivatives? Fine. You want to take just one or two derivatives? Fine. Everything is just fine. So F of X is infinitely differentiable. Thatfs one property. That doesnft talk about the rapid decrease. The second definition is the rapid decrease, and what it says is that for any M and N greater and equal to zero integers -- let me write the definition down and then tell you what it means. X to the N -- the Nth derivative of F of X tends to zero as X tends to plus or minus infinity. I donft know whether I switched the rules of M and N here -- what Ifm writing on the board here with what I wrote in the notes, but it doesnft matter. What this says in words is not hard. In words, this says that any derivative of F tends to zero faster than any power of X. M and N are independent here. The third derivative tends to zero faster than X to the 100. Also faster than X. Also faster than X to one million. The 50th derivative of F tends to zero faster than X to the five millionths or whatever. M and N are independent. It says once again in words, and thatfs the way to think of it, is that any derivative of F tends to zero faster than any power of F. And independently -- the order of the derivative of the power. Okay. Where this comes from Ifll explain in just a second. First question you might ask is are there any such functions? I mean, is it terribly restrictive? This certainly does rule out sines and cosines and constant functions. Sorry, tends to zero faster than any power of X. This already seems to rule out the functions we want, or a lot of functions we want. So itfs easy to write down, but it seems highly restrictive. And again, itfs a little bit surprising that one could build such a powerful theory, such a general theory on such a restricted class. But you can, and thatfs the story thatfs yet to be told. Are there any such functions? Well, Ifm happy to write down at least one of them. A galcean is such a function. Any galcean -- Ifll take the galcean that wefve worked with so many times. It doesnft matter how itfs normalized. Any such galcean will work, actually. As it turns out, of course the function tends to zero as X tends to plus or minus infinity, but in fact, the exponential is going down so fast that any derivative of the exponential is also gonna tend to zero faster than any power of X. You know what happens here If I start taking derivatives of E to the minus X squared, Ifm gonna get powers of X out front, but always also -- theyfre gonna be multiplied by this E to the minus Pi X squared, which is gonna kill them of. And in fact, any derivative -- any power of X times any derivative of this function will tend to zero. Check this out. Do some calculations. Do some experiments. Do some graphs. Work it out. Just try it. Ifm not gonna prove that for you, but itfs true. So thatfs -- itfs not a vacuous class. At least there are functions in the class. Thatfs good news. Another important class of functions, as it turns out, are the smooth functions, which vanish identically as you get close to infinity. Not just tend to zero, but actually which are equal to zero beyond a certain finite interval. So C, another class of functions -- these are the infinitely differentiable functions, which are equal to zero, identically zero, outside some finite interval. Different intervals for different functions. Ifm not saying all the functions vanish outside of minus one to one or minus ten to ten or whatever, but a function is in this class if it is infinitely differentiable and it goes, as they say in the biz, smoothly down to zero. So the graph is something like this. You can do whatever it does, but it does smoothly and it tends to zero smoothly. Itfs identically zero beyond some certain point from A to B. And again, different intervals for different functions. I use the word C because the mathematical term for this is functions of compact support. The support of a function is the place where the function is non zero and has a bounded, closed interval, and thatfs called in the mathematical biz compact set or compact interval, and so these functions are called compact support. You will see that terminology certainly in the mathematics literature, and you also see it in the engineering literature. Itfs pretty widely adopted. Thatfs what the C stands for, compact support. And there are plenty of functions like this, too. Itfs not hard to come up with functions like that. So there are -- at least the function of the class S is populated. There are functions which are rapidly increasing. Itfs not a vacuous theory. Where does it come from? Why these properties among all possible properties and phenomena that nature has given you? Fourier analysis is a highly developed subject. Therefs so many things to choose from. Why this? How did this work? Like I say, genius does -- who can account for it? I can tell you what itfs based on. Itfs based on the relationship between smoothness and rate of decrease -- itfs based on the derivative formula. That turned out to be sort of the fundamental property of Fourier transforms upon which this theory could be erected and the other things could be understood. The connection here comes through the derivative theorem. There are various versions of the derivative theorem, but the one that we looked at before was the Fourier transform of the Nth derivative -- I used M before, so let me use M here -- was at S is two Pi IS to the M times the Fourier transform of F to S. Therefs actually a similar formula for the Fourier transform of the derivative of the Fourier transform and so on. I wonft write that down. Again, these are in the notes. We used this when we talked about the heat equation. I said this was actually a fundamental property of Fourier transforms. It turns differentiation into multiplication. The derivative formula [inaudible] high enough elevation if youfre looking at it relates smoothness -- that is to say the degree of differentiability to multiplication -- to powers of S. So saying something like this quantity tending to zero, meaning a power of S times in this case the Fourier transform tending to zero says something about how differentiable the function is. Itfs a relationship between differentiability and measurements in terms of powers of S of rates of decrease. Ifm not really gonna say anything more about it than that. That plus about a 250 IQ will lead you to develop a fantastic theory. But anything short of 250 IQ probably will leave you short. But this turned out to be the derivative theorem and explaining the relationship between differentiability, between the degree of smoothness and the sort of rate of decay turned out to be the crucial thing upon which this thing hinged. So it was sort of the connection from differential equations, and actually again, thatfs not so -- that comes out a lot in the applications. We didnft do a lot of this, but in Heavisidefs work, for example, where he introduced the so-called operational calculus delta functions and things like that, it was often in the context of differential equations -- of solving differential equations, and he would work on solutions of differential equations by taking transforms. Differentiation turns into multiplication under many of these transforms. Youfve seen probably when you were a kid the [inaudible] transform. It has the same sort of property. Other [inaudible] transforms have a similar kind of property. Again, it was one of the phenomena that was out there, but why this phenomena and not other phenomena? Thatfs what it took a genius to see. I want to do one more thing today, and then Ifm gonna show you next time how wefre gonna use this and how wefre gonna proceed from this very restricted class of very smooth functions with very good properties to functions which are very wild and things that are not functions at all, like deltas and things like that and how wefre gonna use this to actually define the Fourier transform for objects which are extremely general. But let me take this opportunity to finish up today with a derivation of Parsevalfs identity. Itfs off the direct trajectory that wefre talking about now, but it comes up often enough, and itfs somehow always allied to the theory that I think itfs worthwhile just going through the derivation. So Parsevalfs identity for functions in class S -- for functions for which there is no issue of convergence, anything. Parsevalfs identity says again, the integral from minus infinity to infinity F of S squared DS is equal to the integral from minus infinity to infinity of F of T squared DT. And by the way, for functions in the class S, there is no question about convergence of these intervals, because the functions are dying off so fast that the integral of F squared converges. And furthermore, because the class has the property that if you take the Fourier transform you stay within the class, this integral also exists. This integral also exists. I should have said that, actually. Part of this point here is that not only does the fact that this satisfies the properties that wefre interested in is the function satisfies the property of rapidly decreasing, also therefore Fourier transforms do, also, and thatfs where this relationship also comes in. Sorry, I should have been a little bit more explicit about that. Wefll come back to it. Ifm actually gonna prove a slightly more general version of this, and itfll take me but a moment. Actually, itfs quite a straightforward derivation. The advantage is that everything is justified. All the manipulation Ifm gonna do of integrals is completely justified. So let me prove a slightly more general one, namely the integral from minus infinity to infinity of the Fourier transform of S times the Fourier transform of G bar DS, an inner product of the Fourier transforms is the same as the inner product of the functions -- integral from minus infinity to infinity F of T times G of t bar DT. So I allow for complex value functions here, and of course the Fourier transform generally is a complex value. So how dose this follow? Well, let me use X here instead of T, not that it matters. Ifm gonna invoke Fourier inversion. That is to prove this, Ifm gonna write G of X as the inverse Fourier transform of the Fourier transform. Ifm gonna write G of X as the integral from minus infinity to infinity of E to the plus two Pi ISX, the Fourier transform of G of S DS. I can do that because I can. In this case, I can. If I assume that F and G are in class S, Fourier inversion works. Everything is fine. Okay. Then what about the integral from minus infinity to infinity of F of T, F of X, G of X bar DX? Well, G of X bar -- sorry, let me do it over here. G of X bar is the integral from minus infinity to infinity. I take the complex conjugate of the inside. So thatfll be E to the minus two Pi ISX, the complex conjugate of the Fourier transform S DS. So I take F of X times G of X bar integral minus infinity to infinity, thatfs integral from minus infinity to infinity F of X times the integral from minus infinity to infinity E to the minus two Pi ISX F G of X bar DX and then the result is integrated with respect to -- DS, sorry -- the result is integrated with respect to X. Now, in respect to X. Now, let the rigor police through the door. I donft give a damn because everything here converges in every possible good sense. I can interchange limits of integration. I can whip them suckers around with absolutely no fear of retribution. That is to say, brothers and sisters, that this is equal to the integral from minus infinity to infinity, the integral from minus infinity to infinity -- letfs put them all together and then take them all apart -- E to the minus two Pi ISX F of X, G of X, F of G of S bar DS DX and now, thatfs putting all the integrals together. Now I take the integrals apart and turn that into an iterated single intervals again, and I can do that because everything converges in every possible good sense. You donft scare me, you rigor police, you. Minus infinity to infinity of F of G of S bar integral from minus infinity to infinity E to the minus two Pi ISX F of X DX the result integrated with respect to S. I swapped the limits of the order of integration. You see, whatfs inside that integral is nothing but the Fourier transform of F. This is equal to the integral from minus infinity to infinity, the Fourier transform of G of S bar the Fourier transform of F of S DS and I am done. Where do we start? Where do we finish? We started with the integral of the inner product -- the integral of the Fourier transform of F times the Fourier transform of G -- no, I didnft. Herefs where I started. I started with -- there I started! Okay. Up there. I started with the integral of F times G bar and then ended with the integral of the Fourier transform of F times the Fourier transform of G bar and that is what I wanted to get. Ainft life good? And it all works and itfs all perfectly justified because the class S is the perfect class for all Fourier analysis, and next time youfre gonna see how we get from what seems to be a highly restricted class to all the functions that society needs to operate. All that and more, coming up.