The amount of free, useful, precise information coming from this channel is remarkable and something to be grateful for. It legitimizes RU-vid education.
It is not "free". Most likely, Professor Brunton has these lectures as one of the deliverables of many of his NSF grants. Thus, this is paid by the US taxpayer. :)
I have absolutely no clue what you're talking about but I love listening. Even without understanding it's very evident you're a talented and efficient teacher.
I always struggle in order to understand deeply what Fourier transform really is, but now after watching your video I'm very confident in what's really is .Thanks a lot
The video is very nice. Thank you! Just a small remark: The indexing of f and f hat in the matrix vector multiplication is wrong. Should count up to f_{n-1} not f_{n}.
@@Eigensteve Or conversely, shouldn't you simply make the summation from 0 to n? Since for f_0 to f_n you now have n+1 sample points, and x is an n+1 size vector. By making your summation to j=0:n, it is summing over n+1 points which is the standard notation used in approximation theory.
@@iiillililililillil8759 you can change summation range if you pull out the j = 0 term and add it in front of your sum :) similar to how it is done in series solutions for certain differential equations
I like your insight that this should actually be called the Discrete Fourier SERIES. Thank you for your way of relating the matrix to the computation. Your perspective help me see how the matrix is related to the tensor and quantum mechanics.
Dear Steve I really enjoy your teaching format and also your wonderful explanation. Just one suggestion, It would be great if you could have at least one practical lecture at the end of each series of lectures, e.g for Fourier series transformation lecture designing one lecture which shows a real problem is great and enhance the level of understanding. Stay motivated and Many thanks for your consideration
I have to give you credit for giving the absolute best educational videos I have ever seen. The screen is awesome, the audio is great, you explain thoroughly and clearly, you write clearly, your voice is not annoying and everything makes sense. Thank you mr sir Steve.
Oh my goodness! Stumbled onto video 1 in this playlist this evening. and I can't stop. Steve, you're amazing. I actually finally feel like I understand what a fourier series is and why it works. can't wait to get to the end. This is easily the best set of lecture on this topic i've ever experienced. HUGE thanks!
Your ability to explain something this abstract in such a simple manner is simply astounding. However i was more impressed by your mirror writing skills. hats off sir..very very good video.. Subscribing to you.
Mr. Brunton. Thank you for clear, concise, organized presentation of DFT. Appreciative of how much time and effort such a presentation / explanation takes to create and deliver. Appreciative of the format you use and precision in getting explanation correct. Explanation of terms and where terms originate has always been helpful in your presentations. Going through the whole DFT, FFT series again to refresh my thinking on the topics. Thanks again. (Erik Gottlieb)
It took me 5min and 55sec to discover that you're writing correctly, I was wondering why are you writing the inverse way! Thank you for the great presentation!
Hi Professor Brunton, Just wanted to let you know I took your AMATH 301 course at UW in 2012. It really kicked my butt but learned so much. I still use the RK4 for work once in a while. You and Prof. Kutz were both outstanding. Wish you both well!
The last time I tried to give a similar lecture I messed up the indexing much more than this, it was a little comforting to see you do it too. It made me wonder if it was worth it to count from 0 always when teaching linear algebra (probably not).
Thanks for the feedback... yeah, I know that when I make mistakes in class, it actually resonates with some of the students. I hope some of that comes through here.
One of my friends posed me an interpolation problem and I instinctively decided to try a DFT. I used some for loops and got the job done, but I never thought that you could build a matrix using fundamental frequencies. That's clean. Then when it came time to using the algorithm, I realized that it was super slow! Granted, it was an interpolation on some 2D data, but still. My laptop couldn't handle an interpolation over fairly small grids (at 35x35, I was waiting seconds for an answer), which blew my mind. But on further inspection, a for loop (or matrix multiplication) is like O(n^2) but likely all the way to O(n^3) after naive implementation details, so it makes sense. What I'm trying to say is, I can see why you think so highly of the FFT, and I'm super excited to learn how it works, and maybe even implement it myself 🙌🏽. You rock, prof!
I *finally* understand it. Memorizing it for exams is not good enough for me, i want to *get* it. Now I do, and see all the great applications for it. Filtering out specific frequencies, isolating specific frequencies, or the same with a broad spectrum of frequencies will be extremely easy with it. Either just calculate a few values individually, or just take/throw away a chunk of the resulting vector. Great videos!
Excellent video! The video was conceptually very clear and to the point. You are an amazing teacher, Prof Brunton! I loved your control systems videos too!
Some videos ago I was concerned at the implications of this being called the DFT, as it not repeating would be problematic for me, and from my understanding of others' implementations, it is supposed to repeat, so I was happy to hear you clear up the easy to make mistake that this was an actual transform and not a series. Things make sense again now. It's still weird that its mislabeled though.
A nice way to think about the mathematical sums, which Prof. Brunton doesn't explicitly mention, is that each of the n+1 rows in the matrix as a vector that functions as a basis function, together which span the space of all n+1 element vectors. Hence all you're doing is taking the inner (dot) product of the original signal with each of those n+1 basis functions (the vectors), i.e. projecting the orignal signal against each of those basic functions to see how much of it is along each of those (vector space) directions.
Heya! I really enjoy the pacing of your lectures. It's also nice for me to get a quick recap of some signal processing before assembling my own lectures. It is also helping me fill in the gaps of knowledge I have around data science, where my training is in Functional Analysis and Operator Theory. This past fall I dug through the literature for my Tomography class looking for a direct connection between the Fourier transform and the DFT. Mostly this is because in Tomography you talk so much about the Fourier transform proper, that abandoning it for what you called a Discrete Fourier series seemed unnatural. There is indeed a route from the Fourier transform to DFT, where you start by considering Fourier transforms over the Schwartz space, then Fourier transforms over Tempered Distributions. Once you have the Poisson summation formula you can take the Fourier transform of a periodic function, which you view as a regular tempered distribution, and split it up over intervals using its period. The Fourier integral would never converge in the truest sense against a periodic function, but it does converge as a series of tempered distributions in the topology of the dual of the Schwartz space. Hunter and Nachtergaele's textbook Applied Analysis (not to be confused with Lanczos' text of the same name) has much of the required details. They give their book away for free online: www.math.ucdavis.edu/~hunter/book/pdfbook.html
I think I'm just going to watch all your videos for my machine learning course this semester instead of my professor's lecture which was so painful and frustrating....
Omg, when I first learned DFT in class I was so confused, but I watched your video and now everything makes sense. Thank you so much. Please continue to make videos!
Thank you,sir. I really got some new knowledge from your videos,which I never know when I studied this theory in my class. Maybe that's because my terchers just want us to understand the theory without applications,but in yout videos I just found a new world of how to use the mothods of math to solve problems in the real world. Thank you again!
Thank you so much for this series of videos. Just a small suggestion; to be consistent, it seems that the vector should have points from f_0 to f_(n-1)
Thank you for the explanation focused on the implementation of DFT. Fourier series makes much more sense to me in general as well! Now I will attempt to code it :)
At 10:49 corrected the matrix size to be n but then the vector size became n+1; needs another correction but I'm still watching! Edit: I saw the same catch in the comments below, but I think the solutions given weren't the best: My solution is as follows: n should be kept the same as it is the number of samples, also the summation should go until n-1 to give n points and nxn matrix size, but the summation formula should contain f_{j+1} keeping everything else the same. This way you don't even need the x_{0} data point. Still liked the video a lot...
Great video. One of the better ones. I wish you explained the exact meaning of the coefficient in the exponent though ... e.g. I never really understood the relationship between sample frequency and number of data points (N). Seems like they will always be the same.
About your App idea, I think you should consider making it for new Teslas because they contain a very powerful FSD computer on which the required DSP blocks can be computed efficiently. in addition to that, Teslas are connected to the internet so thousands of hours of engine sounds can be accessed to feed a NN if a ML-based approach is selected. Anyway, I think you should contact Tesla. They're very innovative and would be glad to hear about this idea.
Tesla Engineers have known that since they were undergraduate students. But they, now, are not quite happy with the idea of given the customers the power to diagnose deeply, that would be quite disruptive for the business; from their perspective: Smart Aleck customers are a pain on the neck.
In DFT, you can tell there's a linear system of equations (whose dimensions are n*inf) that's being solved through inner products, by eliminating all terms except 1 on each equation, since the complex basis vectors are orthogonal to each other. Thats pretty straightforward and intuitive. However, when f is continuous, Fourier treats it the exact same way, which seems wrong, since the e^(iωx) and e^(i(ω+dω)x) vectors arent orthogonal to each other anymore, so even if we use inner product, there will still exist some non-zero 'remainders' on each equation which we cant get rid of. Also, any F.T. of a function f in the [-inf,+inf] domain is problematic, since the inner product of any pair of 2 basis vectors diverges. Do we assume then, that we extend our domain to [-inf,+inf] in such a way that the I.P. remains 0? Unfortunately, noone explains those.
Thank you for the presentation with clarity and intuition. I have a question, @ 9:14 you mentioned something about the fundamental frequency wn. If we are given a piece of signal like you drew, how do we decide what frequencies to look for in that signal? and hence how do we decide what fundamental frequency we can set wn to be? In other words, how do we know if we should look for frequency content from 10 - 20 hz instead of 100-110hz?
it's really complicated. When we use Binary Algebra, we can get s formula of a function almost immediately. This is a formula for the first 16 primary numbers: y (n) = 5 a3 a2 a1 a0 +5 a3 a2 a1+ 5 a3 a2 a0 + 9 a3 a2 + 1 a3 a1 a0 + 5 a3 a1 + 5 a3 a0 + 21 a3 + + 1 a2 a1 a0 + 3 a2 a1 + 1 a2 a0 + 9 a2 + 1 a1 a0 + 3 a1+ 1 a0 + 2
@@samarendra109 I had to pause the video to look in the comments to see if he was writing backwards, It was driving me crazy, small obsessive compulsive attack XD
He is writing on a piece of glass and he flips the video after. He is a lefty, which you can see in his early unflipped videos. His part is also the other way.
Very useful lecture. Thank you so much, Steve! One question by the way, why the number of f hat equals the number of f ? I can't really understand the point here. In my opinion, the number of calculated Fourier coefficients can be different from the one of sampling points.
Sounds like a good question to me. Maybe some of the values are so small that they can be neglected? I'd be interested for him, or someone else who knows this math, to talk about it here in the comments.
In any kind of complex maths explanation, I value preciseness the most. This guy has a good visualization but should have been prepared better if he is interested to make the video helpful.
To understand how important the FFT algorithm is, it helps nations know when other countries are performing underground nuclear tests from anywhere in the world. hope that helps :)
As far as I understand, when we take the inverse discrete fourier transform, we end up with the function values at x_0, x_1, x_2, ..., x_n, but how would you determine what the values of x_0, x_1, ... ,x_n are? I need to know this for my masters thesis please help me if you can.
Hello ! Thanks for your video. I had a question : So if you start with datas from a periodic analogous signal x(t) of period T, frequency w and you want to discretize it with sampling frequency f_s. I know you use DFT but how to you link the frequencies of your discrete and analogue signals ? Is the frequency w_n you're showing here the frequency of the continuous signal ? Thank you !
Good question! There are deep connections between the discrete and continuous Fourier transform, but you can derive the discrete from continuous and vice versa (taking the limit of infinitesimal data spacing).
Hi Steve, do you have a lecture to the connection between fourier series and DFT? their form seem so alike. do they actually connect each other? interpretation wise. Many Thanks!
They do! The important thing to notice is the continuous FT is described as an integral (an infinite sum) whereas the DFT is defined as a finite sum. Otherwise they're almost identical Would recommend 3blue1brown's video on this
Does Mr. Brunton have a more conceptual video on why that fundamental frequency is defined, why we sample it with harmonics proportional to it etc.? Thanks
Hi, great video. Question: you say you multiply the vector with the matrix, but to make dimesions match, shouldn't you multiply the matrix with the vector ?
Hi Steve, at 13:07, if your increase your sample data to 2n, then your DFT matrix first row will be 2n of 1s, and f0_hat will be doubled, is that right? Thank you!
one thing i don't really understand is why there is a "j" in the exponential e^{2\pi1k/n}. Aren't e^{2\pi1k/n} sort of like the basis vectors we are projecting onto? Why do we need to raise each of those to the j's?
This is excellent, thank you very much. A question - does it matter if the spacing between your independent variable samples isn't even/periodic? If it does, how do you approach that scenario?