top | item 11207183

Ten lessons I wish I had learned before teaching differential equations (1997) [pdf]

338 points| JMStewy | 10 years ago |math.toronto.edu

118 comments

order

mightybyte|10 years ago

I couldn't agree with the last point more.

> A course taught as a bag of tricks is devoid of educational value. One year later, the students will forget the tricks, most of which are useless anyway. The bag of tricks mentality is, in my opinion, a defeatist mentality...In an elementary course in differential equations, students should learn a few basic concepts that they will remember for the rest of their lives...

I hated the DE cleass I took in college and it was largely because I felt like it was nothing but a bag of tricks. I very distinctly remember one problem that seemed unsolvable until the teacher showed that you had to substitute a "2" with "1/2 + 3/2". And then, to make matters worse, he put the exact same problem on the test. So we were being rewarded, not for really understanding the core basic concepts, but for memorizing the tricks needed to solve specific problems.

JackFr|10 years ago

I was working as a programmer at the NIH for a senior scientist who was an MD & math PhD. In the course of the work he was doing, we came across a differential equation that needed solving. My undergraduate math degree was not enough to crack it in a day. I tried Mathematica, which also choked on it (probably should have tried Mathematica first.) My boss spent a couple of hours and came up with no explicit solution. I was preparing to move forward with numerical approach, but my boss made a phone call to a friend who came into his office and copied the problem off the whiteboard.

Two days later he had an explicit solution based on two really non-obvious (bizarre) substitutions. I came away very, very impressed. The guy earned a 'we are indebted to ...' footnote in the paper.

I guess the point is someone has to come up with the tricks.

orbitingpluto|10 years ago

Nothing so quickly emphasizes that a bag of tricks can turn into actual math more than a linear algebra course. In a math degree you can literally take the same course twice. At a first year level you will learn to perform all these tricks with matrices and get a hint of inner product spaces. In 2nd or 3rd year you'll do the reverse and justify why matrix algebra works in the first place.

But the harder a subject, the longer it feels like learning a bag of tricks. The first partial differential equations course feels like you are working on only 3 problems for 4 months.

edit: I had a prof whose first DE course was at the graduate level. At the oral exam he was asked to give an example of a differential equation and all he could do was point to phi on the blackboard.

batbomb|10 years ago

I'm not sure what that problem would have been, but an integral part (no pun intended) of DEs is expansions. It's not part of a bag of tricks, it's a very common technique used to solve a problem.

thoth|10 years ago

>I felt like it was nothing but a bag of tricks... had to substitute a "2" with "1/2 + 3/2".

Many techniques in math are "tricks" like this. Think of solving a quadratic by completing the square, or integrating by substitution or integrating by partial fractions, etc. You could arrive at these techniques on your own, but that is a lot of trial-and-error, deep understanding of theory, and applying it, which all takes a huge amount of time. Meanwhile, previous mathematicians figured this out and we get to benefit from their work. ;)

Maybe your instructor didn't present it well - plopping out the answer without a good enough explanation of the technique, why it works, etc.

maaku|10 years ago

I had an almost identical experience with my ODE class. This is exactly when I started hating math, and moved out of computer science (which was part of the math department at the university I was at).

It took me a long time to return to learning higher maths, this time on my own according to the needs of my job, which is far less an ideal environment than University.

If ODE contains "a few basic concepts that they will remember for the rest of their lives" I never learned it, and I wonder what those concepts would be.

chairleader|10 years ago

Indeed, Rota's approach to teaching seems to mirror the application of mathematics in general...

He acknowledges that concepts have prerequisites, and that certain topics are misleading or dead ends. You might even call the act of building a pedagogical scope and sequence for a topic graph traversal!

I wish I could have taken one of his classes. I wonder if there are other mathemetician/philosophers out there who have taken the problem of teaching as seriously as the math itself.

vpribish|10 years ago

you just made me angry at all my college math classes again. grr.

did anyone have an undergrad math education that was NOT like this?

conceit|10 years ago

on the upside it made some people feel very smart

mynegation|10 years ago

My major is computational math, from 15 years ago, from leading Russian university, so it is just anecdata, and by no means should be generalized.

I absolutely love mathematics, for me it is the embodiment of pure beauty. Still, I positively, absolutely hated the sophomore course of ODEs. The way it was taught was extremely abstract: here is the equation, this is integration, this is separation, this is your SLP, now go deal with it.

It was totally pointless and life-sucking. It was not until I got to the 3rd year and learned about specific applications in physics (like heat dissipation, strings, and springs), and later in finance (stochastic calculus) and biology (e.g. Lotka-Volterra) when I realized how many wonderful and extremely useful applications they have.

Have this course started with that, things would be completely different.

my5thaccount|10 years ago

I took, we called it Diff-e-q, my freshman year in college and the professor died a couple weeks into the course and the new guy was not pleased about having to teach it. We did the applied aspects with the springs and the little ants running away from a candle heating the corner of a plate.

What I didn't like about it was just all the memorization. I had no desire to memorize a bunch of formulas that I knew full well in the real world I'd look up in a table or type into a computer. What I wanted to learn how to do was solve problems using math, not memorize patterns of formulas to apply to problems.

So I didn't memorize them and instead went to work and earned money to pay for college. Still passed the class but it was one of my lowest grades. It's a hard class even without the memorization.

jimmyswimmy|10 years ago

Strongly agree. The way ODEs are taught at the sophomore level violates the beauty of math by teaching a plug-and-pray method of solution. For this form of equation, try this form of solution, if it doesn't work, try this one, then this. If none work, oops. At least once the Laplace method is taught things get a little better.

lamontcg|10 years ago

Heh, all these years I thought it was just me.

Hated sophomore ODEs.

Much later took a mathematical physics course that actually taught me ODEs. Also took a year of applied mathematics that really solidified PDEs and taught me complex analysis. Took another graduate course in calculus of variations that was useful. Another very senior-level physics course taught me greens functions, method of steepest descent, etc.

The Physics and Applied Math profs were vastly better at teaching Math than the Math profs. Only real problem was in the freshman Physics courses where they taught you sloppy vector calculus before you'd seen that from the freshman calculus courses (and generally if you tried to pick up Math directly from the Physics courses that were teaching concepts it was all sloppy physics math -- it worked but you never quite understood why...)

agumonkey|10 years ago

I wonder if there are people that really click and appreciate purely abstract definitions of mathematical concepts. Many books and courses (especially in Europe it seems, watching U.S.A classes, Strang's for instance, feel a lot more pragmatic and applied). Sure after a few years of confusion one might finally click and realize all that was hidden behind a lemma, an identity or a symbol, and then that definition goes from painful to beautiful.

bitL|10 years ago

I hated ODEs until I took MIT's robotics course and had to solve non-linear ones using non-linear optimization techniques. Since then I love them ;-)

graycat|10 years ago

In the OP, the author Gian-Carlo Rota started out with:

> One of many mistakes of my youth was writing a textbook in ordinary differential equations. It set me back several years in my career in mathematics. However, it had a redeeming feature: it led me to realize that I had no idea what a differential equation is.

Wow! Good to see that he wrote this. Looking at his book,

Garrett Birkhoff and Gian-Carlo Rota, Ordinary Differential Equations, Ginn and Company, Boston, 1962.

I got the same impression! I couldn't see what the heck they were driving at. Instead, they seemed to flit around with a lot of tiny topics of little or no interest for little or no reason.

Want to understand ordinary differential equations, read Coddington:

Earl A. Coddington, An Introduction to Ordinary Differential Equations, Prentice-Hall, Englewood Cliffs, NJ, 1961.

Then for more, to make such equations much more important, read some deterministic optimal control theory, e.g., Athans and Falb

Michael Athans and Peter L. Falb, Optimal Control: An Introduction to the Theory and Its Applications, McGraw-Hill Book Company, New York,

that, BTW, also has some good introductory, but very useful, material on ordinary differential equations.

More generally, want to know what to study in a subject that will be useful? Okay, one approach is to go to more advanced material that is an application of that subject and see what that material emphasizes for prerequisites, e.g., sometimes quite clear in an appendix.

E.g., Athans and Falb say quite clearly what is important in ordinary differential equations for their work.

tomekowal|10 years ago

Please, write another textbook.

The internet changed drastically the process of writing books. I saw people making profit from books available online for free. I saw books written chapter by chapter with errors found quickly by first readers. I saw systems that allow commenting parts that are not clear enough with comments how to clarify them.

If you promise to deliver a textbook that teaches skills relevant to engineers, they will fund the time it takes to write the book. I would spare couple of dollars even if you said that it will take 5 years. I believe, there will be even companies that will give you funds upfront. If you reach out for help there will be people who will help you collect example problems from different fields to replace couple of "salt tank problems".

I am not a mathematician, so I don't know how mathematics textbooks are written and how much effort goes into them, so feel free to point out that this idea is stupid.

Terribledactyl|10 years ago

Rota died almost 17 years ago, I'm sorry.

jimhefferon|10 years ago

> I don't know how mathematics textbooks are written and how much effort goes into them

They are written in many ways by many different people.

But some of us have started to write books that are Free, in the sense that software is Free. I have a couple and in addition to making the text and the source available I also sell one of them on Amazon and it does OK (see http://joshua.smcvt.edu/linearalgebra), because lots of people prefer a paper version when they really get down to studying.

p4wnc6|10 years ago

I'm proud to report that at my undergrad institution, Rose-Hulman Institute of Technology, they very successfully adhered to these rules (and I was taking ODEs there way back in 2005).

They had a custom textbook created for their 2-course ODE sequence that several of the faculty collaborated on. Though it did contain content on uniqueness theorems and some proofs, far and away the biggest two items hammered in were (a) linear equations with constant coefficients, and (b) Laplace transform methods.

They also offered (at the time) a 3rd, optional course called Boundary Value Problems that was focused on several physics-motivated BVPs like with Laplace's equation, heat equation, wave equation, Young's modulus, and others, and that course heavily used Fourier and Laplace methods.

We did have word problems, but they were almost exclusively "salt tank" problems. Literally, every word problem described a tank of water or pre-mixed brine solution, with some description of either more salt or more water being added or removed, either gradually or in discrete injections.

The fact that every problem was an infamous "salt-tank problem" essentially made its status as a word problem irrelevant. This seems like it wouldn't be that helpful but actually it was really nice. You got so used to the different pieces that comprised the modeling problem that when you went off and did something in other courses, like circuit systems or conservation systems in mechanical engineering, you knew how to translate the problem to 'salt tank' form, which really covered a huge range of practical problems.

As a math major, one fault I noticed of this method was that it did not make the connections to linear algebra very clear. It took me another few semesters afterward to catch up on that part, but I can understand how engineering majors cared less about that.

I don't know what Rose-Hulman does for this curriculum now, but it would be cool to somehow take a "snapshot" of their methods for it and compare it with other experiences like this OP.

tnecniv|10 years ago

> It took me another few semesters afterward to catch up on that part, but I can understand how engineering majors cared less about that.

It's definitely something engineers care about, but not at the undergraduate level.

analog31|10 years ago

I once told a math teacher at a Big Ten university, that I thought their undergrad math instruction for engineers was weak. As an example, I said that I didn't think students learned any engineering applications of differential equations. He looked at me with a straight face and said: "There are no engineering applications of differential equations."

zerohp|10 years ago

My undergrad was at a Big Ten university. The analog signal processing course in EE teaches how to use differential equations to solve circuits. Right after they taught it, we learned laplace transform and never looked at a differential equation again.

agf|10 years ago

That is really, really hard to believe. I can't say that I remember any discussion of applications in my differential equations course (also at a Big Ten school), but I'm positive the professor could have provided them.

I didn't have a particular need for examples in a course, as differential equations were held up as the holy grail of math by my father, an optical engineer -- he used them at work fairly often, and frequently they were what made him better able to solve a problem than an engineer who wasn't comfortable using them.

gk1|10 years ago

I took the differential equation course over a summer, which means an entire semester was condensed into four weeks. I passed the class but I couldn't tell you a single application for differential equations at the time.

Don't know if the professor could've done anything better or if they had no choice but to plow through it because of the time constraints.

giardini|10 years ago

"He looked at me with a straight face and said: "There are no engineering applications of differential equations."

This made me laugh; I think he wanted you to laugh too.

Some DE courses could be better described as _histories_ of the applications of mathematics. Consequently they appear to be little more than a patchwork of tricks and hints. My DE professor, as adept with applications as theory, weaved this patchwork together quite skillfully and held our interest.

BTW his DE skills paid his bills extremely well (oil companies have lots of DEs to solve), while the university position allowed time for theory and was frosting on the cake.

lagadu|10 years ago

That answer doesn't surprise me in the slightest. I'm a software engineer so I forgot everything taught in those classes the moment I turned the exam in but I've a friend who's a professor of something that mechanical engineers take related to fluids (don't ask me details, we don't even live in the same country any longer); at one point he told me that anything other than numerical analysis is absolutely worthless in the real world because you can't find analytical solutions for almost anything.

drjesusphd|10 years ago

> I do not know how to properly motivate the Laplace transform

I feel like this is impossible without going to the complex plane. Like the author said, taking the inverse Laplace transform is no joke.

I feel like I never properly understood the Laplace transform until I learned about Landau damping. This is when waves exist, but are damped in a collisionless plasma. This damping is not disspiation and the energy does not get converted into heat. The usual way of presenting this is to show that if one Fourier transforms in time, you get the wrong answer. The fact that the system begins at a certain state, and is thus an initial value problem, needs to be respected.

semi-extrinsic|10 years ago

The Laplace transform is somehow the continuous analog of a Taylor series expansion. You don't need complex analysis to motivate it. I sketched this for my students when TAing once upon a time, heavily inspired by this [1] nice MIT lecture.

I'm very surprised this isn't standard material. It makes the parallel between Laplace and Fourier transforms so much more intuitive, because you get Taylor series as a parallel to Fourier series.

[1] https://youtu.be/zvbdoSeGAgI

Kinnard|10 years ago

"We are kidding ourselves if we believe that the purpose of undergraduate teaching is the transmission of information. Information is an accidental feature of an elementary course in differential equations"

graycat|10 years ago

Yes, and once students discover this, then good luck in getting students in the class!

And, if regard the material on differential equations as essentially nonsense, then good luck getting NSF grants for research in the subject!

Actually, can communicate a lot of good information in a course in differential equations, but to do this apparently need some exposure to some of the leading applications of differential equations.

fdej|10 years ago

In my opinion, the first course on ODEs shouldn't be concerned with the transmission of information.

That is where the wave equation comes in, in the first course on PDEs.

backprojection|10 years ago

> Some thirty or so years ago, Bessel functions were included in the syllabus, but in our day they are out of the question. > Teaching a subject of which no honest examples can be given is, in my opinion, demoralizing.

I don't get this. Differential equations theory is about proving existence and uniqueness of solutions. If you have to use numerical techniques to actually compute the solution, then that's perfectly fine. After all, even if the solution is explicit, like sin(x), or especially a special function, then we still need to use numerical techniques to actually evaluate that explicit solution.

has2k1|10 years ago

The article is not questioning the "theory", the identified problem is the teaching at undergraduate level. As it was/is commonly taught, it is neither pure nor applied. Rather it is categorisations and tricks, little of which has any practical value.

As a body of work Differential equations are so messy that theorists landed on so many disparate results. As such Differential equations courses are commonly taught as a "survey of the land" type of courses, so they tend to be incoherent. On the other hand if the teaching focused on practicality there is a lot of commonality among the practical cases.

CamperBob2|10 years ago

I don't get it, either. Bessel functions certainly do have engineering applications.

pkrumins|10 years ago

One lesson academics should learn: pdf-naming-skills.pdf.

I've been collecting interesting scientific papers and publications since early 2000 (I've a collection of 10,000 or so) and I've not yet seen a single academic, not even a computer scientist, who understands how to name your documents right so that when I download them I could quickly find them. I've to rename every single pdf. It's infuriating.

Someone should teach academics an SEO course.

argonaut|10 years ago

Even worse: instead of having a discussion on college mathematics pedagogy and differential equations, you decided this was the most relevant thing to discuss, made even more totally irrelevant by the fact that 1) it's a futile suggestion, 2) the original author is dead, and 3) you haven't heard of full-text search or file renaming.

vlasev|10 years ago

I remember writing a little script in Python that used a pdf library to try to grab the title from inside the pdf and rename the file. It worked half of the time. I too have too many pdfs and it's annoying to have to rename every single one of them.

HarryHirsch|10 years ago

What's wrong with any of the pieces of reference management software out there? I use Endnote, others may have different preferences. Reference management and document retrieval is a problem that popped up together with the first library, and library classification has evolved alongside. I wouldn't say it's a solved problem, but it's a known problem with known attempts at solution. Attempting to roll your own is so reminiscent of OpenSSL, and they keep getting into trouble with their habit of inventing the wheel.

GFK_of_xmaspast|10 years ago

Everybody has a different system, and why should they standardize on yours?

thoth|10 years ago

I also got tired of organizing papers and websites (various articles I print to pdf), I just throw it all into Mendeley Desktop now and apply my own tags and so some minor modifications on author/title. I go with the "last name - year - title" naming and this works well enough for me.

I have an order of magnitude fewer docs that you; I can't imagine hand organizing that many!

pgbovine|10 years ago

you're welcome :)

http://pgbovine.net/publications.htm

this is something i try to do for all my papers, although of course people will still need to manually re-name to fit their own conventions. but at least it's better than Guo.pdf

gaur|10 years ago

The academic equivalent of the dreaded "resume.pdf".

xxs|10 years ago

When you download you have the option to pick a name (and not use the data provided by content-disposition)

swehner|10 years ago

The last point may well be the best, "TEACH CONCEPTS, NOT TRICKS"

sonabinu|10 years ago

I am taking a refresher class in calculus II from a community college and am pained by the tricks being taught and feel that the students are being deprived of learning core concepts. I am taking the trouble to read up and not too focused on the solution techniques that the professor emphasizes. However this doesn't do too much for grades. I wish he focused on concepts and application.

solipsism|10 years ago

Could someone give some examples of applications of numerical methods for solving differential equations that are relevant to a HN crowd? Also, where might I find some introductory material that teaches it well, according to the the suggestions in the OP?

cossatot|10 years ago

I'm not sure what the HN crowd finds useful as a group, but I personally use a couple of different techniques in my work (geophysics). One of them is finite difference and finite element methods. I have a book called 'Introduction to Numerical Geodynamic Modeling' by Teras Gerya that teaches finite difference modeling of plate tectonic phenomena, particularly of 2nd order differential equations (Poisson equation and variants) that are useful in modeling heat flow, diffusion and so forth through space and time. It's a great book but written for a specialized audience. I've used finite element models a lot, and the occasional boundary element method, but never written any.

I also use Green's functions, which are equations that describe the response of a medium to an impulse (think the propagation of sound waves from a source, though I do different stuff), by using convolution.

But I think jofer is the only other geophysicist on HN so we're probably not representative. Nonetheless, a lot of HNers have a physics, classical engineering or chemistry background and use similar tools... just not to find out what happened tens of millions of years ago.

GregBuchholz|10 years ago

http://longnow.org/essays/richard-feynman-connection-machine...

"By the end of that summer of 1983, Richard had completed his analysis of the behavior of the router, and much to our surprise and amusement, he presented his answer in the form of a set of partial differential equations. To a physicist this may seem natural, but to a computer designer, treating a set of boolean circuits as a continuous, differentiable system is a bit strange. Feynman's router equations were in terms of variables representing continuous quantities such as "the average number of 1 bits in a message address." I was much more accustomed to seeing analysis in terms of inductive proof and case analysis than taking the derivative of "the number of 1's" with respect to time. Our discrete analysis said we needed seven buffers per chip; Feynman's equations suggested that we only needed five. We decided to play it safe and ignore Feynman.

The decision to ignore Feynman's analysis was made in September, but by next spring we were up against a wall. The chips that we had designed were slightly too big to manufacture and the only way to solve the problem was to cut the number of buffers per chip back to five. Since Feynman's equations claimed we could do this safely, his unconventional methods of analysis started looking better and better to us. We decided to go ahead and make the chips with the smaller number of buffers.

Fortunately, he was right. When we put together the chips the machine worked. The first program run on the machine in April of 1985 was Conway's game of Life."

mathgenius|10 years ago

Does anyone else get the urge to spend the next few hours (or days) trying to work out (at least the basics of) everything Rota is talking about here?

Mathematicians have so much fun..

yiyus|10 years ago

Although I agree that word problems can be somewhat distracting in a DE course, I do not think they should be totally avoided. One of the main reasons it was difficult to grasp the concepts we studied in our first DE courses is that they were too abstract. When later we studied other subjects, such as electric circuits or fluid dynamics, everything started making much more sense.

In my opinion, the ideal way of learning would be to first have very basic (only conceptual) introductory courses of applied fields, where we find some basic equations that we do not know how to solve. And then, we study DE to learn the techniques to solve these problems, avoiding direct references but keeping in mind where we are going with all this.

musgravepeter|10 years ago

I thought the comment on exterior differential forms was interesting. I always wanted to delve into those and better understand what a dx all by itself was when separated from dy/dx by simple manipulation. Loved his comment "We justify this sudden introduction of differentials by saying that this is 'just another way or rewriting the differential equation,' or some equally atrocious lie."

I am now lusting after http://www.amazon.ca/Exterior-Analysis-Using-Applications-Di... but it's a bit pricy for a indulgence purchase!

seansmccullough|10 years ago

I got an A- in my differential equations class in college. I still wasn't sure what a differential equations was at the end. My pattern matching skills got a good workout, though.

vlasev|10 years ago

I had a similar experience with similar grade and outcome. I finally started understanding things when I took PDEs and so on.

antman|10 years ago

Can anybody comment on that? (page 8 paragraph 1 ):

Professional mathematicians have avoided facing up to density functions by a variety of escapes, such as Stieltjes integrals, measures, etc. But the fact is that the current notation for density functions in physics and engineering is provably superior, and we had better face up to it squarely

gjm11|10 years ago

In physics and engineering, you traditionally talk about things like "delta functions", write expressions involving them as if they are actual functions, etc. This is notationally very convenient but may be misleading because these things are not really functions.

So, what are they really? Well, the key things you can do with them are (1) "boring" linear algebra operations (you can add and subtract them, and multiply them by scalars) and (2) multiplying by some function and taking the integral. E.g., what delta(x) -- the Dirac delta function -- really is, is a thing such that when you compute integral f(x) delta(x) dx, you get f(0).

And so pure mathematicians have ways of dealing with them that make this property more explicit. The theory of distributions says: no, these aren't functions, they're linear functionals on the space of functions (e.g., the delta function is the thing that maps f to f(0)). So now you're no longer allowed to write them as integrals, which means that the very close analogy between "distributions" and ordinary functions is obscured, and e.g. if you need to do a change of variables you can no longer just do it the same way you already know about from doing integrals.

Alternatively, the theory of signed measures says: no, these aren't functions, they're kinda like probability distributions except that the total "weight" doesn't need to be 1 and the density can be negative in places. They are naturally applied not to points but to sets of points. (E.g., the delta function is the signed measure that gives a measure of 1 to any set including 0 and a measure of 0 to any other set.) Now you are allowed to write those integrals, but instead of writing integral f(x) delta(x) dx you need to write integral f(x) dH(x) where H(x) is the "Heaviside step function", so instead of delta(x) appearing there you have (morally) its integral, and again if you want to change variables or something you need to know a new set of rules for what you do to the measure.

Note: I have skated over some technicalities. They are quite important technicalities. Sorry about that.

The sloppy non-rigorous physicists' and engineers' notation, where you just pretend the damn thing is a function and manipulate it as you would any other function, is more convenient. (Right up to the point where you do some manipulation that is safe for actual functions but gives nonsense when applied to singular things like delta functions, and get the wrong answer.)

It's a little like calculus notation. The "Leibniz" notation we all use these days writes derivatives as dy/dx as if dx and dy were just small numbers (compare: we write integrals against distributions as integral f(x) delta(x) dx as if delta were just a function), which is kinda nonsensical if you take it too seriously but very convenient because it makes things like dz/dy dy/dx = dz/dx "obvious", which is not just coincidence but has something to do with the fact that derivatives really are kinda like quotients (in fact, they are limits of quotients). Similarly, using "function" notation for distributions lets you write things like "integral f(x) delta(x-3) dx" and see that "of course" that's f(3), and this convenience isn't mere coincidence but has something to do with the fact that distributions really are kinda like functions (and in fact every distribution "is" a limit of functions).

Newton had a different notation for derivatives. It didn't have a conceptual error baked into it (pretending that derivatives just are quotients), but it turns out that that's a useful conceptual error and that's part of why everyone uses Leibniz's notation these days.

al2o3cr|10 years ago

+1 to teaching concepts - but I don't agree with the author's opposition to word problems and learning to apply a "bag of tricks". I'd argue that "deciphering vaguely-phrased word problems and figuring out which of a selection of tricks to apply" is the MOST transferrable skill somebody can take away from a math class, because it's a major component of working in lots of other fields. For instance, the biggest difficulties I've observed in novice developers are in breaking apart a big challenge ("write a program that solves this Sudoku board") into digestible / implementable pieces and in understanding which piece of information they already know can get the result they want.

amoonki|10 years ago

I agree, but I think the "word problems" the author was referring to are much lower quality than the ones you're thinking of. I imagined some highly-contrived exercises where all of the relevant information is already pre-processed for you, removing any need for problem decomposition. For example, "if the angle between the ground and a tree's shadow is 45 degrees and a 50 ft tall telephone pole that's 10 ft away from the tree casts a shadow..." (substitute a similar differential equation problem). As you point out, half the fun is defining a problem and breaking it down, and these kinds of word problems don't give you a chance to do that.

danidiaz|10 years ago

Where can I find an expansion of the intuitive explanation given for integrating factors?

> It is of the utmost importance to explain the relation between the solutions of the differential equation and the solutions of the system. The solutions of the system are trajectories, they are parametric curves endowed with a velocity given by the vector field. The solutions of the corresponding differential equation are integral curves, and their graphs are the graphs of the trajectories deprived of velocity. Often, instead of solving the differential equation, it is more convenient to solve the corresponding autonomous system.

selimthegrim|10 years ago

Read Arnold, read Arnold, he is the master of us all.

bzbarsky|10 years ago

Here's an attempted expansion.

If you have an equation of the form dy/dx = f(x) and you want "solve" it, what you are typically looking for is to write y = g(x), right? In other words, the solution to this differential equation is some curve in the x-y plane. This applies more generally, e.g. to situations where you end up with an "implicit" solution like h(y) = g(x): you still get an equation relating x and y which can then be represented as some set of points in the x-y plane (the ones that satisfy that equation).

Now say f(x) happens to have the form a(x,y)/b(x,y). You can consider the system of two differential equations: dy/dt = a(x,y), dx/dt = b(x,y). Solving this system gives x and y as functions of t. Picking any particular value of t gives values of x and y, which gives you a point in the x-y plane. The first key point is that the set of points produced by this procedure as you plug in all possible values of t is exactly the set of points for which the h(y) = g(x) equation above holds. In other words, the solution to the two-equation system encapsulates all the information about the solution to the original equation.

The second key point is that the solution to the two-equation system has _more_ information than the solution to the original equation. In particular, it has the actual values of dx/dt and dy/dt for every given value of t, which don't correspond to anything in our original problem. Their _ratio_ does correspond to something in our original problem: the slope of the tangent line to the solution curve (dy/dx). But the exact values themselves are somewhat arbitrary, as long as their ratio is correct. Put another way, our original problem's solution is a curve in the x-y plane, while the solution of our two-equation system is a curve together with a description for how fast to move along it as t changes. That's the "velocity" bit in Rota's article.

OK, but if how fast we move along the curve doesn't really matter, maybe we can choose to move along it in a nice way that makes it particularly simple to figure out what the shape of the curve is. Our only constraint is that at any given point along the curve the ratio of dx/dt and dy/dt is fixed, because in our original problem we have a fixed dy/dx if we're given values of x and y. So if, at every point (x,y) we multiply dx/dt and dy/dt by the same number (which can depend on x and y) then we get a system of two equations that has different solutions for x and y as functions of t, but the graph of the resulting thing in the x-y plane still looks the same. That's the integrating factor bit; we just formalize it by saying that we multiply both dx/dt and dy/dt by the same function q(x,t), which is exactly what it means to multiply them both at every point by some number that might depend on that point.

The hard part, of course, is choosing a q(x,y) that makes things work out nicely and makes it easy to solve our two equations to get x(t) and y(t).

Here's a concrete example that might help:

Say dy/dx = x/y. We rewrite this in the form dy/dt = x, dx/dt = y. This isn't terribly convenient to solve, so we multiply by q(x,y) = 1/(2xy) to get a new system: dy/dt = 1/(2y), dx/dt = 1/(2x). At this point, maybe you just look at it and go, ah, y = sqrt(t + C1), x = sqrt(t + C2), or maybe you figure out some other way to get there. In any case, now you see that t + C1 = y^2, t + C2 = x^2, so x^2 - y^2 = C for some constant (C2-C1, but both are arbitrary, so this is just some single arbitrary constant). And that's your (implicit) solution for the original differential equation: a hyperbola, or more precisely a family of hyperbolas each of which satisfies the equation.

To illustrate the point about velocities, let's just consider C = 1, so x^2 - y^2 = 1. The point (sqrt(2), 1) lies on this curve. At this point, dy/dx = x/y = sqrt(2). On our original formulation of the parametric system, dy/dt = sqrt(2), dx/dt = 1 at this point. In our reformulation with the integrating factor, dy/dt = 1/2 and dx/dt = 1/(2*sqrt(2)). So the two formulations have us moving along the hyperbola at different speeds at this point as t changes, but they're moving along the same hyperbola.

Does that help at all?

RickHull|10 years ago

Wow. I managed to skip Diff-e-q in my C.S. curriculum, and I always wondered if was the worse for it. Perhaps I was prescient.

XaspR8d|10 years ago

I somehow managed to get a bachelor's in mathematics without ever taking a DE class (snuck through on a year when program requirements were being rewritten). Everyone I've ever told this to has been aghast, and yet when I ask what I missed out on, no one really has a response other than "I thought everyone had to take it".

gajomi|10 years ago

I would be very curious to see if Gian Carlo Rota had anything to say about Stephen Strogatz's view on this. Strogatz's text (which was written three years before this article, right before he left MIT), is much beloved by many scientists and engineers, but most mathematician's will have complaints about it.

ataspinar|10 years ago

Do you have a link?

rafinha|10 years ago

"FORGET ABOUT EXISTENCE AND UNIQUENESS OF SOLUTIONS" what? most important thing about differential equations.

kevinr|10 years ago

As an engineer, I promise you, I give not one single solitary conscious fuck about the uniqueness of solutions to differential equations. Mostly I just hit things with Fourier or Laplace transforms as appropriate until they stop moving.

evanb|10 years ago

For ODEs you can simply think of solving the autonomous system numerically. Since I have many numerical algorithms to solve such a system, solutions exist. Since (most of) those algorithms are totally deterministic and offer no choices anywhere along the way (except maybe for some initial conditions) the solutions are unique.

For PDE's it's much more interesting, as the author points out.

chris_wot|10 years ago

Very late to the conversation, but I emailed this to my uncle and he wrote back:

At ANSTO I worked with solution of simultaneous first order differential equations, as arose from Newtons law of cooling for a 4-body calorimeter. That was fun.

jonesb6|10 years ago

I 100% believe DE 1 and 2 are courses used to weed out computer science students who don't meet a certain criteria. Whether this is good or bad is highly debatable imo.

forkandwait|10 years ago

a = F/m

Want motivation? Just saying.

rorykoehler|10 years ago

It seems there are lots of posts complaining about how ODE's are taught here. I am planning to study them in the next months. I can only learn maths through applied mathematics and always need to know the why before I can get the how. Can anyone please point me in the right direction for online materials which will help me self-learn, taking into account my learning preferences?

raverbashing|10 years ago

I would say skip studying "pure ODEs" then examine what you need specifically depending on your area

madengr|10 years ago

I remember my undergrad signals and systems class. Instructor said if I use Laplace transforms on any of the problems, I would get no partial credit. I got an A+ for the course.

Also got A+ in DE, but I still don't think a grokked it.

dbpokorny|10 years ago

> As a matter of fact, the need for proving existence theorems was not felt until the end of the nineteenth century, and I refuse to believe that someone like Cauchy or Riemann did not think of them. More probably, they thought about the possibility of proving existence theorems, but they rejected it as inferior mathematics.

...

> Most often, some student will retort with the dreaded question: “So what?”

Insecure snob.