Sure, but I try really hard not to do that (and tell my students so). Trick questions belong in homework, not on timed tests.
Hmm, I dunno. On the one hand I understand that timed tests produce stress and so forth, on the other hand, does it mean all questions in tests have to be trivial? What is the purpose of the test anyway? I don’t have a strong opinion on this (also I feel liked my perspective is “privileged” since I finished most of my math tests well ahead of time). I can say that, when wearing the hat of an employer, I want candidates that can solve non-trivial questions and I have to test for it myself since a university degree is not much of an indication (then again, my recruiting experience is for pretty elitist groups).
My best childhood nemories of math are things that they either don’t teach in school or teach very late.
I remember a geometry class in elementary school in which the teacher was explaining dimension. “There are 1 dimensional shapes, 2 dimensional shapes and 3 dimensional shapes.” (some examples) I chimed in: “and 4 dimensional shapes!” The teacher laughed: “there is no such thing as 4 dimensional shapes.” Of course I got annoyed and argued but it didn’t help.
School teaches that math is all about following rules. But the best parts of math are about *breaking* the rules!
A: You cannot take the square root of -1.
B: Yes I can! I get i!
A: What on earth is i?!
B: i is this number I just made up!
Hyperbolic geometry is basically “screw the parallel postulate. Let’s decide it’s false and see what happens. Wow, equilateral triangle with three angles of 10 degrees, awesome!”
Differential manifolds came from “Let’s study surfaces in Euclidean space. On second throught, screw Euclidean space, let’s throw it out and leave *only* the surface.”
Math is not about following rules. Math is about making up your own rules. It is about playing games in your head and seeing what happens. Trying to discover something beautiful, strange or unexpected. The only strict rule in math is consistency: you have to keep track of the rules you’re playing with.
The single most important thing about math is something they don’t teach you: math is art. It is constrained art, but art nevertheless.
probably they’ve realised that answering each of those questions requires years of study and more mathematics than they care to know.
If you spend 12 years showing kids that learning is boring and that mathematics is drilling and memorizing formulae you’re doing the terrible thing he’s talking about.
Even though mathematics isn’t drilling and memorizing formulae, sometimes it really is just easier to let someone else learn the math to do the science. Comparative advantages exist, and so does Google for the simpler questions.
Lockhart doesn’t have a monopoly on math. Yeah, maybe when he does it it’s beautiful art. But when most people do it they’re just trying to calculate the damn tip so they can go home and sleep.
It would be a horrible mistake to try to teach everyone how to do real math. Not everyone is a Ravenclaw! Most people don’t and won’t care! It’s not at all useful in the real world! For the exactsame reasons that we shouldn’t force kids to learn art if they don’t want to, we shouldn’t force them to learn pure math if they don’t want to.
Also the kind of math you need to do science is in large part just memorizing formulas. You really think the average statistician knows measure theory? Or that the average biologist knows statistics?
More generally I am opposed to forcing people to learn the things that give you intellectual fuzzies. We shouldn’t force kids to listen to Mozart, look at Picasso, learn about special relatively, or read Euclid. And I think most of the argument for “let’s teach kids how math really works” is just “I really like this mostly useless intellectual exercise and therefore everyone should know it”.
Also also we already tried doing the thing Lockhart is talking about (sort of). It was called New Math and it failed miserably. Turns out most teachers don’t know pure math, and can’t learn it.
I agree we shouldn’t force kids to learn real math. But:
1. We should expose them to real math and give the chance to learn it if they want to.
2. We shouldn’t force kids to learn rote-memorization-pseudo-math also.
Deligne is the standard recommendation, and also really good as far as I can tell and have heard. I never got around to doing a principled study of it (hurts me to admit it, but it’s a little too mathy for my tastes), and probably “all” you really need (it’s quite long).
cc @nostalgebraist, @mark-gently and I think @alexyar is technically in a mathematical physics group, but I’m pretty sure she hates physics.
I second this recommendation (note that there are two volumes).
Somewhat off topic (but reminded to me by that book), I once spent quite a bit of time translating the theory of harmonic superspaces to mathematical language (i.e. describing them as Grassmanian-like objects in supertwistor space instead of the mess of indexes used by physicists) and the result was quite beautiful, but unfortunately I haven’t published it anywhere.
Sure, but I try really hard not to do that (and tell my students so). Trick questions belong in homework, not on timed tests.
Hmm, I dunno. On the one hand I understand that timed tests produce stress and so forth, on the other hand, does it mean all questions in tests have to be trivial? What is the purpose of the test anyway? I don’t have a strong opinion on this (also I feel liked my perspective is “privileged” since I finished most of my math tests well ahead of time). I can say that, when wearing the hat of an employer, I want candidates that can solve non-trivial questions and I have to test for it myself since a university degree is not much of an indication (then again, my recruiting experience is for pretty elitist groups).
There’s a difference between a difficult question and a trick question.
One big difference is that a trick question has more random effects. On a difficult question, the good students who have mastered the material will do well, and the students who have not mastered the material will do poorly.
On a trick question, the students who get lucky and/or try the right thing first will do well, and the students who get tripped up or tricked by the wording, or who don’t guess the right trick the first time, will do poorly. (This is one reason “timed test” matters; it takes the luck of “did you try the right thing first?” out of the running).
My tests are certainly difficult enough to discriminate. (I typically have a dispersal of scores pretty evenly over the range of scores I consider “acceptable”, with a few hanging off the bottom). Which means the tests are “hard enough.”
We could have a different argument about whether my mapping between scores and letter grades is “too generous” or “too harsh”, but that’s entirely a product of the curve I set and has little to do with how hard the tests are. Which is “hard enough that they generate a clear signal.”
In general, that sounds fair, although if the score is comprised from many questions then the random factors average out. In particular, this specific question doesn’t sound like it should take a large fraction from what I assume is a multiple hour test. Assuming that it is a only small fraction, I’m not sure that the requirement to verify that L’Hospital’s rule is applicable before using it is sufficient to make a “trick” question the way you defined it.
Of course, I never taught a course in anything, so my opinions are quite dilettante.
Is producing some interesting reactions in my mind, since I don’t understand all the applications so far, so lots of elegant facts about them are not obvious at all and just look miraculous.
So when I come across tidbits like “the sum of the eigenvalues is equal to the trace of the matrix” and “the product of the eigenvalues is equal to the determinant of the matrix”
I’m just like
To see that the trace is the sum of the e-values and the det is product of the e-values, it’s enough to understand why
1. Trace and det are invariant under changes of basis.
2. All matrices have a basis in which they have the Jordan normal form. In fact, most matrices (i.e. all but a set of zero measure) have a basis in which they are diagonal (all of this is true over algebraically closed fields, e.g. complex numbers).
The above two facts explain the “miracles” since in Jordan normal form checking these facts is trivial. Now, let’s see how to understand 1 & 2.
1. Algebraically, these are trivial. Namely, det(PAP^{-1}) = det P det A det P^{-1} = det P det A (det P)^{-1} = det A. Also tr(PAP^{-1}) = tr(AP^{-1}P) = tr A. All you need is the formula for changing bases (which follows more or less straight from the definitions) and basic properties of det and tr (although it takes some juggling to show that det is multiplicative). Geometrically, you can see it by observing that applying A to any n-dimensional shape multiplies its volume by |det A|, whereas the sign of det A says whether A preserves orientation (for real matrices; you can in principle derive invariance in the complex case from invariance in the real case by analytic continuation; alternatively, over any field det A is just the action of A on the 1-dimensional space of alternating n-forms). tr A is a bit more tricky but you can see it’s actually the derivative of det: d/dt (det A(t)) = (det A) (tr dA/dt).
2. Proving Jordan normal form is a bit technical so let me just do some hand waving to explain why most matrices are diagonalizable. Introducing an indeterminate x, we can compute the polynomial p(x) = det (A - xI) (characteristic polynomial of A). For A of size nxn, this is a polynomial of degree n. The e-values are just the roots of this polynomial since det(A - lambda I) = 0 iff A - lambda I is degenerate matrix iff it has a kernel i.e. exists non-zero v s.t. (A - lambda I)v = 0 but this is the definition of an e-value / e-vector. Now, most polynomials of degree n have n different roots. This means you have n e-vectors with different e-values and it’s not hard to see they have to be linearly independent*, so they form a basis. In this basis, A becomes diagonal.
* Assume they are dependent. Then, Sum_i c_i v_i = 0. Applying A, Sum_i c_i lambda_i v_i = 0. Multiplying the first equation by lambda_1 and subtracting, we conclude that a subset of size n-1 is linearly dependent. Continuing by induction we arrive at a contradiction.
This is so cool and weird! And straightforward enough that I, a set theory layperson, was able to make a lot of sense of this result and its proof.
An important thing to note is that “[y]ou have to run the program in a different universe in order that it will compute your desired function“ - in particular, you have to compute it in within a nonstandard model of arithmetic. What that might mean, insofar as computation is ultimately a physical process, is not at all clear.
It means that your Turing machine fails to halt on every input but it’s impossible to prove it in PA. These pathologies disappear if you set a time bound. But, yeah, it’s a cool theorem.
See also my alternative proof, which Hamkins added to the original post.
Is producing some interesting reactions in my mind, since I don’t understand all the applications so far, so lots of elegant facts about them are not obvious at all and just look miraculous.
So when I come across tidbits like “the sum of the eigenvalues is equal to the trace of the matrix” and “the product of the eigenvalues is equal to the determinant of the matrix”
I’m just like
You eventually start thinking about this the other way. The determinant is the product of the eigenvalues. It happens that, given a matrix, there’s this really annoying formula for calculating the determinant without figuring out what the eigenvalues are explicitly.
(There are a number of advantages to this approach. The big one is probably that it still makes sense if you’re thinking about “an operator” and not actually thinking about a matrix.)
I don’t think this is the ideal way to think about the determinant, in terms of “Platonic mathematical beauty.” The ideal way to think about the determinant, is via alternating n-forms.
Given a vector space V over a field K, a multilinear form or rank r is a function \alpha: V^r -> K which is linear in every argument when fixing all other arguments. A form is call alternating when substituting two equal vectors to any positions yields zero:
An alternating form is always antisymmetric, i.e. switching any two arguments flips the sign of the result. For K of characteristic other than 2, any antisymmetric form is alternating. For K of characteristic 2 (i.e. 1+1=0 in K), antisymmetric forms and symmetric forms are the same thing while alternating forms are a proper subspace of either (for V of non-zero dimension).
Alternating forms of any given rank obviously form a vector space of their own. For V of finite dimension n, the space of alternating forms of rank n is always 1-dimensional.
Now, consider a linear operator A: V->V. This operator induces a linear operator B on multilinear forms defined by
(B\alpha)(v_1, v_2 … v_r) = \alpha(A v_1, A v_2 … A v_r)
On the space of alternating forms of rank n, the operator B has to be multiplication by a scalar since this space is 1-dimensional. This scalar is the determinant of A.
If we ever build a hypercomputer, algorithmically incompressible strings will suddenly become very interesting! :)
I think that a hypercomputer is not only physically impossible but logically incoherent. What does it mean for an object in the physical world to be a hypercomputer? We don’t have “direct” access to the properties of physical objects: our only access if through testable models. These models have to be computable since our mind is computable*. In other words, if someone handed you a black box claiming it is a halting oracle, there would be no way for you to verify it is indeed a halting oracle. Moreover, there would be no way for you to verify the box is anything incomputable.
* That is, any precisely defined function that you can evaluate using your mind is computable, even if you could think for infinite time and had a perfectly reliable infinite external memory module. Some might argue that since you actually have neither infinite time nor infinite space for your mental computations, we should embrace the more radical position of ultrafinitism. I think this is debatable.
I decided that this tumblr blog doesn’t serve the functions I hope it would, and that it is a net negative influence on my productivity. Some of the posts will remain, but I will no longer produce new ones, and within a few days from now I will stop reading asks / messages / replies / reblogs.
That said, I will remember fondly some of the conversations I had here. Feel absolutely free to contact me in the future, at any time and for any reason: by e-mail (rot13 of gbc.fdhnex@tznvy.pbz) or via Facebook.