These are FAQs about the rules of math contests (e.g. MOP qualification, IMO criteria, and so on) that I often get asked. Please note: this is not an official source in any way; it represents my best knowledge and may become out of date. See the official AMC website for authoritative answers.
If you are looking for training advice, see Contest FAQs.
Starting in 2011 and as of 2019, the IMO selection procedure is as follows.
In the USA the main olympiad event is the USA(J)MO held in April each year. Roughly 250 students qualify for the USAMO and 250 students qualify for the USA(J)MO via scores on previous short-answer contests. Then things get interesting:
- In USAMO of year $N-1$, about 60 students in grades 11 and below and the IMO team of that year are invited to the three-week Mathematical Olympiad Summer Program, abbreviated MOP.
- At the end of MOP, an exam called the TSTST (yes, I know) is given. The acronym stands for “Team Selection Test Selection Test”. The top approximately 30 students are then selected for the year-round TST’s for the IMO of year $N$.
- Two TST’s are given in December and January, administered by schools. Each of the two exams is worth 21 points.
- Contestants participate in RMM Day 1, worth 21 points.
- Contestants participate in the APMO, which is weighted by 0.6; hence it is also worth 21 points (despite being a five-problem exam).
- Finally, the two days of USAMO of year $N$ serves as the final TST, worth 42 points.
- The sum of scores over these five tests (six days, maximum 126 points) determine that year’s IMO team.
In particular, winning the USAMO is neither necessary nor sufficient for qualifying for the American IMO team, since the IMO team is determined by several different exams.
The IMO selection process takes an entire year. (In particular, it is impossible to qualify for the IMO without first having attended MOP the previous year.) The IMO team is known shortly after the USAMO grading in May.
Here are the details for IMO 2014 (the year which I participated), A series of tests/camps take place and at each step the number of contestants is reduced. Interestingly, the Taiwan national olympiad is not involved in the selection process at all.
- The first test worth mentioning is the APMOC, an exam in the same format as the APMO.
- High scores on the APMOC are invited to another camp at which the APMO is given in March.
- The top 30 scores on the APMO are invited to the first TST camp in late March.
- The top 15 scores at the first TST camp are invited to the second TST camp two weeks later.
- The top 10 scores at the second TST camp are invited to the third TST camp two weeks later.
- The IMO team is decided by a weighted sum of the scores at the second and third TST camp plus an oral interview.
Each camp lasts roughly from Friday to Tuesday; the geography of Taiwan makes commuting convenient. Each TST camp features
- Three Team Selection Quizzes (formally “individual study”), which last 110 minutes with two olympiad problems
- A full mock IMO with six problems.
Thus, the IMO team is decided by $6 \cdot 2 = 12$ mock IMO problems and $2 \cdot 3 \cdot 2 = 12$ quiz problems. The weighting is roughly so that a mock IMO problem is worth three quiz problems. The selection process takes a few months. The team is known shortly after the third camp, a couple of days into May.
For more details, see my report on Taiwan IMO.
Don’t take my word for it: try reading the USAMO 2003 rubric.
In general, for each problem the solution is graded according to the rubric:
- 7: Problem solved.
- 6: Tiny slip (and contestant could repair)
- 5: Small gap or mistake, but non-central
- 2: Lots of genuine progress
- 1: Significant non-trivial progress
- 0: “Busy work”, special cases, lots of writing
The most important difference is that USAMO grading is designed to reward complete solutions. Basically, we split the solution into either top-down or bottom-up: meaning the score is “7-” (has all the main ideas) or “0+” (missing at least one main idea). The scores of 3 and 4 are absent for this reason (typically used only for two-part problems or strange cases). This means your total score is essentially equal to 7 times the number of complete solutions plus or minus some pocket change.
(Pocket change tends to also be pretty small, though it depends a lot on who writes the rubric for each problem. In particular, we don’t give many sympathy points, and demand real progress towards a solution to get partial credit, not just vague ideas or lots of writing.)
One thing that is perhaps also surprising: rubrics often do NOT use standard addition. Meaning, we do not break the problem into bite-sized chunks and award the sum of the chunks. Here are a few examples of what I mean.
- Consider a problem that is two parts. In grader’s internal jargon, an example rubric for that problem might colloquially be summarized as “2+2=7”, meaning doing either half of the problem is worth 2 points, getting both halves is worth 7. The exact parameters depend on the problem and the captain. On both extremes, I have seen 3+5=7 and 1+1=7 rubrics as well.
- On a rubric, there might be three or four different starting observations that are worth 1 or 2 points. Often, a 0+ solution gets the maximum, rather than the sum, of these little partial bits: “1+1+1+2=2”, say.
- If you forget to do some easy step (e.g. the base case of an induction), and the rubric penalizes a top-down solution by 1 point for this, it is not necessarily the case that the rubric would award 1 point for doing that step for a 0+ solution.
There is no style: a totally correct and complete solution receives 7 points. However, it is still in your best interest to write a solution clearly to minimize the chance the grader doesn’t understand you. Graders don’t give points for things they can’t comprehend ;)
If you want to see more examples of problem rubrics, you should check out the reports of the USEMO. These rubrics are designed by veteran graders and released to the public, so the standards for these rubrics is likely to be similar to what you might encounter at USAMO.
No. It is a long-standing policy of AMC that decisions of the USAMO graders are final and may not be appealed. I am also not able to look up your exact score distribution for you.
I won’t claim that we graders never make mistakes, but all papers are read independently by at least two graders (and sometimes more if the solution is especially different or unusual); these graders must agree before a score is assigned. Then the top papers on each contest (anyone within one problem or so of an award or MOP invitation) are taken and re-graded again, independently, by a third grader. Nearly all the graders are either past USAMO winners or IMO medalists or high-scorers of some national olympiad.
I will say that USAMO grading tends to be stricter than IMO grading, or most other national olympiads for that matter.
All that being said, this is a good incentive to learn to write your solutions neatly. You don’t want to be the one with a correct solution that the graders can’t understand!
There are no hard criteria set in stone, but in general, any well-known result which does not trivialize the problem is okay. If you are still unsure, it is strictly better to include the proof. See English handout for more discussion.
Enough to convince the grader that you understand the solution to the problem.
Whenever you say something like “it’s easy to see X”, the grader has to ask themselves whether the student actually understands why X is true, or doesn’t know and is just “bluffing”. In particular, this is a math issue, not a style issue.
As a loose approximation, the official solutions file for USAMO is about as terse as you can be. See English handout for more discussion.
Before I say anything I want to say that the criteria for MOP invitations are not especially well-defined. Each year, the exact number and choice of students is determined based on the exact scores for that year.
That being said, as of 2016 the criteria for MOP is roughly as follows:
- IMO team members (see CR-1)
- The next approximately 12 non-graduating USAMO students
- The next approximately 12 USAMO students in 9th and 10th grades
- The top approximately 12 students on USAJMO
- Some varying number of non-graduating female contestants from either USAMO or USAJMO (these students represent USA at the European Girls’ Math Olympiad). The exact cutoffs for each contest are determined based on the scores for that year.
Young students in 8th grades and below receive invitations if and only if the moon is full and the wind is blowing south-south-east. All selection is done by ID number, without student names.
As a consequence of the procedure spelled out in CR-1, it is not possible for a student to attend MOP for the first time as a graduating 12th grader. This is because the IMO selection process begins at MOP of the previous year.
At MOP, there are often four tracks of classes and tests. The tracks are usually called Black, Blue, Green, Red as a convenient shorthand, a tradition started around 2002. The rosters for these classes correspond approximately with the first four bullets in CR-7 in that order, but only loosely.
Nevertheless, students will often colloquially refer to the cutoffs by color as well, in a slight abuse of terminology. (For example, “blue cutoff” would mean the minimum score for an 11th grader to attend MOP through USAMO, even though this student could end up in any class group depending on the whims of the assistant academic director.)
Some students (not staff) even refer to the 5th cutoff in CR-7 as the “pink” cutoff, despite there being no analogous class or testing group. This imaginary color was coined by students around 2007, not necessarily out of good will, and stuck around. Students have told me they find “pink” offensive, so I avoid it. But the name is so entrenched I think it will never die, hence I decided to explain its history here.
Finally, I better explain that color groups are not that important. Once the USA students arrive at MOP, they are all treated equally in team selection: no matter their color group, past USA(J)MO score, gender, or what-have-you. The team selection is not based on a caste system. As a matter of principle, anyone attending MOP of year N and still in high school could qualify for IMO of year N+1.
As of 2016, ties are generally only broken for USAMO winners, since exactly 12 winners are invited to the state dinner. This did not occur in 2020 or 2021 (due to COVID) and I don’t know if it will continue.
- General IMO regulations, section 7
- 2002 IMO report by Tom Verhoeff
- IMO 2019/3 blog post by Michael Ng
- Quora answer by Carlos Shine
- Reddit post on IMO 2019/5
- IMO 2020 instructions (unusual virtual procedure)
Basically, the outline of the idea is: before the exam, a marking scheme (rubric) is set for each problem, to cover the typical cases of what progress will be worth what points. Then, the leaders of each country get to see the solutions of their country’s students, while there is a number of coordinators from the IMO host country for each problem. Both the coordinators and the leaders read through these scripts in advance, and then meet at an appointment where they discuss the scores. There are a lot of scripts that are not in English, in which case the leader is expected to do any necessary translation.
The idea is that the leaders try to make the best case they can from their student’s work (while of course still being honest), while the coordinators for the problem try to maintain uniform fairness. We sometimes say jokingly that it’s an “organized fight”, with the analogy that leaders are defense attorneys and coordinators are prosecutors, but in practice it’s a lot friendlier than this. I remember being a coordinator at the virtual IMO 2020 (problem 5), and there were quite a few cases where I offered a higher score than the leader asked for, because I managed to decipher a part of the student’s paper that matched a rubric item. (It’s quite a different experience having your country’s six papers for each of six problems versus 100 papers of the same problem — you start seeing the common patterns and similarities.) On the other hand, it’s not uncommon for leaders to spend a lot of effort trying to see how partial solutions of their students could be completed, particularly if the student’s approach is novel or uncommon.
At every IMO, the jury is presented a list of about 30 problems, and select six of them to comprise the IMO. (The shortlist is in turn selected from all proposed problems, of which modern IMO’s have a few hundred. The selection is nowadays much more competitive than back in early 2000s.) The list for each year is called the IMO Shortlist.
The shortlist is typically divided into four categories Algebra, Combinatorics, Geometry, Number Theory (with about 6-8 problems in each category) and spans IMO-easy to IMO-hard. The problems are numbered A1, A2, … in loosely ascending difficulty; similarly the combinatorics problems are labeled C1, C2, … and so on. (This means you can refer to a problem being “C8 level” in a way that is meaningful from year to year.) All the posted shortlists for the past several years are indexed on AoPS, and recent shortlists are available in their entirety on the official IMO website.
An important property of the IMO shortlist is that problems which are shortlisted are confidential for one year if they do not appear on IMO. For example, the IMO 2020 shortlist is confidential until after IMO 2021 concludes. The reason for this is that the IMO Shortlist is a valuable source of good problems, so many countries will use shortlisted problems either for training purposes, or in their team selection tests. Thus, it is critical that the security of the shortlist is not compromised. (So, if you see a shortlist problem posted on AoPS prematurely, you should immediately report it.)
Shortlisted problems from previous years of the IMO are extensively studied by top students, to the point where IMO team members from countries like the USA will often know from memory which problem “2011 G4” refers to, etc.