Additive combinatorics

Additive combinatorics is an area of combinatorics in mathematics. One major area of study in additive combinatorics are inverse problems: given the size of the sumset A + B is small, what can we say about the structures of $$A$$ and $$B$$? In the case of the integers, the classical Freiman's theorem provides a partial answer to this question in terms of multi-dimensional arithmetic progressions.

Another typical problem is to find a lower bound for $$|A + B|$$ in terms of $$|A|$$ and $$|B|$$. This can be viewed as an inverse problem with the given information that $$|A+B|$$ is sufficiently small and the structural conclusion is then of the form that either $$A$$ or $$B$$ is the empty set; however, in literature, such problems are sometimes considered to be direct problems as well. Examples of this type include the Erdős–Heilbronn Conjecture (for a restricted sumset) and the Cauchy–Davenport Theorem. The methods used for tackling such questions often come from many different fields of mathematics, including combinatorics, ergodic theory, analysis, graph theory, group theory, and linear algebraic and polynomial methods.

History of additive combinatorics
Although additive combinatorics is a fairly new branch of combinatorics (in fact the term additive combinatorics was coined by Terence Tao and Van H. Vu in their book in 2000's), an extremely old problem Cauchy–Davenport theorem is one of the most fundamental results in this field.

Cauchy–Davenport theorem
Suppose that A and B are finite subsets of the cyclic group $$\mathbb{Z}/p\mathbb{Z} $$ for a prime $$p$$, then the following inequality holds.
 * $$ |A+B| \ge \min(|A|+|B|-1,p) $$

Vosper's theorem
Now we have the inequality for the cardinality of the sum set $$A + B$$, it is natural to ask the inverse problem, namely under what conditions on $$A$$ and $$B$$ does the equality hold? Vosper's theorem answers this question. Suppose that $$|A|,|B| \ge 2$$ (that is, barring edge cases) and
 * $$ |A+B| \le |A|+|B|-1 \le p-2, $$

then $$A$$ and $$B$$ are arithmetic progressions with the same difference. This illustrates the structures that are often studied in additive combinatorics: the combinatorial structure of $$A+B$$ as compared to the algebraic structure of arithmetic progressions.

Plünnecke–Ruzsa inequality
A useful theorem in additive combinatorics is Plünnecke–Ruzsa inequality. This theorem gives an upper bound on the cardinality of $$ |nA-mA| $$ in terms of the doubling constant of $$A$$. For instance using Plünnecke–Ruzsa inequality, we are able to prove a version of Freiman's Theorem in finite fields.

Operations on sets
Let A and B be finite subsets of an abelian group, then the sum set is defined to be
 * $$ A+B = \{a+b : a \in A, b \in B\}. $$

For example, we can write $$\{1,2,3,4 \} + \{1, 2, 3\} = \{2,3,4,5,6,7 \}$$. Similarly we can define the difference set of A and B to be
 * $$ A-B = \{a-b : a \in A, b \in B\}. $$

Here we provide other useful notations.
 * $$ kA = \underbrace{A+A+\cdots+A}_{k\text{ terms }} = \{a_1+\cdots+a_k : a_1 \in A, \dots, a_k \in A\}. $$

Not to be confused with
 * $$ k \cdot A = \{ka : a \in A\} $$

Doubling constant
Let A be a subset of an abelian group. The doubling constant measures how big the sum set $$ |A+A|$$ is compared to its original size |A|. We define the doubling constant of A to be
 * $$ K = \dfrac{|A+A|}{|A|}.$$

Ruzsa distance
Let A and B be two subsets of an abelian group. We define the Ruzsa distance between these two sets to be the quantity
 * $$ d(A,B) = \log \dfrac{|A-B|}{\sqrt{|A||B|}}. $$

Ruzsa triangle inequality tells us that the Ruzsa distance obeys the triangle inequality:
 * $$d(B,C) \le d(A,B) + d(A,C).$$

However, since $$d(A,A)$$ cannot be zero, note that the Ruzsa distance is not actually a metric.