Processing math: 100%
Skip to main content

Section 7.2 From Linear to General

In this section, we will take two ideas we already used with linear congruences, and see how they can be modified to apply in any polynomial situation.

Subsection 7.2.1 Combining solutions

One of the most important things we can do is study congruences with prime (power) modulus, because we can combine their solutions to get solutions for any congruences when we combine the Chinese Remainder Theorem and Fundamental Theorem of Arithmetic (recall Proposition 6.5.1). Even more interestingly, we can combine the numbers of solutions.

Informally, if you want to get the total number of solutions of a polynomial congruence, just write the modulus as a product of prime powers n=∏ki=1peii, find out how many solutions the congruence has with each prime power modulus, then multiply those numbers for the total number of solutions.

Example 7.2.1.

For instance, if f(x)≑0 has 2 solutions modulo 3, 1 solution modulo 5, and 3 solutions modulo 7, it would have 2β‹…1β‹…3=6 solutions modulo 105=3β‹…5β‹…7.

We will state this for the general case of a coprime factorization of n, though again the prime power factorization is usually the most useful.

For all \(i\text{,}\) among the \(N_i\) solutions to the \(i\)th congruence choose a solution \(a_i\text{,}\) so that

\begin{equation*} f(a_i)\equiv 0\text{ (mod }n_i)\text{.} \end{equation*}

Since the moduli \(n_i\) for these congruences are coprime, we can use the Chinese Remainder Theorem to obtain one number \(a\) such that \(a\equiv a_i\) (mod \(n_i\)) for all \(i\text{.}\)

Since polynomials are exclusively made up of addition and multiplication, and addition and multiplication are well-defined, we also have \(f(a)\equiv f(a_i)\equiv 0\) (mod \(n_i\)), so as promised we have a solution

\begin{equation*} f(a)\equiv 0\text{ (mod }\prod_{i=1}^k n_i\text{)}\text{.} \end{equation*}

Each such set of \(a_i\) will yield a solution, and if \(\{a_i\}_{i=1}^k\neq \{b_i\}_{i=1}^k\) then if \(a_j\not\equiv b_j\) (mod \(n_j\)) they certainly are not equivalent modulo \(\prod_{i=1}^k n_i\) either.

Now multiply how many solutions there are for each \(n_i\) to get the total number of combinations of solutions. If there are \(N_i\) solutions modulo \(n_i\text{,}\) we would get \(\prod_{i=1}^k N_i\text{.}\) There aren't any additional answers, because any answer to the β€˜big’ congruence automatically also satisfies the β€˜little’ ones; if \(\prod_{i=1}^k n_i\mid f(a)\text{,}\) then certainly \(n_i\mid f(a)\) as well.

Subsection 7.2.2 Prime power congruences

We have already discussed prime power congruences in Subsection 6.5.2. Recall that in Examples 6.5.3 and 6.5.4 we took the (obvious) solution of 4x≑1 (mod 7) (namely, x=[2]), and got solutions (mod 49) and even (mod 343) from it relatively easily.

But that is essentially the same as asking for solutions to 4xβˆ’1≑0, a linear congruence. Let's see if we can generalize this method for more general polynomial congruences.

The key was taking the already known fact 7∣1βˆ’4β‹…2 and then cancelling out 7 from the entire congruence to get that

1βˆ’4β‹…27≑4k (mod 7).

We were able to solve the resulting congruence βˆ’1≑4k (mod 7), which had solution k≑5 (mod 7). Finally, we plugged that back in to get a solution to 4x≑1 (mod 72), which was

2+7k=2+7β‹…5=37 (mod 72)

as the solution.

Can we use this approach to get solutions to more advanced congruences as well, like the simple quadratics we've started exploring in this chapter? The answer is yes, with a minor caveat. The preceding discussion was just a basic form of the following.

If \(p\) and \(f'(x_{e-1})\) are relatively prime, then by Proposition 5.1.1 any linear congruence of the form \(f'(x_{e-1})k\equiv b \text{ (mod }p)\) with coefficient \(a=f'(x_{e-1})\text{,}\) unknown \(k\text{,}\) and known \(b\) can be solved (uniquely modulo \(p\text{,}\) given the \(\gcd\) condition). Since \(x_{e-1}\) is a known zero of \(f(x)\) for modulus \(p^{e-1}\text{,}\) we know that as an integer (not modulo anything) \(p^{e-1}\mid f(x_{e-1})\text{.}\)

This means that \(-\frac{f(x_{e-1})}{p^{e-1}}\) exists in \(\mathbb{Z}\text{,}\) so if we set \(b=-\frac{f(x_{e-1})}{p^{e-1}}\) there will indeed be a solution \(k\) to the congruence \(\frac{f(x_{e-1})}{p^{e-1}}+k\cdot f'(x_{e-1})\equiv 0\text{ (mod }p\text{)}\text{.}\) Then the only question becomes why \(x_{e}=x_{e-1}+kp^{e-1}\) is actually a solution to \(f(x)\equiv 0\text{ (mod }p^{e})\text{.}\)

To see this, think of \(f\) as a polynomial with terms of the form \(c_i x^i\text{,}\) e.g. \(f(x)=\sum_{i=0}^n c_i x^i\text{.}\) Then \(f(x_{e-1}+kp^{e-1})\) can be expanded out term-by-term in the following form:

\begin{equation*} f(x_e)=f(x_{e-1}+kp^{e-1})=\sum_{i=0}^n c_i (x_{e-1}+kp^{e-1})^i\text{.} \end{equation*}

Let's break this down on a term-by-term basis in the sum. Each term will look like

\begin{equation*} c_i(x_{e-1}+kp^{e-1})^i = c_i x_{e-1}^i+c_i (x_{e-1}^{i-1}\cdot kp^{e-1})\cdot i+\text{ terms with at least }p^{(e-1)2} \end{equation*}

Since \(e\geq 2\) in this context, the extra terms (from Taylor or binomial series) 2 One way or another one of these series will have to enter in, unfortunately; [C.2.1, Section 4.3] has more of a binomial theorem-esque treatment, while [C.2.13, Theorem 4.7] and [C.5.1, Theorem 6.2] more explicitly invoke Taylor series. involving \(p^{(e-1)/2}\) will be divisible by at least \(p^e\) and hence be irrelevant in that modulus, so each term in the sum will be equivalent to

\begin{equation*} c_i x_{e-1}^i+c_i \cdot i x_{e-1}^{i-1}\cdot kp^{e-1}\text{ (mod }p^e)\text{.} \end{equation*}

We're nearly done with the individual terms. Recall Proposition 5.2.6 where we are allowed to cancel a nonzero divisor from β€œall three sides” of a congruence. That motivates dividing each term and the modulus by \(p^{e-1}\) to get

\begin{equation*} \frac{c_i x_{e-1}^i}{p^{e-1}}+c_i \cdot i x_{e-1}^{i-1}\cdot k\text{ (mod }p)\text{.} \end{equation*}

Now add up the terms of the sum for all \(i\) to find out something about \(f(x_e)\text{.}\) Summing up the \(x_{e-1}^i\) will give us \(f(x_{e-1})\text{,}\) and summing up \(i x_{e-1}^{i-1}\) is adding terms that look like the derivative of polynomials, so modulo \(p\) we have

\begin{equation*} \frac{f(x_{e-1})}{p^{e-1}}+f'(x_{e-1})\cdot k\text{,} \end{equation*}

which is divisible by \(p\) by hypothesis.

Now we have that

\begin{equation*} f(x_e)/p^{e-1}=f(\text{the sum})/p^{e-1}\equiv 0\text{ (mod }p)\text{,} \end{equation*}

so we multiply everything by \(p^{e-1}\) (this time actually using Proposition 5.2.6) and get \(f(x_{e})\equiv 0\text{ (mod }p^e)\) as desired.

Let's use Hensel's Lemma to take solutions to x2+1≑0 (mod 5) and turn them into solutions modulo 25 and 125.

Example 7.2.4.

First we solve x2+1≑0 (mod 25). By inspection, the solutions modulo 5 are [2],[3] (or [Β±2]). So solutions modulo 25 will look like 3+kβ‹…5 or 2+kβ‹…5. Further, fβ€²(x)=2x, so for either solution modulo 5 the technical derivative condition is met.

Let x1=3. Then the condition for k is

f(x1)5+kβ‹…(2x1)≑0 (mod 5)

which simplifies to 2+6k≑0, which solves to kβ‰‘βˆ’2≑3. Then our solution to the congruence modulo 25 would be

x2=x1+3β‹…5≑18 (mod 25)

And indeed 182+1=325 is divisible by twenty-five.

Now try the same procedure with x1=2 to get the solution x2≑7 in Exercise 7.7.3. (If you get stuck, see Example 16.1.3.)

Example 7.2.5.

We can try the same process with e=3. Taking (from the previous example) x2≑7 yields, as a condition for k,

72+125+14k≑0 (mod 5)

This gives k=2, and indeed

x3=x2+2β‹…52=57

yields

572+1=3250≑0 (mod 125).

It's good practice to try the same process with x1=18 instead in Exercise 7.7.3.

This is a very powerful technique. What is most interesting is that this is even interpretable as Newton's method in calculus. How? Note that the result above can be rearranged as

xe=xeβˆ’1βˆ’f(xeβˆ’1)fβ€²(xeβˆ’1)

since peβˆ’1∣f(xeβˆ’1) and the technical condition is tantamount to saying fβ€²(xeβˆ’1) has an inverse. (Unlike in the Newton case, it is also possible for there to be solutions here if gcd(p,fβ€²(xeβˆ’1))β‰ 1, but only if f(xeβˆ’1)peβˆ’1 itself is also divisible by p. We omit details of this case, which then yields additional solutions for each successive e.)

If you didn't notice this, don't feel bad! When we had the linear congruence f(x)=4xβˆ’1 in Examples 6.5.3 and 6.5.4, the derivative was just fβ€²(x)=4 and it was not at all obvious that anything more than a trick was involved. Still, it's another fascinating place where ideas from calculus can invade the world of number theory.