You are currently browsing the tag archive for the ‘factorial function’ tag.
I’ve just uploaded to the arXiv my paper “Products of consecutive integers with unusual anatomy“. This paper answers some questions of Erdős and Graham which were initially motivated by the study of the Diophantine factorial equation
The equation (1) ties into the general question of what the anatomy (prime factorization) of the product looks like. This is a venerable topic, with the first major result being the Sylvester-Schur theorem from 1892 that the largest prime factor of
is greater than
. Another notable result is the Erdős-Selfridge theorem that the product
is never a perfect power for
.
Erdős and Graham were able to show that solutions to (1) were somewhat rare, in that the set of possible values of had density zero. For them, the hardest case to treat was when the interval
was what they called bad, in the sense that
was divisible by the square of its largest prime factor. They were able, with some effort, to show that the union of all bad intervals also had density zero, which was a key ingredient in to prove the previous result about solutions to (1). They isolated a subcase of the bad intervals, which they called the very bad intervals, in which the product
was a powerful number (divisible by the square of every prime factor).
A later paper of Luca, Saradha, and Shorey made the bounds more quantitative, showing that both the set of values of , as well as the union of bad intervals, had density
for some absolute constant
. In the other direction, just by considering the case
, one can show that the number of possible values of
up to
is
, where
is the constant
It was conjectured by Erdős and Graham that all of these lower bounds are in fact sharp (up to multiplicative factors); this is Erdos Problem 380 (and a portion of Erdos Problem 374). The main result of this paper is to confirm this conjecture in two cases and come close in the third:
Theorem 1
- The number of numbers up to
that lie in a bad interval of length
is
of the number of bad points up to
.
- The number of numbers up to
that lie in a very bad interval of length
is
.
- The number of numbers up to
of the form
for a solution to (1) is
.
Not surprisingly, the methods of proof involve many standard tools in analytic number theory, such as the prime number theorem (and its variants in short intervals), zero density estimates, Vinogradov’s bounds on exponential sums, asymptotics for smooth numbers, the large sieve, the fundamental lemma of sieve theory, and the Burgess bound for character sums. There was one point where I needed a small amount of algebraic number theory (the classification of solutions to a generalized Pell equation), which was the one place where I turned to AI for assistance (though I ended up rewriting the AI argument myself). One amusing point is that I specifically needed the recent zero density theorem of Guth and Maynard (as converted to a bound on exceptions to the prime number theorem in short intervals by Gafni and myself); previous zero density theorems were barely not strong enough to close the arguments.
A few more details on the methods of proof. It turns out that very bad intervals, or intervals solving (1), are both rather short, in that the bound holds. The reason for this is that the primes
that are larger than
(in the very bad case) or
for a large constant
(in the (1) case) cannot actually divide any of the
unless they divide it at least twice. This creates a constraint on the fractional parts of
and
that turns out to be inconsistent with the equidistribution results on those fractional parts coming from Vinogradov’s bounds on exponential sums unless
is small. In the very bad case, this forces a linear relation between two powerful numbers; expressing powerful numbers as the product of a square and a cube, matters then boil down to counting solutions to an equation such as
The situation with bad intervals is more delicate, because there is no obvious way to make small in all cases. However, by the large sieve (as well as the Guth–Maynard theorem), one can show that the contribution of large
is negligible, and from bounds on smooth numbers one can show that the interval
contains a number with a particularly specific anatomy, of the form
where
are all primes of roughly the same size, and
is a smoother factor involving smaller primes. The rest of the bad interval creates some congruence conditions on the product
. Using some character sum estimates coming from the Burgess bounds, we find that the residue of
becomes fairly equidistributed amongst the primitive congruence classes to a given modulus when one perturbs the primes
randomly (there are some complications from exceptional characters of Siegel zero type, but we can use a large values estimate to keep their total contribution under control). This allows us to show that the congruence conditions coming from the bad interval are restrictive enough to make non-trivial bad intervals quite rare compared to bad points. One innovation in this regard is to set up an “anti-sieve”: the elements of a bad interval tend to have an elevated chance of being divisible by small primes, and one can use moment methods to show that an excessive number of small prime divisors is somewhat rare. This can be compared to standard sieve arguments, which often seek to limit the event that a number has an unexpectedly deficient number of small prime divisors.
I’ve just uploaded to the arXiv the paper “Decomposing a factorial into large factors“. This paper studies the quantity , defined as the largest quantity such that it is possible to factorize
into
factors
, each of which is at least
. The first few values of this sequence are
This quantity was introduced by Erdös, who asked for upper and lower bounds on
; informally, this asks how equitably one can split up
into
factors. When factoring an arbitrary number, this is essentially a variant of the notorious knapsack problem (after taking logarithms), but one can hope that the specific structure of the factorial
can make this particular knapsack-type problem more tractable. Since
Some further exploration of was conducted by Guy and Selfridge. There is a simple construction that gives the lower bound
- (i) Is
for all
?
- (ii) Is
for all
? (At
, this conjecture barely fails:
.)
- (iii) Is
for all
?
In this note we establish the bounds , where
is the explicit constant
The upper bound argument for (3) is simple enough that it could also be modified to establish the first conjecture (i) of Guy and Selfridge; in principle, (ii) and (iii) are now also reducible to a finite computation, but unfortunately the implied constants in the lower bound of (3) are too weak to make this directly feasible. However, it may be possible to now crowdsource the verification of (ii) and (iii) by supplying a suitable set of factorizations to cover medium sized , combined with some effective version of the lower bound argument that can establish
for all
past a certain threshold. The value
singled out by Guy and Selfridge appears to be quite a suitable test case: the constructions I tried fell just a little short of the conjectured threshold of
, but it seems barely within reach that a sufficiently efficient rearrangement of factors can work here.
We now describe the proof of the upper and lower bound in (3). To improve upon the trivial upper bound (1), one can use the large prime factors of . Indeed, every prime
between
and
divides
at least once (and the ones between
and
divide it twice), and any factor
that contains such a factor therefore has to be significantly larger than the benchmark value of
. This observation already readily leads to some upper bound of the shape (4) for some
; if one also uses the primes
that are slightly less than
(noting that any multiple of
that exceeds
, must in fact exceed
) is what leads to the precise constant
.
For previous lower bound constructions, one started with the initial factorization and then tried to “improve” this factorization by moving around some of the prime factors. For the lower bound in (3), we start instead with an approximate factorization roughly of the shape
The general approach of first locating some approximate factorization of (where the approximation is in the “adelic” sense of having not just approximately the right magnitude, but also approximately the right number of factors of
for various primes
), and then moving factors around to get an exact factorization of
, looks promising for also resolving the conjectures (ii), (iii) mentioned above. For instance, I was numerically able to verify that
by the following procedure:
- Start with the approximate factorization of
,
by
. Thus
is the product of
odd numbers, each of which is at least
.
- Call an odd prime
-heavy if it divides
more often than
, and
-heavy if it divides
more often than
. It turns out that there are
more
-heavy primes than
-heavy primes (counting multiplicity). On the other hand,
contains
powers of
, while
has none. This represents the (multi-)set of primes one has to redistribute in order to convert a factorization of
to a factorization of
.
- Using a greedy algorithm, one can match a
-heavy prime
to each
-heavy prime
(counting multiplicity) in such a way that
for a small
(in most cases one can make
, and often one also has
). If we then replace
in the factorization of
by
for each
-heavy prime
, this increases
(and does not decrease any of the
factors of
), while eliminating all the
-heavy primes. With a somewhat crude matching algorithm, I was able to do this using
of the
powers of
dividing
, leaving
powers remaining at my disposal. (I don’t claim that this is the most efficient matching, in terms of powers of two required, but it sufficed.)
- There are still
-heavy primes left over in the factorization of (the modified version of)
. Replacing each of these primes with
, and then distributing the remaining
powers of two arbitrarily, this obtains a factorization of
into
terms, each of which are at least
.
However, I was not able to adjust parameters to reach in this manner. Perhaps some readers here who are adept with computers can come up with a more efficient construction to get closer to this bound? If one can find a way to reach this bound, most likely it can be adapted to then resolve conjectures (ii) and (iii) above after some additional numerical effort.
UPDATE: There is now an active Github project to track the latest progress, coming from multiple contributors.
In this supplemental set of notes we derive some approximations for , when
is large, and in particular Stirling’s formula. This formula (and related formulae for binomial coefficients
will be useful for estimating a number of combinatorial quantities in this course, and also in allowing one to analyse discrete random walks accurately.

Recent Comments