Search All of the Math Forum:

Views expressed in these public forums are not endorsed by NCTM or The Math Forum.

Notice: We are no longer accepting new posts, but the forums will continue to be readable.

Topic: (no subject)
Replies: 0

 Angela Richardson Posts: 42 From: UK Registered: 6/22/11
(no subject)
Posted: Oct 10, 2011 6:42 AM
 att1.html (1.3 K)

understand what I am trying to ask ( I know I have worded it badly).

Supposing that n were known to have two prime factors, and that the computer had
a database of all the primes <?n. Then, unless n is square, one factor would be
<?n. If an algorithm were to reduce the number of possible primes <?n which
could be a factor of n by 1/2 with each step, then the worst-case-scenario
computing time is minimised. As n?? the number of primes to check approaches 2?n
/(ln?ln2). Hence, if each step were to eliminate half the possibilities, the
number of steps taken to factorise n would tend towards log_2(2?n  /
(ln(n)?ln(2))).

In reality there is probably no algorithm that can eliminate half the possible
factors in each step, so how much can this lower bound be improved, still
assuming an infinite database of primes?