The Math Forum

Search All of the Math Forum:

Views expressed in these public forums are not endorsed by NCTM or The Math Forum.

Math Forum » Discussions » Math Topics » discretemath

Notice: We are no longer accepting new posts, but the forums will continue to be readable.

Topic: (no subject)
Replies: 0  

Advanced Search

Back to Topic List Back to Topic List  
Angela Richardson

Posts: 42
From: UK
Registered: 6/22/11
(no subject)
Posted: Oct 10, 2011 6:42 AM
  Click to see the message monospaced in plain text Plain Text   Click to reply to this topic Reply
att1.html (1.3 K)

I have already asked on math.stackexchange, but hope someone here will
understand what I am trying to ask ( I know I have worded it badly).
Supposing that n were known to have two prime factors, and that the computer had
a database of all the primes <?n. Then, unless n is square, one factor would be
<?n. If an algorithm were to reduce the number of possible primes <?n which
could be a factor of n by 1/2 with each step, then the worst-case-scenario
computing time is minimised. As n?? the number of primes to check approaches 2?n
/(ln?ln2). Hence, if each step were to eliminate half the possibilities, the
number of steps taken to factorise n would tend towards log_2(2?n  /
In reality there is probably no algorithm that can eliminate half the possible
factors in each step, so how much can this lower bound be improved, still
assuming an infinite database of primes?

Point your RSS reader here for a feed of the latest messages in this topic.

[Privacy Policy] [Terms of Use]

© The Math Forum at NCTM 1994-2018. All Rights Reserved.