According to Moore's law,it implies that computer processing speed doubles every 18 months. How many times a computer today is faster than one ten years ago?Using the same amount of time, how much bigger a problem we can solve with a computer today than with one ten years ago given that the problem takes originally 100 nano second.
10 years = 120 months = 6.666 x 18 months.
Let original speed of processing = s, so 10 years later, speed = s x 2^6.666 = 101.6s. That is 101.6 times faster.
Time required to solve problem = 100/101.6 approx. = 1 nano second.