Well the other day I improved the performance of a particualr piece of code from 34sec to 2 sec and I was calcuating the percentage for the same i.e (34-2)/34 i.e 94.11 percentage and when I told this number out in a meeting people were not that amazed.. I am wondering if that was a wrong number that I communicated..
How do you generally measure the improvement and look good at the same time?
Speed (or throughput) is proportional to the reciprocal of time. So it's actually a factor of 34/2 = 17x
faster (which you can state as a (34-2)/2 = 1600%
speed increase if you want to sound impressive).