Big O notation
Big O Notation (or the Big O) is used to describe how long and complex an operation will be based on its input.
Complexity could mean that an operation takes N amount of time, or N amount of memory, N CPU resources, etc.
There are some notations to describe this:
O(n)-> The complexity grows linearly based on the size of the input.O(n^2)-> Grows at a square ratio of its input.O(n^3)-> Grows at a cube ratio of its input.O(n^4)-> And so on.
In the previous notations, the complexity always grow at a minimum rate of O(n) and grows higher and higher, but what if the complexity grows at a lower rate than that?
This is where logarithm notations can help describe the complexity.
But first, what is logarithm or log?
A logarithm is the exponent on which a number is raised, for example:
b^p = n
2^3 = 2x2x2
2^3 = 8In this case, p is the logarithm
Another example:
log(10)^10,000 = x
10^x = 10,000
10^4 = 10,000
log(10)^10,000 = 4Now that we know that the log is just an exponent to raise a base (p) we can say that:
O(log(n))-> grows at a logarithmic rate based on its input.
complexity described in
O(log(n))is used to define “efficient” algorithms.
For example, this binary search example:
On each iteration of this loop the input size is halved, which means that the exponent or p of this function is O(log2(n))
Which means that we only need 10 iteratios to find the value in a list of 1024 elements.
Also,
O(log(log(n)))-> Grows at a double logarithm rate. or the complexity increases slowerO(log(log(log(n))))-> and so on, similar toO(n^x).
Last updated