Big O Notation
Big O notation is used in Computer Science to describe the performance or complexity of an algorithm.
Big O specifically describes the worst-case (not exactly) scenario, and can be used to describe the execution time required or the space used (e.g. in memory or on disk) by an algorithm.
O(1) describes an algorithm that will