# Big O Notation

Big O notation is used in Computer Science to describe the performance or complexity of an algorithm.

Big O specifically describes the worst-case (not exactly) scenario, and can be used to describe the execution time required or the space used (e.g. in memory or on disk) by an algorithm.

*O(1)* describes an algorithm that will