latency - TarisMajor/5143-OpSystems GitHub Wiki
Latency
Figure: Explaining how latency is measured.
Latency refers to the time it takes for a system to respond to an input or request, often measured in milliseconds (ms). It is a critical metric in real-time applications such as online gaming, video conferencing, and financial trading, where delays can degrade user experience or lead to missed opportunities.
Latency can arise at various stages in a system, including network latency (delays in data transmission between systems), disk latency (the time it takes to access data from storage), and processing latency (the time it takes for the CPU to process data). Reducing latency involves optimizing hardware, software, and network components to ensure that data is processed as quickly as possible.
Sources:
Arpaci-Dusseau, R. H., & Arpaci-Dusseau, A. C. (2018). Operating Systems: Three Easy Pieces. Stallings, W. (2013). Data and Computer Communications.