Dotted Decimal Notation
Definition - What does Dotted Decimal Notation mean?
Dotted decimal notation is a system of presenting numbers that is a little different from the common conventions in arithmetic as it is taught in schools. Specifically, dotted decimal notation is used in various IT contexts, including in Internet Protocol addresses.
Techopedia explains Dotted Decimal Notation
At its core, dotted decimal notation is just another way of recording and displaying a big number. It is a particular method of notation that can replace "octets" or sets bytes (eight individual bits) in machine language. Dotted decimal notation effectively puts numbers in different containers that are separated by dots or decimals.
In changing a set of bytes into dotted decimal notation, the ones and zeros making up the bite of information are enumerated into decimal-delineated numbers between 0 and 255. In IP version 4 addresses, which are 32 bits long, this results in four numbers, for example: 0.0.172.1
The dotted decimal notation system used so commonly in IP addressing is just one of several choices for representing numbers differently. Another common one is the hexadecimal system, in which traditional numbers are augmented by letters of the alphabet in a base-16 system.
Techopedia Deals: The Complete Android Developer Course - Build 14 Apps
Join thousands of others with our weekly newsletter
The 4th Era of IT Infrastructure: Superconverged Systems:
Approaches and Benefits of Network Virtualization:
Free E-Book: Public Cloud Guide:
Free Tool: Virtual Health Monitor:
Free 30 Day Trial – Turbonomic: