The natural unit of information (symbol nat), sometimes also nit or nepit, is a unit of information or entropy, based on natural logarithms and powers of e, rather than the powers of 2 and base 2 logarithms, which define the bit. This unit is also known by its unit symbol, the nat. The nat is the natural unit for information entropy. The International System of Units, by assigning the same units (joule per kelvin) both to heat capacity and to thermodynamic entropy implicitly treats information entropy as a quantity of dimension one, with 1 nat = 1. Physical systems of natural units that normalize Boltzmann's constant to 1 are effectively measuring thermodynamic entropy in nats.
When the Shannon entropy is written using a natural logarithm,
it is implicitly giving a number measured in nats.
One nat is equal to 1/ln 2 shannons (or bits) ≈ 1.44 Sh or, equivalently, 1/ln 10 hartleys ≈ 0.434 Hart. The factors 1.44 and 0.434 arise from the relationships