The apparent brightness of an object is often measured in magnitudes. The magnitude scale is a logarithmic scale, which means that each whole number difference in magnitude corresponds to a factor of 2.512 in brightness. The brighter the object, the lower the magnitude. For example, the apparent brightness of the Sun is -26.7 magnitude, while the apparent brightness of the full Moon is -12.7 magnitude.
The apparent brightness of an object can be used to estimate its distance. For example, if two stars have the same absolute brightness (i.e. they emit the same amount of light), then the star that appears brighter will be closer to the observer.
The apparent brightness of an object can also be used to estimate its size. For example, if two stars have the same apparent brightness, then the star that is physically larger will be farther away from the observer.
The apparent brightness of an object is an important factor in determining how visible it is. For example, a star with a high apparent brightness will be more easily seen than a star with a low apparent brightness.