Shannon discovered that information can be used as a measure of entropy and probability. This is both a physical and a quantitative definition which involves how much information a system can carry. And it’s not concerned with the meaning of that information.
The more information a system carries the less entropy it contains, which also happens to be the least probable state of the system. Likewise, the most probable state of a system is noise, which carries little information.
Life has a lot of information in it and this makes life a very unlikely event (tends to zero). The Universe on the other hand has tons of time to ‘play’ (tends to infinite). This makes the number of life forms in the Universe to be a zero times infinity indeterminate.