2023-9 trading log

18
Entropy is a measure of the amount of disorder or randomness in a system. In thermodynamics, it is a measure of the unavailability of a system's thermal energy for conversion into mechanical work, often interpreted as the degree of disorder or randomness in the system. In information theory, entropy is a measure of the uncertainty or randomness of information content in a message or signal. The greater the entropy in a message or signal, the less predictable it is.

Disclaimer

The information and publications are not meant to be, and do not constitute, financial, investment, trading, or other types of advice or recommendations supplied or endorsed by TradingView. Read more in the Terms of Use.