[Home ] [Archive]   [ فارسی ]  
:: Main :: About :: Current Issue :: Archive :: Search :: Submit :: Contact ::
Main Menu
Home::
Journal Information::
Articles archive::
For Authors::
For Reviewers::
Registration::
Contact us::
Site Facilities::
::
Search in website

Advanced Search
..
Receive site information
Enter your Email in the following box to receive the site news and information.
..
:: Search published articles ::
Showing 1 results for Ashtab

Manije Sanei Tabas, Mohammadhosein Dehghan, Fatemeh Ashtab,
Volume 28, Issue 1 (9-2023)
Abstract

Variance and entropy are distinct metrics that are commonly used to measure the uncertainty of random variables. While the variance shows how a random variable spreads more than expected, the entropy measure measures the uncertainty of an information approach, in other words, it measures the average amount of information of a random variable.
 For both uniform and normal distributions, variance is a ratio of power entropy. Finding such a monotonic relationship between variance and entropy for a larger class of these two distributions is very important and useful in signal processing, machine learning, information theory, probability and statistics, for example, it is used to reduce the errors of estimators and choose a strategy. gives, on average, the greatest or nearly greatest reduction in the entropy of the distribution of the target location, and the effectiveness of this method is tested using simulations with mining assay models. In this article, the upper bound of the variance for single-mode distributions whose tails are heavier than the tails of exponential distributions is created with the help of power entropy


Page 1 from 1     

مجله اندیشه آماری Andishe _ye Amari
Persian site map - English site map - Created in 0.06 seconds with 25 queries by YEKTAWEB 4710