Abstract: Higher ground-level ozone (GLO) concentration adversely affects human health, vegetations as well as activities in the ecosystem. In Malaysia, most of the analysis on GLO concentration are carried out using the average value of GLO concentration, which refers to the centre of distribution to make a prediction or estimation. However, analysis which focuses on the higher value or extreme value in GLO concentration is rarely explored. Hence, the objective of this study is to classify the tail behaviour of GLO using generalized extreme value (GEV) distribution estimation the return level using the corresponding modelling (Gumbel, Weibull, and Frechet) of GEV distribution. The results show that Weibull distribution which is also known as short tail distribution and considered as having less extreme behaviour is the best-fitted distribution for four selected air monitoring stations in Peninsular Malaysia, namely Larkin, Pelabuhan Kelang, Shah Alam, and Tanjung Malim; while Gumbel distribution which is considered as a medium tail distribution is the best-fitted distribution for Nilai station. The return level of GLO concentration in Shah Alam station is comparatively higher than other stations. Overall, return levels increase with increasing return periods but the increment depends on the type of the tail of GEV distribution’s tail. We conduct this study by using maximum likelihood estimation (MLE) method to estimate the parameters at four selected stations in Peninsular Malaysia. Next, the validation for the fitted block maxima series to GEV distribution is performed using probability plot, quantile plot and likelihood ratio test. Profile likelihood confidence interval is tested to verify the type of GEV distribution. These results are important as a guide for early notification on future extreme ozone events.
Abstract: The forensic use of handwriting depends on the analysis, comparison, and evaluation decisions made by forensic document examiners. When using biometric technology in forensic applications, it is necessary to compute Likelihood Ratio (LR) for quantifying strength of evidence under two competing hypotheses, namely the prosecution and the defense hypotheses wherein a set of assumptions and methods for a given data set will be made. It is therefore important to know how repeatable and reproducible our estimated LR is. This paper evaluated the accuracy and reproducibility of examiners' decisions. Confidence interval for the estimated LR were presented so as not get an incorrect estimate that will be used to deliver wrong judgment in the court of Law. The estimate of LR is fundamentally a Bayesian concept and we used two LR estimators, namely Logistic Regression (LoR) and Kernel Density Estimator (KDE) for this paper. The repeatability evaluation was carried out by retesting the initial experiment after an interval of six months to observe whether examiners would repeat their decisions for the estimated LR. The experimental results, which are based on handwriting dataset, show that LR has different confidence intervals which therefore implies that LR cannot be estimated with the same certainty everywhere. Though the LoR performed better than the KDE when tested using the same dataset, the two LR estimators investigated showed a consistent region in which LR value can be estimated confidently. These two findings advance our understanding of LR when used in computing the strength of evidence in handwriting using forensics.
Abstract: The main objective of this study is to find a suitable approach to monitor the land infrastructure growth over a period of time using multispectral satellite images. Bi-temporal change detection method is unable to indicate the continuous change occurring over a long period of time. To achieve this objective, the approach used here estimates a statistical model from series of multispectral image data over a long period of time, assuming there is no considerable change during that time period and then compare it with the multispectral image data obtained at a later time. The change is estimated pixel-wise. Statistical composite hypothesis technique is used for estimating pixel based change detection in a defined region. The generalized likelihood ratio test (GLRT) is used to detect the changed pixel from probabilistic estimated model of the corresponding pixel. The changed pixel is detected assuming that the images have been co-registered prior to estimation. To minimize error due to co-registration, 8-neighborhood pixels around the pixel under test are also considered. The multispectral images from Sentinel-2 and Landsat-8 from 2015 to 2018 are used for this purpose. There are different challenges in this method. First and foremost challenge is to get quite a large number of datasets for multivariate distribution modelling. A large number of images are always discarded due to cloud coverage. Due to imperfect modelling there will be high probability of false alarm. Overall conclusion that can be drawn from this work is that the probabilistic method described in this paper has given some promising results, which need to be pursued further.
Abstract: In this research article, a comprehensive analysis is performed to compare the diagnostic performance of soluble transferrin receptor (sTfR) and sTfR/log ferritin index tests in the differential diagnosis of iron deficiency anemia (IDA) and anemia of chronic disease (ACD). The analysis is performed for both sTfR and sTfR/log ferritin index using a set of 11 studies. The overall odds ratios for sTfR and sTfR/log ferritin index were 36.79 and 119.32 respectively, using 95% confidence interval. The relative sensitivity, specificity. positive likelihood ratio (LR) and negative LR values for sTfR in relation to sTfR/log ferritin index were 81% vs 85%, 84% vs 93%, 6.31 vs 13.95 and 0.18 vs 0.14 respectively. The summary receiver operating characteristic (SROC) curves are also plotted for both sTfR and sTfR/log ferritin index. The area under SROC curves for sTfR and sTfR/log ferritin index was found to be 0.9296 and 0.9825 respectively. Although both tests are useful, the sTfR/log ferritin index seems to be more effective when compared with sTfR.
Abstract: A simple, rapid and non-invasive electromagnetic sensor (C-FAST device) was- patented; for diagnosis of HCV RNA. Aim: To test the validity of the device compared to standard HCV PCR. Subjects and Methods: The first phase was done as pilot in Egypt on 79 participants; the second phase was done in five centers: one center from Egypt, two centers from Pakistan and two centers from India (800, 92 and 113 subjects respectively). The third phase was done nationally as multicenter study on (1600) participants for ensuring its representativeness. Results: When compared to PCR technique, C-FAST device revealed sensitivity 95% to 100%, specificity 95.5% to 100%, PPV 89.5% to 100%, NPV 95% to 100% and positive likelihood ratios 21.8% to 38.5%. Conclusion: It is practical evidence that HCV nucleotides emit electromagnetic signals that can be used for its identification. As compared to PCR, C-FAST is an accurate, valid and non-invasive device.
Abstract: In this paper, two-sided uniformly normal distribution techniques were used in the derivation of monotone likelihood ratio. The approach mainly employed the parameters of the distribution for a class of all size a. The derivation technique is fast, direct and less burdensome when compared to some existing methods.
Abstract: Extreme temperature of several stations in Malaysia is
modelled by fitting the monthly maximum to the Generalized
Extreme Value (GEV) distribution. The Mann-Kendall (MK) test
suggests a non-stationary model. Two models are considered for
stations with trend and the Likelihood Ratio test is used to determine
the best-fitting model. Results show that half of the stations favour a
model which is linear for the location parameters. The return level is
the level of events (maximum temperature) which is expected to be
exceeded once, on average, in a given number of years, is obtained.
Abstract: In this work the characteristics of spatial signal detec¬tion from an antenna array in various sample cases are investigated. Cases for a various number of available prior information about the received signal and the background noise are considered. The spatial difference between a signal and noise is only used. The performance characteristics and detecting curves are presented. All test-statistics are obtained on the basis of the generalized likelihood ratio (GLR). The received results are correct for a short and long sample.
Abstract: Recent years have seen a growing trend towards the
integration of multiple information sources to support large-scale
prediction of protein-protein interaction (PPI) networks in model
organisms. Despite advances in computational approaches, the
combination of multiple “omic" datasets representing the same type
of data, e.g. different gene expression datasets, has not been
rigorously studied. Furthermore, there is a need to further investigate
the inference capability of powerful approaches, such as fullyconnected
Bayesian networks, in the context of the prediction of PPI
networks. This paper addresses these limitations by proposing a
Bayesian approach to integrate multiple datasets, some of which
encode the same type of “omic" data to support the identification of
PPI networks. The case study reported involved the combination of
three gene expression datasets relevant to human heart failure (HF).
In comparison with two traditional methods, Naive Bayesian and
maximum likelihood ratio approaches, the proposed technique can
accurately identify known PPI and can be applied to infer potentially
novel interactions.
Abstract: This paper proposes a smart design strategy for a sequential detector to reliably detect the primary user-s signal, especially in fast fading environments. We study the computation of the log-likelihood ratio for coping with a fast changing received signal and noise sample variances, which are considered random variables. First, we analyze the detectability of the conventional generalized log-likelihood ratio (GLLR) scheme when considering fast changing statistics of unknown parameters caused by fast fading effects. Secondly, we propose an efficient sensing algorithm for performing the sequential probability ratio test in a robust and efficient manner when the channel statistics are unknown. Finally, the proposed scheme is compared to the conventional method with simulation results with respect to the average number of samples required to reach a detection decision.
Abstract: In this paper we consider the problem of change
detection and non stationary signals tracking. Using parametric
estimation of signals based on least square lattice adaptive filters we
consider for change detection statistical parametric methods using
likelihood ratio and hypothesis tests. In order to track signals
dynamics, we introduce a compensation procedure in the adaptive
estimation. This will improve the adaptive estimation performances
and fasten it-s convergence after changes detection.
Abstract: The zero inflated models are usually used in modeling
count data with excess zeros where the existence of the excess zeros
could be structural zeros or zeros which occur by chance. These type
of data are commonly found in various disciplines such as finance,
insurance, biomedical, econometrical, ecology, and health sciences
which involve sex and health dental epidemiology. The most popular
zero inflated models used by many researchers are zero inflated
Poisson and zero inflated negative binomial models. In addition, zero
inflated generalized Poisson and zero inflated double Poisson models
are also discussed and found in some literature. Recently zero
inflated inverse trinomial model and zero inflated strict arcsine
models are advocated and proven to serve as alternative models in
modeling overdispersed count data caused by excessive zeros and
unobserved heterogeneity. The purpose of this paper is to review
some related literature and provide a variety of examples from
different disciplines in the application of zero inflated models.
Different model selection methods used in model comparison are
discussed.
Abstract: In recent years, the use of vector variance as a
measure of multivariate variability has received much attention in
wide range of statistics. This paper deals with a more economic
measure of multivariate variability, defined as vector variance minus
all duplication elements. For high dimensional data, this will increase
the computational efficiency almost 50 % compared to the original
vector variance. Its sampling distribution will be investigated to make
its applications possible.
Abstract: In cognitive radio (CR) systems, the primary user (PU) signal would randomly depart or arrive during the sensing period of a CR user, which is referred to as the high traffic environment. In this paper, we propose a novel spectrum sensing scheme based
on the cyclostationarity of PU signals in high traffic environments. Specifically, we obtain a test statistic by applying an estimate of spectral autocoherence function of the PU signal to the generalized- likelihood ratio. From numerical results, it is confirmed that the proposed scheme provides a better spectrum sensing performance compared with the conventional spectrum sensing scheme based on the energy of the PU signals in high traffic environments.
Abstract: Numerous divergence measures (spectral distance, cepstral
distance, difference of the cepstral coefficients, Kullback-Leibler
divergence, distance given by the General Likelihood Ratio, distance
defined by the Recursive Bayesian Changepoint Detector and the
Mahalanobis measure) are compared in this study. The measures are
used for detection of abrupt spectral changes in synthetic AR signals
via the sliding window algorithm. Two experiments are performed;
the first is focused on detection of single boundary while the second
concentrates on detection of a couple of boundaries. Accuracy of
detection is judged for each method; the measures are compared
according to results of both experiments.
Abstract: The Comparison analysis of the Wald-s and Bayestype sequential methods for testing hypotheses is offered. The merits of the new sequential test are: universality which consists in optimality (with given criteria) and uniformity of decision-making regions for any number of hypotheses; simplicity, convenience and uniformity of the algorithms of their realization; reliability of the obtained results and an opportunity of providing the errors probabilities of desirable values. There are given the Computation results of concrete examples which confirm the above-stated characteristics of the new method and characterize the considered methods in regard to each other.