Abstract: Copyright protection and ownership proof of digital multimedia are achieved nowadays by digital watermarking techniques. A text watermarking algorithm for protecting the property rights and ownership judgment of color images is proposed in this paper. Embedding is achieved by inserting texts elements randomly into the color image as noise. The YIQ image processing model is found to be faster than other image processing methods, and hence, it is adopted for the embedding process. An optional choice of encrypting the text watermark before embedding is also suggested (in case required by some applications), where, the text can is encrypted using any enciphering technique adding more difficulty to hackers. Experiments resulted in embedding speed improvement of more than double the speed of other considered systems (such as least significant bit method, and separate color code methods), and a fairly acceptable level of peak signal to noise ratio (PSNR) with low mean square error values for watermarking purposes.
Abstract: One of the key problems facing in the analysis of Computed Tomography (CT) images is the poor contrast of the images. Image enhancement can be used to improve the visual clarity and quality of the images or to provide a better transformation representation for further processing. Contrast enhancement of images is one of the acceptable methods used for image enhancement in various applications in the medical field. This will be helpful to visualize and extract details of brain infarctions, tumors, and cancers from the CT image. This paper presents a comparison study of five contrast enhancement techniques suitable for the contrast enhancement of CT images. The types of techniques include Power Law Transformation, Logarithmic Transformation, Histogram Equalization, Contrast Stretching, and Laplacian Transformation. All these techniques are compared with each other to find out which enhancement provides better contrast of CT image. For the comparison of the techniques, the parameters Peak Signal to Noise Ratio (PSNR) and Mean Square Error (MSE) are used. Logarithmic Transformation provided the clearer and best quality image compared to all other techniques studied and has got the highest value of PSNR. Comparison concludes with better approach for its future research especially for mapping abnormalities from CT images resulting from Brain Injuries.
Abstract: In this paper, we present a technique of secure watermarking of grayscale and color images. This technique consists in applying the Singular Value Decomposition (SVD) in LWT (Lifting Wavelet Transform) domain in order to insert the watermark image (grayscale) in the host image (grayscale or color image). It also uses signature in the embedding and extraction steps. The technique is applied on a number of grayscale and color images. The performance of this technique is proved by the PSNR (Pick Signal to Noise Ratio), the MSE (Mean Square Error) and the SSIM (structural similarity) computations.
Abstract: Hard seeds will not grow and can cause mold in sprouting process. Thus, the hard seeds need to be separated from the normal seeds. Near infrared hyperspectral imaging in a range of 900 to 1700 nm was implemented to develop a model by partial least squares discriminant analysis to discriminate the hard seeds from the normal seeds. The orientation of the seeds was also studied to compare the performance of the models. The model based on hilum-up orientation achieved the best result giving the coefficient of determination of 0.98, and root mean square error of prediction of 0.07 with classification accuracy was equal to 100%.
Abstract: Crop coefficient (Kc) is an important factor contributing to estimation of evapotranspiration, and is also used to determine the irrigation schedule. This study investigated and determined the monthly Kc of winter wheat (Triticum aestivum L.) using five vegetation indices (VIs): Normalized Difference Vegetation Index (NDVI), Difference Vegetation Index (DVI), Soil Adjusted Vegetation Index (SAVI), Infrared Percentage Vegetation Index (IPVI), and Ratio Vegetation Index (RVI) of four basins in Golestan province, Iran. 14 Landsat-8 images according to crop growth stage were used to estimate monthly Kc of wheat. VIs were calculated based on infrared and near infrared bands of Landsat 8 images using Geographical Information System (GIS) software. The best VIs were chosen after establishing a regression relationship among these VIs with FAO Kc and Kc that was modified for the study area by the previous research based on R² and Root Mean Square Error (RMSE). The result showed that local modified SAVI with R²= 0.767 and RMSE= 0.174 was the best index to produce monthly wheat Kc maps.
Abstract: Under the circumstance of environment deterioration, people are increasingly concerned about the quality of the environment, especially air quality. As a result, it is of great value to give accurate and timely forecast of AQI (air quality index). In order to simplify influencing factors of air quality in a city, and forecast the city’s AQI tomorrow, this study used MATLAB software and adopted the method of constructing a mathematic model of PCA-GABP to provide a solution. To be specific, this study firstly made principal component analysis (PCA) of influencing factors of AQI tomorrow including aspects of weather, industry waste gas and IAQI data today. Then, we used the back propagation neural network model (BP), which is optimized by genetic algorithm (GA), to give forecast of AQI tomorrow. In order to verify validity and accuracy of PCA-GABP model’s forecast capability. The study uses two statistical indices to evaluate AQI forecast results (normalized mean square error and fractional bias). Eventually, this study reduces mean square error by optimizing individual gene structure in genetic algorithm and adjusting the parameters of back propagation model. To conclude, the performance of the model to forecast AQI is comparatively convincing and the model is expected to take positive effect in AQI forecast in the future.
Abstract: Predictive data analysis and modeling involving machine learning techniques become challenging in presence of too many explanatory variables or features. Presence of too many features in machine learning is known to not only cause algorithms to slow down, but they can also lead to decrease in model prediction accuracy. This study involves housing dataset with 79 quantitative and qualitative features that describe various aspects people consider while buying a new house. Boruta algorithm that supports feature selection using a wrapper approach build around random forest is used in this study. This feature selection process leads to 49 confirmed features which are then used for developing predictive random forest models. The study also explores five different data partitioning ratios and their impact on model accuracy are captured using coefficient of determination (r-square) and root mean square error (rsme).
Abstract: This paper presents a denoising method called EMD-Custom that was based on Empirical Mode Decomposition (EMD) and the modified Customized Thresholding Function (Custom) algorithms. EMD was applied to decompose adaptively a noisy signal into intrinsic mode functions (IMFs). Then, all the noisy IMFs got threshold by applying the presented thresholding function to suppress noise and to improve the signal to noise ratio (SNR). The method was tested on simulated data and real ECG signal, and the results were compared to the EMD-Based signal denoising methods using the soft and hard thresholding. The results showed the superior performance of the proposed EMD-Custom denoising over the traditional approach. The performances were evaluated in terms of SNR in dB, and Mean Square Error (MSE).
Abstract: This paper aimed to establish econometrical equation models for the Nile delta region in Egypt, which will represent a basement for future predictions of Lumpy skin disease outbreaks and its pathway in relation to climate change. Data of lumpy skin disease (LSD) outbreaks were collected from the cattle farms located in the provinces representing the Nile delta region during 1 January, 2015 to December, 2015. The obtained results indicated that there was a significant association between the degree of the LSD outbreaks and the investigated climate factors (temperature, wind speed, and humidity) and the outbreaks peaked during the months of June, July, and August and gradually decreased to the lowest rate in January, February, and December. The model obtained depicted that the increment of these climate factors were associated with evidently increment on LSD outbreaks on the Nile Delta of Egypt. The model validation process was done by the root mean square error (RMSE) and means bias (MB) which compared the number of LSD outbreaks expected with the number of observed outbreaks and estimated the confidence level of the model. The value of RMSE was 1.38% and MB was 99.50% confirming that this established model described the current association between the LSD outbreaks and the change on climate factors and also can be used as a base for predicting the of LSD outbreaks depending on the climatic change on the future.
Abstract: One of the biggest challenges in nonparametric
regression is the curse of dimensionality. Additive models are known
to overcome this problem by estimating only the individual additive
effects of each covariate. However, if the model is misspecified, the
accuracy of the estimator compared to the fully nonparametric one
is unknown. In this work the efficiency of completely nonparametric
regression estimators such as the Loess is compared to the estimators
that assume additivity in several situations, including additive and
non-additive regression scenarios. The comparison is done by
computing the oracle mean square error of the estimators with regards
to the true nonparametric regression function. Then, a backward
elimination selection procedure based on the Akaike Information
Criteria is proposed, which is computed from either the additive or
the nonparametric model. Simulations show that if the additive model
is misspecified, the percentage of time it fails to select important
variables can be higher than that of the fully nonparametric approach.
A dimension reduction step is included when nonparametric estimator
cannot be computed due to the curse of dimensionality. Finally, the
Boston housing dataset is analyzed using the proposed backward
elimination procedure and the selected variables are identified.
Abstract: To encourage building owners to purchase electricity at the wholesale market and reduce building peak demand, this study aims to develop models that predict day-ahead hourly electricity consumption and demand using artificial neural network (ANN) and support vector machine (SVM). All prediction models are built in Python, with tool Scikit-learn and Pybrain. The input data for both consumption and demand prediction are time stamp, outdoor dry bulb temperature, relative humidity, air handling unit (AHU), supply air temperature and solar radiation. Solar radiation, which is unavailable a day-ahead, is predicted at first, and then this estimation is used as an input to predict consumption and demand. Models to predict consumption and demand are trained in both SVM and ANN, and depend on cooling or heating, weekdays or weekends. The results show that ANN is the better option for both consumption and demand prediction. It can achieve 15.50% to 20.03% coefficient of variance of root mean square error (CVRMSE) for consumption prediction and 22.89% to 32.42% CVRMSE for demand prediction, respectively. To conclude, the presented models have potential to help building owners to purchase electricity at the wholesale market, but they are not robust when used in demand response control.
Abstract: A relative efficiency is defined as Ridge Estimate in the general linear model. The relative efficiency is based on the Mean square error. In this paper, we put forward a parameter of Ridge Estimate and discussions are made on the relative efficiency between the ridge estimation and the General Ridge Estimate. Eventually, this paper proves that the estimation is better than the general ridge estimate, which is based on the MSE.
Abstract: The construction of Intensity-Duration-Frequency (IDF) curves is one of the most common and useful tools in order to design hydraulic structures and to provide a mathematical relationship between rainfall characteristics. IDF curves, especially those in Peninsular Malaysia, are often built using moving windows of rainfalls. However, these windows do not represent the actual rainfall events since the duration of rainfalls is usually prefixed. Hence, instead of using moving windows, this study aims to find regionalized distributions for IDF curves of extreme rainfalls based on storm events. Homogeneity test is performed on annual maximum of storm intensities to identify homogeneous regions of storms in Peninsular Malaysia. The L-moment method is then used to regionalized Generalized Extreme Value (GEV) distribution of these annual maximums and subsequently. IDF curves are constructed using the regional distributions. The differences between the IDF curves obtained and IDF curves found using at-site GEV distributions are observed through the computation of the coefficient of variation of root mean square error, mean percentage difference and the coefficient of determination. The small differences implied that the construction of IDF curves could be simplified by finding a general probability distribution of each region. This will also help in constructing IDF curves for sites with no rainfall station.
Abstract: Hydrologic models are increasingly used as tools to
predict stormwater quantity and quality from urban catchments.
However, due to a range of practical issues, most models produce
gross errors in simulating complex hydraulic and hydrologic systems.
Difficulty in finding a robust approach for model calibration is one of
the main issues. Though automatic calibration techniques are
available, they are rarely used in common commercial hydraulic and
hydrologic modelling software e.g. MIKE URBAN. This is partly
due to the need for a large number of parameters and large datasets in
the calibration process. To overcome this practical issue, a
framework for automatic calibration of a hydrologic model was
developed in R platform and presented in this paper. The model was
developed based on the time-area conceptualization. Four calibration
parameters, including initial loss, reduction factor, time of
concentration and time-lag were considered as the primary set of
parameters. Using these parameters, automatic calibration was
performed using Approximate Bayesian Computation (ABC). ABC is
a simulation-based technique for performing Bayesian inference
when the likelihood is intractable or computationally expensive to
compute. To test the performance and usefulness, the technique was
used to simulate three small catchments in Gold Coast. For
comparison, simulation outcomes from the same three catchments
using commercial modelling software, MIKE URBAN were used.
The graphical comparison shows strong agreement of MIKE URBAN
result within the upper and lower 95% credible intervals of posterior
predictions as obtained via ABC. Statistical validation for posterior
predictions of runoff result using coefficient of determination (CD),
root mean square error (RMSE) and maximum error (ME) was found
reasonable for three study catchments. The main benefit of using
ABC over MIKE URBAN is that ABC provides a posterior
distribution for runoff flow prediction, and therefore associated
uncertainty in predictions can be obtained. In contrast, MIKE
URBAN just provides a point estimate. Based on the results of the
analysis, it appears as though ABC the developed framework
performs well for automatic calibration.
Abstract: Banda Sea Collision Zone (BSCZ) is the result of the
interaction and convergence of Indo-Australian plate, Eurasian plate
and Pacific plate. This location is located in eastern Indonesia. This
zone has a very high seismic activity. In this research, we will
calculate the rate (λ) and Mean Square Error (MSE). By this result,
we will classification earthquakes distribution in the BSCZ with the
point process approach. Chi-square is used to determine the type of
earthquakes distribution in the sub region of BSCZ. The data used in
this research is data of earthquakes with a magnitude ≥ 6 SR for the
period 1964-2013 and sourced from BMKG Jakarta. This research is
expected to contribute to the Moluccas Province and surrounding
local governments in performing spatial plan document related to
disaster management.
Abstract: In this paper, the unstable angle of attack of a
FOXTROT aircraft is controlled by using Genetic Algorithm based
flight controller and the result is compared with the conventional
techniques like Tyreus-Luyben (TL), Ziegler-Nichols (ZN) and
Interpolation Rule (IR) for tuning the PID controller. In addition, the
performance indices like Mean Square Error (MSE), Integral Square
Error (ISE), and Integral Absolute Time Error (IATE) etc. are
improved by using Genetic Algorithm. It was established that the
error by using GA is very less as compared to the conventional
techniques thereby improving the performance indices of the
dynamic system.
Abstract: Presently various computational techniques are used
in modeling and analyzing environmental engineering data. In the
present study, an intra-comparison of polynomial and radial basis
kernel functions based on Support Vector Regression and, in turn, an
inter-comparison with Multi Linear Regression has been attempted in
modeling mass transfer capacity of vertical (θ = 90O) and inclined (θ
multiple plunging jets (varying from 1 to 16 numbers). The data set
used in this study consists of four input parameters with a total of
eighty eight cases, forty four each for vertical and inclined multiple
plunging jets. For testing, tenfold cross validation was used.
Correlation coefficient values of 0.971 and 0.981 along with
corresponding root mean square error values of 0.0025 and 0.0020
were achieved by using polynomial and radial basis kernel functions
based Support Vector Regression respectively. An intra-comparison
suggests improved performance by radial basis function in
comparison to polynomial kernel based Support Vector Regression.
Further, an inter-comparison with Multi Linear Regression
(correlation coefficient = 0.973 and root mean square error = 0.0024)
reveals that radial basis kernel functions based Support Vector
Regression performs better in modeling and estimating mass transfer
by multiple plunging jets.
Abstract: Predicting earnings management is vital for the capital
market participants, financial analysts and managers. The aim of this
research is attempting to respond to this query: Is there a significant
difference between the regression model and neural networks’
models in predicting earnings management, and which one leads to a
superior prediction of it? In approaching this question, a Linear
Regression (LR) model was compared with two neural networks
including Multi-Layer Perceptron (MLP), and Generalized
Regression Neural Network (GRNN). The population of this study
includes 94 listed companies in Tehran Stock Exchange (TSE)
market from 2003 to 2011. After the results of all models were
acquired, ANOVA was exerted to test the hypotheses. In general, the
summary of statistical results showed that the precision of GRNN did
not exhibit a significant difference in comparison with MLP. In
addition, the mean square error of the MLP and GRNN showed a
significant difference with the multi variable LR model. These
findings support the notion of nonlinear behavior of the earnings
management. Therefore, it is more appropriate for capital market
participants to analyze earnings management based upon neural
networks techniques, and not to adopt linear regression models.
Abstract: In the present study, the kinetics of thermal
degradation of a phenolic and lignin reinforced phenolic foams, and
the lignin used as reinforcement were studied and the activation
energies of their degradation processes were obtained by a DAEM
model. The average values for five heating rates of the mean
activation energies obtained were: 99.1, 128.2, and 144.0 kJ.mol-1 for
the phenolic foam; 109.5, 113.3, and 153.0 kJ.mol-1 for the lignin
reinforcement; and 82.1, 106.9, and 124.4 kJ.mol-1 for the lignin
reinforced phenolic foam. The standard deviation ranges calculated
for each sample were 1.27-8.85, 2.22-12.82, and 3.17-8.11 kJ.mol-1
for the phenolic foam, lignin and the reinforced foam, respectively.
The DAEM model showed low mean square errors (
Abstract: In general, classical methods such as maximum
likelihood (ML) and least squares (LS) estimation methods are used
to estimate the shape parameters of the Burr XII distribution.
However, these estimators are very sensitive to the outliers. To
overcome this problem we propose alternative robust estimators
based on the M-estimation method for the shape parameters of the
Burr XII distribution. We provide a small simulation study and a real
data example to illustrate the performance of the proposed estimators
over the ML and the LS estimators. The simulation results show that
the proposed robust estimators generally outperform the classical
estimators in terms of bias and root mean square errors when there
are outliers in data.