LOWL: Logic and OWL, an Extension

Current research on semantic web aims at making intelligent web pages meaningful for machines. In this way, ontology plays a primary role. We believe that logic can help ontology languages (such as OWL) to be more fluent and efficient. In this paper we try to combine logic with OWL to reduce some disadvantages of this language. Therefore we extend OWL by logic and also show how logic can satisfy our future expectations of an ontology language.

Assessment of Time-Lapse in Visible and Thermal Face Recognition

Although face recognition seems as an easy task for human, automatic face recognition is a much more challenging task due to variations in time, illumination and pose. In this paper, the influence of time-lapse on visible and thermal images is examined. Orthogonal moment invariants are used as a feature extractor to analyze the effect of time-lapse on thermal and visible images and the results are compared with conventional Principal Component Analysis (PCA). A new triangle square ratio criterion is employed instead of Euclidean distance to enhance the performance of nearest neighbor classifier. The results of this study indicate that the ideal feature vectors can be represented with high discrimination power due to the global characteristic of orthogonal moment invariants. Moreover, the effect of time-lapse has been decreasing and enhancing the accuracy of face recognition considerably in comparison with PCA. Furthermore, our experimental results based on moment invariant and triangle square ratio criterion show that the proposed approach achieves on average 13.6% higher in recognition rate than PCA.

Using the V-Sphere Code for the Passive Scalar in the Wake of a Bluff Body

The objective of this research was to find the diffusion properties of vehicles on the road by using the V-Sphere Code. The diffusion coefficient and the size of the height of the wake were estimated with the LES option and the third order MUSCL scheme. We evaluated the code with the changes in the moments of Reynolds Stress along the mean streamline. The results show that at the leading part of a bluff body the LES has some advantages over the RNS since the changes in the strain rates are larger for the leading part. We estimated that the diffusion coefficient with the computed Reynolds stress (non-dimensional) was about 0.96 times the mean velocity.

Micromachining with ECDM: Research Potentials and Experimental Investigations

Electro Chemical Discharge Machining (ECDM) is an emerging hybrid machining process used in precision machining of hard and brittle non-conducting materials. The present paper gives a critical review on materials machined by ECDM under the prevailing machining conditions; capability indicators of the process are reported. Some results obtained while performing experiments in micro-channeling on soda lime glass using ECDM are also presented. In these experiments, Tool Wear (TW) and Material Removal (MR) were studied using design of experiments and L–4 orthogonal array. Experimental results showed that the applied voltage was the most influencing parameter in both MR and TW studies. Field emission scanning electron microscopy (FESEM) results obtained on the microchannels confirmed the presence of micro-cracks, primarily responsible for MR. Chemical etching was also seen along the edges. The Energy dispersive spectroscopy (EDS) results were used to detect the elements present in the debris and specimens.

Simulation Model for Predicting Dengue Fever Outbreak

Dengue fever is prevalent in Malaysia with numerous cases including mortality recorded over the years. Public education on the prevention of the desease through various means has been carried out besides the enforcement of legal means to eradicate Aedes mosquitoes, the dengue vector breeding ground. Hence, other means need to be explored, such as predicting the seasonal peak period of the dengue outbreak and identifying related climate factors contributing to the increase in the number of mosquitoes. Simulation model can be employed for this purpose. In this study, we created a simulation of system dynamic to predict the spread of dengue outbreak in Hulu Langat, Selangor Malaysia. The prototype was developed using STELLA 9.1.2 software. The main data input are rainfall, temperature and denggue cases. Data analysis from the graph showed that denggue cases can be predicted accurately using these two main variables- rainfall and temperature. However, the model will be further tested over a longer time period to ensure its accuracy, reliability and efficiency as a prediction tool for dengue outbreak.

The Effect of Innovation Factors to Customer Loyalty by Structural Equation Model

Innovation is being view from four areas of innovation, product, service, technology, and marketing. Whereas customer loyalty is composed of customer expectation, perceived quality, perceived value, corporate image, customer satisfaction, customer trust/confidence, customer commitment, customer complaint, and customer loyalty. This study aimed to investigate the influence of innovation factors to customer loyalty to GSM in the telecom companies where use of products and services. Structural Equation Modeling (SEM) using to analyze innovation factors. It was found the factor of innovation have significant influence on customer loyalty.

Exponential Particle Swarm Optimization Approach for Improving Data Clustering

In this paper we use exponential particle swarm optimization (EPSO) to cluster data. Then we compare between (EPSO) clustering algorithm which depends on exponential variation for the inertia weight and particle swarm optimization (PSO) clustering algorithm which depends on linear inertia weight. This comparison is evaluated on five data sets. The experimental results show that EPSO clustering algorithm increases the possibility to find the optimal positions as it decrease the number of failure. Also show that (EPSO) clustering algorithm has a smaller quantization error than (PSO) clustering algorithm, i.e. (EPSO) clustering algorithm more accurate than (PSO) clustering algorithm.

Speaker Independent Quranic Recognizer Basedon Maximum Likelihood Linear Regression

An automatic speech recognition system for the formal Arabic language is needed. The Quran is the most formal spoken book in Arabic, it is spoken all over the world. In this research, an automatic speech recognizer for Quranic based speakerindependent was developed and tested. The system was developed based on the tri-phone Hidden Markov Model and Maximum Likelihood Linear Regression (MLLR). The MLLR computes a set of transformations which reduces the mismatch between an initial model set and the adaptation data. It uses the regression class tree, as well as, estimates a set of linear transformations for the mean and variance parameters of a Gaussian mixture HMM system. The 30th Chapter of the Quran, with five of the most famous readers of the Quran, was used for the training and testing of the data. The chapter includes about 2000 distinct words. The advantages of using the Quranic verses as the database in this developed recognizer are the uniqueness of the words and the high level of orderliness between verses. The level of accuracy from the tested data ranged 68 to 85%.

HIV Treatment Planning on a case-by-CASE Basis

This study presents a mathematical modeling approach to the planning of HIV therapies on an individual basis. The model replicates clinical data from typical-progressors to AIDS for all stages of the disease with good agreement. Clinical data from rapid-progressors and long-term non-progressors is also matched by estimation of immune system parameters only. The ability of the model to reproduce these phenomena validates the formulation, a fact which is exploited in the investigation of effective therapies. The therapy investigation suggests that, unlike continuous therapy, structured treatment interruptions (STIs) are able to control the increase in both the drug-sensitive and drug-resistant virus population and, hence, prevent the ultimate progression from HIV to AIDS. The optimization results further suggest that even patients characterised by the same progression type can respond very differently to the same treatment and that the latter should be designed on a case-by-case basis. Such a methodology is presented here.

Transmitter Design for LMS-MIMO-MCCDMA Systems with Pilot Channel Estimates and Zero Forcing Equalizer

We propose a downlink multiple-input multipleoutput (MIMO) multi-carrier code division multiple access (MCCDMA) system with adaptive beamforming algorithm for smart antennas. The algorithm used in this paper is based on the Least Mean Square (LMS), with pilot channel estimation (PCE) and the zero forcing equalizer (ZFE) in the receiver, requiring reference signal and no knowledge channel. MC-CDMA is studied in a multiple antenna context in order to efficiently exploit robustness against multipath effects and multi-user flexibility of MC-CDMA and channel diversity offered by MIMO systems for radio mobile channels. Computer simulations, considering multi-path Rayleigh Fading Channel, interference inter symbol and interference are presented to verify the performance. Simulation results show that the scheme achieves good performance in a multi-user system.

Development of Variable Stepsize Variable Order Block Method in Divided Difference Form for the Numerical Solution of Delay Differential Equations

This paper considers the development of a two-point predictor-corrector block method for solving delay differential equations. The formulae are represented in divided difference form and the algorithm is implemented in variable stepsize variable order technique. The block method produces two new values at a single integration step. Numerical results are compared with existing methods and it is evident that the block method performs very well. Stability regions of the block method are also investigated.

The Effects of Whole-Body Vibration Training on Jump Performance in Handball Athletes

This study examined the effects of eight weeks of whole-body vibration training (WBVT) on vertical and decuple jump performance in handball athletes. Sixteen collegiate Level I handball athletes volunteered for this study. They were divided equally as control group and experimental group (EG). During the period of the study, all athletes underwent the same handball specific training, but the EG received additional WBVT (amplitude: 2 mm, frequency: 20 - 40 Hz) three time per week for eight consecutive weeks. The vertical jump performance was evaluated according to the maximum height of squat jump (SJ) and countermovement jump (CMJ). Single factor ANCOVA was used to examine the differences in each parameter between the groups after training with the pretest values as a covariate. The statistic significance was set at p < .05. After 8 weeks WBVT, the EG had significantly improved the maximal height of SJ (40.92 ± 2.96 cm vs. 48.40 ± 4.70 cm, F = 5.14, p < .05) and the maximal height CMJ (47.25 ± 7.48 cm vs. 52.20 ± 6.25 cm, F = 5.31, p < .05). 8 weeks of additional WBVT could improve the vertical and decuple jump performance in handball athletes. Enhanced motor unit synchronization and firing rates, facilitated muscular contraction stretch-shortening cycle, and improved lower extremity neuromuscular coordination could account for these enhancements.

Generalization Kernel for Geopotential Approximation by Harmonic Splines

This paper presents a generalization kernel for gravitational potential determination by harmonic splines. It was shown in [10] that the gravitational potential can be approximated using a kernel represented as a Newton integral over the real Earth body. On the other side, the theory of geopotential approximation by harmonic splines uses spherically oriented kernels. The purpose of this paper is to show that in the spherical case both kernels have the same type of representation, which leads us to conclusion that it is possible to consider the kernel represented as a Newton integral over the real Earth body as a kind of generalization of spherically harmonic kernels to real geometries.

Using Artificial Neural Network to Forecast Groundwater Depth in Union County Well

A concern that researchers usually face in different applications of Artificial Neural Network (ANN) is determination of the size of effective domain in time series. In this paper, trial and error method was used on groundwater depth time series to determine the size of effective domain in the series in an observation well in Union County, New Jersey, U.S. different domains of 20, 40, 60, 80, 100, and 120 preceding day were examined and the 80 days was considered as effective length of the domain. Data sets in different domains were fed to a Feed Forward Back Propagation ANN with one hidden layer and the groundwater depths were forecasted. Root Mean Square Error (RMSE) and the correlation factor (R2) of estimated and observed groundwater depths for all domains were determined. In general, groundwater depth forecast improved, as evidenced by lower RMSEs and higher R2s, when the domain length increased from 20 to 120. However, 80 days was selected as the effective domain because the improvement was less than 1% beyond that. Forecasted ground water depths utilizing measured daily data (set #1) and data averaged over the effective domain (set #2) were compared. It was postulated that more accurate nature of measured daily data was the reason for a better forecast with lower RMSE (0.1027 m compared to 0.255 m) in set #1. However, the size of input data in this set was 80 times the size of input data in set #2; a factor that may increase the computational effort unpredictably. It was concluded that 80 daily data may be successfully utilized to lower the size of input data sets considerably, while maintaining the effective information in the data set.

Optimization of Protein Hydrolysate Production Process from Jatropha curcas Cake

This was the first document revealing the investigation of protein hydrolysate production optimization from J. curcas cake. Proximate analysis of raw material showed 18.98% protein, 5.31% ash, 8.52% moisture and 12.18% lipid. The appropriate protein hydrolysate production process began with grinding the J. curcas cake into small pieces. Then it was suspended in 2.5% sodium hydroxide solution with ratio between solution/ J. curcas cake at 80:1 (v/w). The hydrolysis reaction was controlled at temperature 50 °C in water bath for 45 minutes. After that, the supernatant (protein hydrolysate) was separated using centrifuge at 8000g for 30 minutes. The maximum yield of resulting protein hydrolysate was 73.27 % with 7.34% moisture, 71.69% total protein, 7.12% lipid, 2.49% ash. The product was also capable of well dissolving in water.

Transport and Fate of Copper in Soils

The presence of toxic heavy metals in industrial effluents is one of the serious threats to the environment. Heavy metals such as Cadmium, Chromium, Lead, Nickel, Zinc, Mercury, Copper, Arsenic are found in the effluents of industries such as foundries, electroplating, petrochemical, battery manufacturing, tanneries, fertilizer, dying, textiles, metallurgical and metal finishing. Tremendous increase of industrial copper usage and its presence in industrial effluents has lead to a growing concern about the fate and effects of Copper in the environment. Percolation of industrial effluents through soils leads to contamination of ground water and soils. The transport of heavy metals and their diffusion into the soils has therefore, drawn the attention of the researchers. In this study, an attempt has been made to delineate the mechanisms of transport and fate of copper in terrestrial environment. Column studies were conducted using perplex glass square column of dimension side 15 cm and 1.35 m long. The soil samples were collected from a natural drain near Mohali (India). The soil was characterized to be poorly graded sandy loam. The soil was compacted to the field dry density level of about 1.6 g/cm3. Break through curves for different depths of the column were plotted. The results of the column study indicated that the copper has high tendency to flow in the soils and fewer tendencies to get absorbed on the soil particles. The t1/2 estimates obtained from the studies can be used for design copper laden wastewater disposal systems.

Content Based Image Retrieval of Brain MR Images across Different Classes

Magnetic Resonance Imaging play a vital role in the decision-diagnosis process of brain MR images. For an accurate diagnosis of brain related problems, the experts mostly compares both T1 and T2 weighted images as the information presented in these two images are complementary. In this paper, rotational and translational invariant form of Local binary Pattern (LBP) with additional gray scale information is used to retrieve similar slices of T1 weighted images from T2 weighted images or vice versa. The incorporation of additional gray scale information on LBP can extract more local texture information. The accuracy of retrieval can be improved by extracting moment features of LBP and reweighting the features based on users feedback. Here retrieval is done in a single subject scenario where similar images of a particular subject at a particular level are retrieved, and multiple subjects scenario where relevant images at a particular level across the subjects are retrieved.

Adaptive Kernel Filtering Used in Video Processing

In this paper we present a noise reduction filter for video processing. It is based on the recently proposed two dimensional steering kernel, extended to three dimensions and further augmented to suit the spatial-temporal domain of video processing. Two alternative filters are proposed - the time symmetric kernel and the time asymmetric kernel. The first reduces the noise on single sequences, but to handle the problems at scene shift the asymmetric kernel is introduced. The performance of both are tested on simulated data and on a real video sequence together with the existing steering kernel. The proposed kernels improves the Rooted Mean Squared Error (RMSE) compared to the original steering kernel method on video material.

The Roles of Natural and Anthropogenic Factors of Ecological State in the Lake Peipsi

In this paper we discuss the problems of the long-term management policy of Lake Peipsi and the roles of natural and anthropogenic factors in the ecological state of the lake. The reduction of the pollution during the last 15 years could not give significant changes of the chemical composition of the water, what implicates the essential role that natural factors have on the ecological state of lake. One of the most important factors having impact on the hydrochemical cycles and ecological state is the hydrological regime which is clearly expressed in L. Peipsi. The absence on clear interrelations of climate cycles and nutrients suggest that complex abiotic and biotic interactions, which take place in the lake ecosystem, plays a significant role in the matter circulation mechanism within lake.

Evaluation of State of the Art IDS Message Exchange Protocols

During the last couple of years, the degree of dependence on IT systems has reached a dimension nobody imagined to be possible 10 years ago. The increased usage of mobile devices (e.g., smart phones), wireless sensor networks and embedded devices (Internet of Things) are only some examples of the dependency of modern societies on cyber space. At the same time, the complexity of IT applications, e.g., because of the increasing use of cloud computing, is rising continuously. Along with this, the threats to IT security have increased both quantitatively and qualitatively, as recent examples like STUXNET or the supposed cyber attack on Illinois water system are proofing impressively. Once isolated control systems are nowadays often publicly available - a fact that has never been intended by the developers. Threats to IT systems don’t care about areas of responsibility. Especially with regard to Cyber Warfare, IT threats are no longer limited to company or industry boundaries, administrative jurisdictions or state boundaries. One of the important countermeasures is increased cooperation among the participants especially in the field of Cyber Defence. Besides political and legal challenges, there are technical ones as well. A better, at least partially automated exchange of information is essential to (i) enable sophisticated situational awareness and to (ii) counter the attacker in a coordinated way. Therefore, this publication performs an evaluation of state of the art Intrusion Detection Message Exchange protocols in order to guarantee a secure information exchange between different entities.