Neural-Symbolic Machine-Learning for Knowledge Discovery and Adaptive Information Retrieval

In this paper, a model for an information retrieval system is proposed which takes into account that knowledge about documents and information need of users are dynamic. Two methods are combined, one qualitative or symbolic and the other quantitative or numeric, which are deemed suitable for many clustering contexts, data analysis, concept exploring and knowledge discovery. These two methods may be classified as inductive learning techniques. In this model, they are introduced to build “long term" knowledge about past queries and concepts in a collection of documents. The “long term" knowledge can guide and assist the user to formulate an initial query and can be exploited in the process of retrieving relevant information. The different kinds of knowledge are organized in different points of view. This may be considered an enrichment of the exploration level which is coherent with the concept of document/query structure.

Protein Residue Contact Prediction using Support Vector Machine

Protein residue contact map is a compact representation of secondary structure of protein. Due to the information hold in the contact map, attentions from researchers in related field were drawn and plenty of works have been done throughout the past decade. Artificial intelligence approaches have been widely adapted in related works such as neural networks, genetic programming, and Hidden Markov model as well as support vector machine. However, the performance of the prediction was not generalized which probably depends on the data used to train and generate the prediction model. This situation shown the importance of the features or information used in affecting the prediction performance. In this research, support vector machine was used to predict protein residue contact map on different combination of features in order to show and analyze the effectiveness of the features.

Communication Engineering Curriculum (Past, Present and the Future)

At present time, competition, unpredictable fluctuations have made communication engineering education in the global sphere really difficult. Confront with new situation in the engineering education sector. Communication engineering education has to be reformed and ready to use more advanced technologies. We realized that one of the general problems of student`s education is that after graduating from their universities, they are not prepared to face the real life challenges and full skilled to work in industry. They are prepared only to think like engineers and professionals but they also need to possess some others non-technical skills. In today-s environment, technical competence alone is not sufficient for career success. Employers want employees (graduate engineers) who have good oral and written communication (soft) skills. It does require for team work, business awareness, organization, management skills, responsibility, initiative, problem solving and IT competency. This proposed curriculum brings interactive, creative, interesting, effective learning methods, which includes online education, virtual labs, practical work, problem-based learning (PBL), and lectures given by industry experts. Giving short assignments, presentations, reports, research papers and projects students can significantly improve their non-technical skills. Also, we noticed the importance of using ICT technologies in engineering education which used by students and teachers, and included that into proposed teaching and learning methods. We added collaborative learning between students through team work which builds theirs skills besides course materials. The prospective on this research that we intent to update communication engineering curriculum in order to get fully constructed engineer students to ready for real industry work.

A Case Study on Appearance Based Feature Extraction Techniques and Their Susceptibility to Image Degradations for the Task of Face Recognition

Over the past decades, automatic face recognition has become a highly active research area, mainly due to the countless application possibilities in both the private as well as the public sector. Numerous algorithms have been proposed in the literature to cope with the problem of face recognition, nevertheless, a group of methods commonly referred to as appearance based have emerged as the dominant solution to the face recognition problem. Many comparative studies concerned with the performance of appearance based methods have already been presented in the literature, not rarely with inconclusive and often with contradictory results. No consent has been reached within the scientific community regarding the relative ranking of the efficiency of appearance based methods for the face recognition task, let alone regarding their susceptibility to appearance changes induced by various environmental factors. To tackle these open issues, this paper assess the performance of the three dominant appearance based methods: principal component analysis, linear discriminant analysis and independent component analysis, and compares them on equal footing (i.e., with the same preprocessing procedure, with optimized parameters for the best possible performance, etc.) in face verification experiments on the publicly available XM2VTS database. In addition to the comparative analysis on the XM2VTS database, ten degraded versions of the database are also employed in the experiments to evaluate the susceptibility of the appearance based methods on various image degradations which can occur in "real-life" operating conditions. Our experimental results suggest that linear discriminant analysis ensures the most consistent verification rates across the tested databases.

Challenges to Enable Quick Start of an Environmental Monitoring with Wireless Sensor Network Technology

With the advancement of wireless sensor network technology, its practical utilization is becoming an important challange. This paper overviews my past environmental monitoring project, and discusses the process of starting the monitoring by classifying it into four steps. The steps to start environmental monitoring can be complicated, but not well discussed by researchers of wireless sensor network technology. This paper demonstrates our activity and challenges in each of the four steps to ease the process, and argues future challenges to enable quick start of environmental monitoring.

Continuous Flow Experimental Set-Up for Fouling Deposit Study

The study of the fouling deposition of pink guava juice (PGJ) is relatively new research compared to milk fouling deposit. In this work, a new experimental set-up was developed to imitate the fouling formation in heat exchanger, namely a continuous flow experimental set-up heat exchanger. The new experimental setup was operated under industrial pasteurization temperature of PGJ, which was at 93°C. While the flow rate and pasteurization period were based on the experimental capacity, which were 0.5 and 1 liter/min for the flow rate and the pasteurization period was set for 1 hour. Characterization of the fouling deposit was determined by using various methods. Microstructure of the deposits was carried out using ESEM. Proximate analyses were performed to determine the composition of moisture, fat, protein, fiber, ash and carbohydrate content. A study on the hardness and stickiness of the fouling deposit was done using a texture analyzer. The presence of seedstone in pink guava juice was also analyzed using a particle analyzer. The findings shown that seedstone from pink guava juice ranging from 168 to 200μm and carbohydrate was found to be a major composition (47.7% of fouling deposit consists of carbohydrate). Comparison between the hardness and stickiness of the deposits at two different flow rates showed that fouling deposits were harder and denser at higher flow rate. Findings from this work provide basis knowledge for further study on fouling and cleaning of PGJ.

Evaluation of Chiller Power Consumption Using Grey Prediction

98% of the energy needed in Taiwan has been imported. The prices of petroleum and electricity have been increasing. In addition, facility capacity, amount of electricity generation, amount of electricity consumption and number of Taiwan Power Company customers have continued to increase. For these reasons energy conservation has become an important topic. In the past linear regression was used to establish the power consumption models for chillers. In this study, grey prediction is used to evaluate the power consumption of a chiller so as to lower the total power consumption at peak-load (so that the relevant power providers do not need to keep on increasing their power generation capacity and facility capacity). In grey prediction, only several numerical values (at least four numerical values) are needed to establish the power consumption models for chillers. If PLR, the temperatures of supply chilled-water and return chilled-water, and the temperatures of supply cooling-water and return cooling-water are taken into consideration, quite accurate results (with the accuracy close to 99% for short-term predictions) may be obtained. Through such methods, we can predict whether the power consumption at peak-load will exceed the contract power capacity signed by the corresponding entity and Taiwan Power Company. If the power consumption at peak-load exceeds the power demand, the temperature of the supply chilled-water may be adjusted so as to reduce the PLR and hence lower the power consumption.

Novel Hybrid Method for Gene Selection and Cancer Prediction

Microarray data profiles gene expression on a whole genome scale, therefore, it provides a good way to study associations between gene expression and occurrence or progression of cancer. More and more researchers realized that microarray data is helpful to predict cancer sample. However, the high dimension of gene expressions is much larger than the sample size, which makes this task very difficult. Therefore, how to identify the significant genes causing cancer becomes emergency and also a hot and hard research topic. Many feature selection algorithms have been proposed in the past focusing on improving cancer predictive accuracy at the expense of ignoring the correlations between the features. In this work, a novel framework (named by SGS) is presented for stable gene selection and efficient cancer prediction . The proposed framework first performs clustering algorithm to find the gene groups where genes in each group have higher correlation coefficient, and then selects the significant genes in each group with Bayesian Lasso and important gene groups with group Lasso, and finally builds prediction model based on the shrinkage gene space with efficient classification algorithm (such as, SVM, 1NN, Regression and etc.). Experiment results on real world data show that the proposed framework often outperforms the existing feature selection and prediction methods, say SAM, IG and Lasso-type prediction model.

An Autonomous Collaborative Forecasting System Implementation – The First Step towards Successful CPFR System

In the past decade, artificial neural networks (ANNs) have been regarded as an instrument for problem-solving and decision-making; indeed, they have already done with a substantial efficiency and effectiveness improvement in industries and businesses. In this paper, the Back-Propagation neural Networks (BPNs) will be modulated to demonstrate the performance of the collaborative forecasting (CF) function of a Collaborative Planning, Forecasting and Replenishment (CPFR®) system. CPFR functions the balance between the sufficient product supply and the necessary customer demand in a Supply and Demand Chain (SDC). Several classical standard BPN will be grouped, collaborated and exploited for the easy implementation of the proposed modular ANN framework based on the topology of a SDC. Each individual BPN is applied as a modular tool to perform the task of forecasting SKUs (Stock-Keeping Units) levels that are managed and supervised at a POS (point of sale), a wholesaler, and a manufacturer in an SDC. The proposed modular BPN-based CF system will be exemplified and experimentally verified using lots of datasets of the simulated SDC. The experimental results showed that a complex CF problem can be divided into a group of simpler sub-problems based on the single independent trading partners distributed over SDC, and its SKU forecasting accuracy was satisfied when the system forecasted values compared to the original simulated SDC data. The primary task of implementing an autonomous CF involves the study of supervised ANN learning methodology which aims at making “knowledgeable" decision for the best SKU sales plan and stocks management.

Importance of the Green Belts to Reduce Noise Pollution and Determination of Roadside Noise Reduction Effectiveness of Bushes in Konya, Turkey

The impact of noise upon live quality has become an important aspect to make both urban and environmental policythroughout Europe and in Turkey. Concern over the quality of urban environments, including noise levels and declining quality of green space, is over the past decade with increasing emphasis on designing livable and sustainable communities. According to the World Health Organization, noise pollution is the third most hazardous environmental type of pollution which proceeded by only air (gas emission) and water pollution. The research carried out in two phases, the first stage of the research noise and plant types providing the suction of noise was evaluated through literature study and at the second stage, definite types (Juniperus horizontalis L., Spirea vanhouetti Briot., Cotoneaster dammerii C.K., Berberis thunbergii D.C., Pyracantha coccinea M. etc.) were selected for the city of Konya. Trials were conducted on the highway of Konya. The biggest value of noise reduction was 6.3 dB(A), 4.9 dB(A), 6.2 dB(A) value with compared to the control which includes the group that formed by the bushes at the distance of 7m, 11m, 20m from the source and 5m, 9m, 20m of plant width, respectively. In this paper, definitions regarding to noise and its sources were made and the precautions were taken against to noise that mentioned earlier with the adverse effects of noise. Plantation design approaches and suggestions concerning to the diversity to be used, which are peculiar to roadside, were developed to discuss the role and the function of plant material to reduce the noise of the traffic.

Evaluation of Aerodynamic Noise Generation by a Generic Side Mirror

The aerodynamic noise radiation from a side view mirror (SVM) in the high-speed airflow is calculated by the combination of unsteady incompressible fluid flow analysis and acoustic analysis. The transient flow past the generic SVM is simulated with variable turbulence model, namely DES Detached Eddy Simulation and LES (Large Eddy Simulation). Detailed velocity vectors and contour plots of the time-varying velocity and pressure fields are presented along cut planes in the flow-field. Mean and transient pressure are also monitored at several points in the flow field and compared to corresponding experimentally data published in literature. The acoustic predictions made using the Ffowcs-Williams-Hawkins acoustic analogy (FW-H) and the boundary element (BEM).

Finite Element Modeling of Rotating Mixing of Toothpaste

The objective of this research is to examine the shear thinning behaviour of mixing flow of non-Newtonian fluid like toothpaste in the dissolution container with rotating stirrer. The problem under investigation is related to the chemical industry. Mixing of fluid is performed in a cylindrical container with rotating stirrer, where stirrer is eccentrically placed on the lid of the container. For the simulation purpose the associated motion of the fluid is considered as revolving of the container, with stick stirrer. For numerical prediction, a time-stepping finite element algorithm in a cylindrical polar coordinate system is adopted based on semi-implicit Taylor-Galerkin/pressure-correction scheme. Numerical solutions are obtained for non-Newtonian fluids employing power law model. Variations with power law index have been analysed, with respect to the flow structure and pressure drop.

Assessing Land Cover Change Trajectories in Olomouc, Czech Republic

Olomouc is a unique and complex landmark with widespread forestation and land use. This research work was conducted to assess important and complex land use change trajectories in Olomouc region. Multi-temporal satellite data from 1991, 2001 and 2013 were used to extract land use/cover types by object oriented classification method. To achieve the objectives, three different aspects were used: (1) Calculate the quantity of each transition; (2) Allocate location based landscape pattern (3) Compare land use/cover evaluation procedure. Land cover change trajectories shows that 16.69% agriculture, 54.33% forest and 21.98% other areas (settlement, pasture and water-body) were stable in all three decade. Approximately 30% of the study area maintained as a same land cove type from 1991 to 2013. Here broad scale of political and socioeconomic factors was also affect the rate and direction of landscape changes. Distance from the settlements was the most important predictor of land cover change trajectories. This showed that most of landscape trajectories were caused by socio-economic activities and mainly led to virtuous change on the ecological environment.

Enhanced QoS Mechanisms for IEEE 802.11e Wireless Networks

The quality-of-service (QoS) support for wireless LANs has been a hot research topic during the past few years. In this paper, two QoS provisioning mechanisms are proposed for the employment in 802.11e EDCA MAC scheme. First, the proposed call admission control mechanism can not only guarantee the QoS for the higher priority existing connections but also provide the minimum reserved bandwidth for traffic flows with lower priority. In addition, the adaptive contention window adjustment mechanism can adjust the maximum and minimum contention window size dynamically according to the existing connection number of each AC. The collision probability as well as the packet delay will thus be reduced effectively. Performance results via simulations have revealed the enhanced QoS property achieved by employing these two mechanisms.

Taxonomy of Structured P2P Overlay Networks Security Attacks

The survey and classification of the different security attacks in structured peer-to-peer (P2P) overlay networks can be useful to computer system designers, programmers, administrators, and users. In this paper, we attempt to provide a taxonomy of structured P2P overlay networks security attacks. We have specially focused on the way these attacks can arise at each level of the network. Moreover, we observed that most of the existing systems such as Content Addressable Network (CAN), Chord, Pastry, Tapestry, Kademlia, and Viceroy suffer from threats and vulnerability which lead to disrupt and corrupt their functioning. We hope that our survey constitutes a good help for who-s working on this area of research.

A DCT-Based Secure JPEG Image Authentication Scheme

The challenge in the case of image authentication is that in many cases images need to be subjected to non malicious operations like compression, so the authentication techniques need to be compression tolerant. In this paper we propose an image authentication system that is tolerant to JPEG lossy compression operations. A scheme for JPEG grey scale images is proposed based on a data embedding method that is based on a secret key and a secret mapping vector in the frequency domain. An encrypted feature vector extracted from the image DCT coefficients, is embedded redundantly, and invisibly in the marked image. On the receiver side, the feature vector from the received image is derived again and compared against the extracted watermark to verify the image authenticity. The proposed scheme is robust against JPEG compression up to a maximum compression of approximately 80%,, but sensitive to malicious attacks such as cutting and pasting.

Development of Intelligent Time/Frequency Based Signal Detection Algorithm for Intrusion Detection System

For the past couple of decades Weak signal detection is of crucial importance in various engineering and scientific applications. It finds its application in areas like Wireless communication, Radars, Aerospace engineering, Control systems and many of those. Usually weak signal detection requires phase sensitive detector and demodulation module to detect and analyze the signal. This article gives you a preamble to intrusion detection system which can effectively detect a weak signal from a multiplexed signal. By carefully inspecting and analyzing the respective signal, this system can successfully indicate any peripheral intrusion. Intrusion detection system (IDS) is a comprehensive and easy approach towards detecting and analyzing any signal that is weakened and garbled due to low signal to noise ratio (SNR). This approach finds significant importance in applications like peripheral security systems.

Application of Java-based Pointcuts in Aspect Oriented Programming (AOP) for Data Race Detection

Wide applicability of concurrent programming practices in developing various software applications leads to different concurrency errors amongst which data race is the most important. Java provides greatest support for concurrent programming by introducing various concurrency packages. Aspect oriented programming (AOP) is modern programming paradigm facilitating the runtime interception of events of interest and can be effectively used to handle the concurrency problems. AspectJ being an aspect oriented extension to java facilitates the application of concepts of AOP for data race detection. Volatile variables are usually considered thread safe, but they can become the possible candidates of data races if non-atomic operations are performed concurrently upon them. Various data race detection algorithms have been proposed in the past but this issue of volatility and atomicity is still unaddressed. The aim of this research is to propose some suggestions for incorporating certain conditions for data race detection in java programs at the volatile fields by taking into account support for atomicity in java concurrency packages and making use of pointcuts. Two simple test programs will demonstrate the results of research. The results are verified on two different Java Development Kits (JDKs) for the purpose of comparison.

Importance of Pastoral Human Factor Overloading in Land Desertification: Case Studies in Northeastern Libya

Grazing and pastoral overloading through human factors result in significant land desertification. Failure to take into account the phenomenon of desertification as a serious problem can lead to an environmental disaster because of the damages caused by land encroachment. Therefore, soil on residential and urban areas is affected because of the deterioration of vegetation. Overgrazing or grazing in open and irregular lands is practiced in these areas almost throughout the year, especially during the growth cycle of edible plants, thereby leading to their disappearance. In addition, the large number of livestock in these areas exceeds the capacity of these pastures because of pastoral land overloading, which results in deterioration and desertification in the region. In addition, rare plants, the extinction of some edible plants in the region, and the emergence of plants unsuitable for grazing, must be taken into consideration, as along with the emergence of dust and sand storms during the dry seasons (summer to autumn) due to the degradation of vegetation. These results show that strategic plans and regulations that protect the environment from desertification must be developed. Therefore, increased pastoral load is a key human factor in the deterioration of vegetation cover, leading to land desertification in this region.

The Current Implementation Status of Manufacturing Control Systems for a Key Manufacturing Industry

Manufacturing, production and service industries within Libya have struggled with many problems during the past two decades due to many difficulties. These problems have created a negative impact on the productivity and utilization of many industries around the country. This paper studies the implementation levels of the manufacturing control systems known as Manufacturing Resource Planning (MRPII) being adapted within some Libyan industries. A survey methodology has been applied for this research, based on the survey analysis, the results pointed out that the system within these industries has a modest strategy towards most of the areas that are considered as being very crucial in implementing these systems successfully. The findings also show a variation within these implementation levels with a respect to the key-elements that related to MRPII, giving the highest levels in the emphasise on financial data accuracy. The paper has also identified limitations within the investigated manufacturing and managerial areas and has pointed to where senior managers should take immediate actions in order to achieve effective implementation of MRPII within their business area.