Observer Based Control of a Class of Nonlinear Fractional Order Systems using LMI

Design of an observer based controller for a class of fractional order systems has been done. Fractional order mathematics is used to express the system and the proposed observer. Fractional order Lyapunov theorem is used to derive the closed-loop asymptotic stability. The gains of the observer and observer based controller are derived systematically using the linear matrix inequality approach. Finally, the simulation results demonstrate validity and effectiveness of the proposed observer based controller.

Simulation of the Pedestrian Flow in the Tawaf Area Using the Social Force Model

In today-s modern world, the number of vehicles is increasing on the road. This causes more people to choose walking instead of traveling using vehicles. Thus, proper planning of pedestrians- paths is important to ensure the safety of pedestrians in a walking area. Crowd dynamics study the pedestrians- behavior and modeling pedestrians- movement to ensure safety in their walking paths. To date, many models have been designed to ease pedestrians- movement. The Social Force Model is widely used among researchers as it is simpler and provides better simulation results. We will discuss the problem regarding the ritual of circumambulating the Ka-aba (Tawaf) where the entrances to this area are usually congested which worsens during the Hajj season. We will use the computer simulation model SimWalk which is based on the Social Force Model to simulate the movement of pilgrims in the Tawaf area. We will first discuss the effect of uni and bi-directional flows at the gates. We will then restrict certain gates to the area as the entrances only and others as exits only. From the simulations, we will study the effect of the distance of other entrances from the beginning line and their effects on the duration of pilgrims circumambulate Ka-aba. We will distribute the pilgrims at the different entrances evenly so that the congestion at the entrances can be reduced. We would also discuss the various locations and designs of barriers at the exits and its effect on the time taken for the pilgrims to exit the Tawaf area.

Evaluating Spectral Relationships between Signals by Removing the Contribution of a Common, Periodic Source A Partial Coherence-based Approach

Partial coherence between two signals removing the contribution of a periodic, deterministic signal is proposed for evaluating the interrelationship in multivariate systems. The estimator expression was derived and shown to be independent of such periodic signal. Simulations were used for obtaining its critical value, which were found to be the same as those for Gaussian signals, as well as for evaluating the technique. An Illustration with eletroencephalografic (EEG) signals during photic stimulation is also provided. The application of the proposed technique in both simulation and real EEG data indicate that it seems to be very specific in removing the contribution of periodic sources. The estimate independence of the periodic signal may widen partial coherence application to signal analysis, since it could be used together with simple coherence to test for contamination in signals by a common, periodic noise source.

How to Connect User Research and not so Forthcoming Technology Scenarios – The Extended Home Environment Case Study

This paper draws a methodological framework adopted within an internal Telecomitalia project aimed to identify, on a user centred base, the potential interest towards a technological scenario aimed to extend on a personal bubble the typical communication and media fruition home environment. The problem is that involving user in the early stage of the development of such disruptive technology scenario asking users opinions on something that users actually do not manage even in a rough manner could lead to wrong or distorted results. For that reason we chose an approach that indirectly aim to understand users hidden needs in order to obtain a meaningful picture of the possible interest for a technological proposition non yet easily understandable.

Optimal Data Compression and Filtering: The Case of Infinite Signal Sets

We present a theory for optimal filtering of infinite sets of random signals. There are several new distinctive features of the proposed approach. First, we provide a single optimal filter for processing any signal from a given infinite signal set. Second, the filter is presented in the special form of a sum with p terms where each term is represented as a combination of three operations. Each operation is a special stage of the filtering aimed at facilitating the associated numerical work. Third, an iterative scheme is implemented into the filter structure to provide an improvement in the filter performance at each step of the scheme. The final step of the concerns signal compression and decompression. This step is based on the solution of a new rank-constrained matrix approximation problem. The solution to the matrix problem is described in this paper. A rigorous error analysis is given for the new filter.

Classifier Combination Approach in Motion Imagery Signals Processing for Brain Computer Interface

In this study we focus on improvement performance of a cue based Motor Imagery Brain Computer Interface (BCI). For this purpose, data fusion approach is used on results of different classifiers to make the best decision. At first step Distinction Sensitive Learning Vector Quantization method is used as a feature selection method to determine most informative frequencies in recorded signals and its performance is evaluated by frequency search method. Then informative features are extracted by packet wavelet transform. In next step 5 different types of classification methods are applied. The methodologies are tested on BCI Competition II dataset III, the best obtained accuracy is 85% and the best kappa value is 0.8. At final step ordered weighted averaging (OWA) method is used to provide a proper aggregation classifiers outputs. Using OWA enhanced system accuracy to 95% and kappa value to 0.9. Applying OWA just uses 50 milliseconds for performing calculation.

Analysis of Roasted and Ground Grains on the Seoul (Korea) Market for Their Contaminants of Aflatoxins, Ochratoxin A and Fusarium Toxins by LC-MS/MS

A sensitive and specific method for quantitative determination of aflatoxins(B1, B2, G1,G2), deoxynivalenol, fumonisin(B1,B2), ochratoxin A, zearalenone, T-2 and HT-2 in roasted and ground grains using liquid chromatography combined with tandem mass spectrometry. A double extraction using a phosphate buffer solution followed by methanol was applied to achieve effective co extraction of 11 mycotoxins. A multitoxin immunoaffinity column for all these mycotoxins was used to clean up the extract. The LODs of mycotoxins were 0.1~6.1 μg/kg, LOQs were 0.3~18.4 μg/kg. Forty seven samples collected from Seoul (Korea) for mycotoxin contamination monitoring. The results showed that the occurrence of zearalenone and deoxynivalenol were frequent. Zearalenone was detected in all samples and deoxynivalenol was detected in 80.9 % samples in the range 0.626 ~ 29.264 μg/kg and N.D ~ 48.332 μg/kg respectively. Fumonisins and ochratoxin A were detected in 46.8% samples and 17 % samples respectively, aflatoxins and T-2/HT-2 toxins were not detected all samples.

Fuzzy Trust for Peer-to-Peer Based Systems

Trust management is one of the drawbacks in Peer-to-Peer (P2P) system. Lack of centralized control makes it difficult to control the behavior of the peers. Reputation system is one approach to provide trust assessment in P2P system. In this paper, we use fuzzy logic to model trust in a P2P environment. Our trust model combines first-hand (direct experience) and second-hand (reputation)information to allow peers to represent and reason with uncertainty regarding other peers' trustworthiness. Fuzzy logic can help in handling the imprecise nature and uncertainty of trust. Linguistic labels are used to enable peers assign a trust level intuitively. Our fuzzy trust model is flexible such that inference rules are used to weight first-hand and second-hand accordingly.

A New IT-Convergence Service Design Framework

In many countries, digital city or ubiquitous city (u-City) projects have been initiated to provide digitalized economic environments to cities. Recently in Korea, Kangwon Province has started the u-Kangwon project to boost local economy with digitalized tourism services. We analyze the limitations of the ubiquitous IT approach through the u-Kangwon case. We have found that travelers are more interested in quality over speed in access of information. For improved service quality, we are looking to develop an IT-convergence service design framework (ISDF). The ISDF is based on the service engineering technique and composed of three parts: Service Design, Service Simulation, and the Service Platform.

Solving the Teacher Assignment-Course Scheduling Problem by a Hybrid Algorithm

This paper presents a hybrid algorithm for solving a timetabling problem, which is commonly encountered in many universities. The problem combines both teacher assignment and course scheduling problems simultaneously, and is presented as a mathematical programming model. However, this problem becomes intractable and it is unlikely that a proven optimal solution can be obtained by an integer programming approach, especially for large problem instances. A hybrid algorithm that combines an integer programming approach, a greedy heuristic and a modified simulated annealing algorithm collaboratively is proposed to solve the problem. Several randomly generated data sets of sizes comparable to that of an institution in Indonesia are solved using the proposed algorithm. Computational results indicate that the algorithm can overcome difficulties of large problem sizes encountered in previous related works.

Role of Customers in Stakeholders- Approach in Company Corporate Governance

The purpose of this paper is to explore the relationship between the customers- issues in company corporate governance and the financial performance. At the beginning theoretical background consisting stakeholder theory and corporate governance is presented. On this theoretical background, the empirical research is built, collecting data of 60 Czech joint stock companies- boards considering their relationships with customers. Correlation analysis and multivariate regression analysis were employed to test the sample on two hypotheses. The weak positive correlation between stakeholder approach and the company size was identified. But both hypotheses were not supported, because there was no significant relation of independent variables to financial performance.

Using Genetic Algorithms in Closed Loop Identification of the Systems with Variable Structure Controller

This work presents a recursive identification algorithm. This algorithm relates to the identification of closed loop system with Variable Structure Controller. The approach suggested includes two stages. In the first stage a genetic algorithm is used to obtain the parameters of switching function which gives a control signal rich in commutations (i.e. a control signal whose spectral characteristics are closest possible to those of a white noise signal). The second stage consists in the identification of the system parameters by the instrumental variable method and using the optimal switching function parameters obtained with the genetic algorithm. In order to test the validity of this algorithm a simulation example is presented.

Optimized Data Fusion in an Intelligent Integrated GPS/INS System Using Genetic Algorithm

Most integrated inertial navigation systems (INS) and global positioning systems (GPS) have been implemented using the Kalman filtering technique with its drawbacks related to the need for predefined INS error model and observability of at least four satellites. Most recently, a method using a hybrid-adaptive network based fuzzy inference system (ANFIS) has been proposed which is trained during the availability of GPS signal to map the error between the GPS and the INS. Then it will be used to predict the error of the INS position components during GPS signal blockage. This paper introduces a genetic optimization algorithm that is used to update the ANFIS parameters with respect to the INS/GPS error function used as the objective function to be minimized. The results demonstrate the advantages of the genetically optimized ANFIS for INS/GPS integration in comparison with conventional ANFIS specially in the cases of satellites- outages. Coping with this problem plays an important role in assessment of the fusion approach in land navigation.

Biodiversity of Micromycetes Isolated from Soils of Different Agricultures in Kazakhstan and Their Plant Growth Promoting Potential

The comparative analysis of different taxonomic groups of microorganisms isolated from dark chernozem soils under different agricultures (alfalfa, melilot, sainfoin, soybean, rapeseed) at Almaty region of Kazakhstan was conducted. It was shown that the greatest number of micromycetes was typical to the soil planted with alfalfa and canola. Species diversity of micromycetes markedly decreases as it approaches the surface of the root, so that the species composition in the rhizosphere is much more uniform than in the virgin soil. Promising strains of microscopic fungi and yeast with plant growth-promoting activity to agricultures were selected. Among the selected fungi there are representatives of Penicillium bilaiae, Trichoderma koningii, Fusarium equiseti, Aspergillus ustus. The highest rates of growth and development of seedlings of plants observed under the influence of yeasts Aureobasidium pullulans, Rhodotorula mucilaginosa, Metschnikovia pulcherrima. Using molecular - genetic techniques confirmation of the identification results of selected micromycetes was conducted.

Discovery of Quantified Hierarchical Production Rules from Large Set of Discovered Rules

Automated discovery of Rule is, due to its applicability, one of the most fundamental and important method in KDD. It has been an active research area in the recent past. Hierarchical representation allows us to easily manage the complexity of knowledge, to view the knowledge at different levels of details, and to focus our attention on the interesting aspects only. One of such efficient and easy to understand systems is Hierarchical Production rule (HPRs) system. A HPR, a standard production rule augmented with generality and specificity information, is of the following form: Decision If < condition> Generality Specificity . HPRs systems are capable of handling taxonomical structures inherent in the knowledge about the real world. This paper focuses on the issue of mining Quantified rules with crisp hierarchical structure using Genetic Programming (GP) approach to knowledge discovery. The post-processing scheme presented in this work uses Quantified production rules as initial individuals of GP and discovers hierarchical structure. In proposed approach rules are quantified by using Dempster Shafer theory. Suitable genetic operators are proposed for the suggested encoding. Based on the Subsumption Matrix(SM), an appropriate fitness function is suggested. Finally, Quantified Hierarchical Production Rules (HPRs) are generated from the discovered hierarchy, using Dempster Shafer theory. Experimental results are presented to demonstrate the performance of the proposed algorithm.

Image Restoration in Non-Linear Filtering Domain using MDB approach

This paper proposes a new technique based on nonlinear Minmax Detector Based (MDB) filter for image restoration. The aim of image enhancement is to reconstruct the true image from the corrupted image. The process of image acquisition frequently leads to degradation and the quality of the digitized image becomes inferior to the original image. Image degradation can be due to the addition of different types of noise in the original image. Image noise can be modeled of many types and impulse noise is one of them. Impulse noise generates pixels with gray value not consistent with their local neighborhood. It appears as a sprinkle of both light and dark or only light spots in the image. Filtering is a technique for enhancing the image. Linear filter is the filtering in which the value of an output pixel is a linear combination of neighborhood values, which can produce blur in the image. Thus a variety of smoothing techniques have been developed that are non linear. Median filter is the one of the most popular non-linear filter. When considering a small neighborhood it is highly efficient but for large window and in case of high noise it gives rise to more blurring to image. The Centre Weighted Mean (CWM) filter has got a better average performance over the median filter. However the original pixel corrupted and noise reduction is substantial under high noise condition. Hence this technique has also blurring affect on the image. To illustrate the superiority of the proposed approach, the proposed new scheme has been simulated along with the standard ones and various restored performance measures have been compared.

Development of Accident Predictive Model for Rural Roadway

This paper present the study carried out of accident analysis, black spot study and to develop accident predictive models based on the data collected at rural roadway, Federal Route 50 (F050) Malaysia. The road accident trends and black spot ranking were established on the F050. The development of the accident prediction model will concentrate in Parit Raja area from KM 19 to KM 23. Multiple non-linear regression method was used to relate the discrete accident data with the road and traffic flow explanatory variable. The dependent variable was modeled as the number of crashes namely accident point weighting, however accident point weighting have rarely been account in the road accident prediction Models. The result show that, the existing number of major access points, without traffic light, rise in speed, increasing number of Annual Average Daily Traffic (AADT), growing number of motorcycle and motorcar and reducing the time gap are the potential contributors of increment accident rates on multiple rural roadway.

Constraint Based Frequent Pattern Mining Technique for Solving GCS Problem

Generalized Center String (GCS) problem are generalized from Common Approximate Substring problem and Common substring problems. GCS are known to be NP-hard allowing the problems lies in the explosion of potential candidates. Finding longest center string without concerning the sequence that may not contain any motifs is not known in advance in any particular biological gene process. GCS solved by frequent pattern-mining techniques and known to be fixed parameter tractable based on the fixed input sequence length and symbol set size. Efficient method known as Bpriori algorithms can solve GCS with reasonable time/space complexities. Bpriori 2 and Bpriori 3-2 algorithm are been proposed of any length and any positions of all their instances in input sequences. In this paper, we reduced the time/space complexity of Bpriori algorithm by Constrained Based Frequent Pattern mining (CBFP) technique which integrates the idea of Constraint Based Mining and FP-tree mining. CBFP mining technique solves the GCS problem works for all center string of any length, but also for the positions of all their mutated copies of input sequence. CBFP mining technique construct TRIE like with FP tree to represent the mutated copies of center string of any length, along with constraints to restraint growth of the consensus tree. The complexity analysis for Constrained Based FP mining technique and Bpriori algorithm is done based on the worst case and average case approach. Algorithm's correctness compared with the Bpriori algorithm using artificial data is shown.

Cyber Warriors for Cyber Security and Information Assurance- An Academic Perspective

A virtualized and virtual approach is presented on academically preparing students to successfully engage at a strategic perspective to understand those concerns and measures that are both structured and not structured in the area of cyber security and information assurance. The Master of Science in Cyber Security and Information Assurance (MSCSIA) is a professional degree for those who endeavor through technical and managerial measures to ensure the security, confidentiality, integrity, authenticity, control, availability and utility of the world-s computing and information systems infrastructure. The National University Cyber Security and Information Assurance program is offered as a Master-s degree. The emphasis of the MSCSIA program uniquely includes hands-on academic instruction using virtual computers. This past year, 2011, the NU facility has become fully operational using system architecture to provide a Virtual Education Laboratory (VEL) accessible to both onsite and online students. The first student cohort completed their MSCSIA training this past March 2, 2012 after fulfilling 12 courses, for a total of 54 units of college credits. The rapid pace scheduling of one course per month is immensely challenging, perpetually changing, and virtually multifaceted. This paper analyses these descriptive terms in consideration of those globalization penetration breaches as present in today-s world of cyber security. In addition, we present current NU practices to mitigate risks.

Determination of Severe Loading Condition at Critical System Cascading Collapse Considering the Effect of Protection System Hidden Failure

Hidden failure in a protection system has been recognized as one of the main reasons which may cause to a power system instability leading to a system cascading collapse. This paper presents a computationally systematic approach used to obtain the estimated average probability of a system cascading collapse by considering the effect of probability hidden failure in a protection system. The estimated average probability of a system cascading collapse is then used to determine the severe loading condition contributing to the higher risk of critical system cascading collapse. This information is essential to the system utility since it will assist the operator to determine the highest point of increased system loading condition prior to the event of critical system cascading collapse.