Adaptive Network Intrusion Detection Learning: Attribute Selection and Classification

In this paper, a new learning approach for network intrusion detection using naïve Bayesian classifier and ID3 algorithm is presented, which identifies effective attributes from the training dataset, calculates the conditional probabilities for the best attribute values, and then correctly classifies all the examples of training and testing dataset. Most of the current intrusion detection datasets are dynamic, complex and contain large number of attributes. Some of the attributes may be redundant or contribute little for detection making. It has been successfully tested that significant attribute selection is important to design a real world intrusion detection systems (IDS). The purpose of this study is to identify effective attributes from the training dataset to build a classifier for network intrusion detection using data mining algorithms. The experimental results on KDD99 benchmark intrusion detection dataset demonstrate that this new approach achieves high classification rates and reduce false positives using limited computational resources.

The Story of Mergers and Acquisitions: Using Narrative Theory to Understand the Uncertainty of Organizational Change

This paper examines the influence of communication form on employee uncertainty during mergers and acquisitions (M&As). Specifically, the author uses narrative theory to analyze how narrative organizational communication affects the three components of uncertainty – decreased predictive, explanatory, and descriptive ability. It is hypothesized that employees whose organizations use narrative M&A communication will have greater predictive, explanatory, and descriptive abilities than employees of organizations using non-narrative M&A communication. This paper contributes to the stream of research examining uncertainty during mergers and acquisitions and argues that narratives are an effective means of managing uncertainty in the mergers and acquisitions context.

Agreement Options in Multi-person Decision on Optimizing High-Rise Building Columns

This paper presents a conceptual model of agreement options for negotiation support in multi-person decision on optimizing high-rise building columns. The decision is complicated since many parties involved in choosing a single alternative from a set of solutions. There are different concern caused by differing preferences, experiences, and background. Such building columns as alternatives are referred to as agreement options which are determined by identifying the possible decision maker group, followed by determining the optimal solution for each group. The group in this paper is based on three-decision makers preferences that are designer, programmer, and construction manager. Decision techniques applied to determine the relative value of the alternative solutions for performing the function. Analytical Hierarchy Process (AHP) was applied for decision process and game theory based agent system for coalition formation. An n-person cooperative game is represented by the set of all players. The proposed coalition formation model enables each agent to select individually its allies or coalition. It further emphasizes the importance of performance evaluation in the design process and value-based decision.

A New Heuristic Statistical Methodology for Optimizing Queuing Networks Using Discreet Event Simulation

Most of the real queuing systems include special properties and constraints, which can not be analyzed directly by using the results of solved classical queuing models. Lack of Markov chains features, unexponential patterns and service constraints, are the mentioned conditions. This paper represents an applied general algorithm for analysis and optimizing the queuing systems. The algorithm stages are described through a real case study. It is consisted of an almost completed non-Markov system with limited number of customers and capacities as well as lots of common exception of real queuing networks. Simulation is used for optimizing this system. So introduced stages over the following article include primary modeling, determining queuing system kinds, index defining, statistical analysis and goodness of fit test, validation of model and optimizing methods of system with simulation.

Implementation of Technology Concept for the Reduction of Cyanobacteria in Laboratory

Following the research in the Department of environmental engineering in Faculty of mechanical engineering on Technical University of Kosice and experiences with electrocoagulation style of disposal waste water, there were designed and partly examining the equipment of two stage revitalization on the standing and little fusible water of tenet electrolysis on the little tarns. With the cooperation with vet experts was that manners prove and it is innocuous for animals, during which time cyanobacteria are totally paralyzed. For the implementation of science and research results have been obtained by means EU funds for structural development.

Topology Preservation in SOM

The SOM has several beneficial features which make it a useful method for data mining. One of the most important features is the ability to preserve the topology in the projection. There are several measures that can be used to quantify the goodness of the map in order to obtain the optimal projection, including the average quantization error and many topological errors. Many researches have studied how the topology preservation should be measured. One option consists of using the topographic error which considers the ratio of data vectors for which the first and second best BMUs are not adjacent. In this work we present a study of the behaviour of the topographic error in different kinds of maps. We have found that this error devaluates the rectangular maps and we have studied the reasons why this happens. Finally, we suggest a new topological error to improve the deficiency of the topographic error.

Bottom Up Text Mining through Hierarchical Document Representation

Most of the existing text mining approaches are proposed, keeping in mind, transaction databases model. Thus, the mined dataset is structured using just one concept: the “transaction", whereas the whole dataset is modeled using the “set" abstract type. In such cases, the structure of the whole dataset and the relationships among the transactions themselves are not modeled and consequently, not considered in the mining process. We believe that taking into account structure properties of hierarchically structured information (e.g. textual document, etc ...) in the mining process, can leads to best results. For this purpose, an hierarchical associations rule mining approach for textual documents is proposed in this paper and the classical set-oriented mining approach is reconsidered profits to a Direct Acyclic Graph (DAG) oriented approach. Natural languages processing techniques are used in order to obtain the DAG structure. Based on this graph model, an hierarchical bottom up algorithm is proposed. The main idea is that each node is mined with its parent node.

Mining Frequent Patterns with Functional Programming

Frequent patterns are patterns such as sets of features or items that appear in data frequently. Finding such frequent patterns has become an important data mining task because it reveals associations, correlations, and many other interesting relationships hidden in a dataset. Most of the proposed frequent pattern mining algorithms have been implemented with imperative programming languages such as C, Cµ, Java. The imperative paradigm is significantly inefficient when itemset is large and the frequent pattern is long. We suggest a high-level declarative style of programming using a functional language. Our supposition is that the problem of frequent pattern discovery can be efficiently and concisely implemented via a functional paradigm since pattern matching is a fundamental feature supported by most functional languages. Our frequent pattern mining implementation using the Haskell language confirms our hypothesis about conciseness of the program. The performance studies on speed and memory usage support our intuition on efficiency of functional language.

A Framework of Monte Carlo Simulation for Examining the Uncertainty-Investment Relationship

This paper argues that increased uncertainty, in certain situations, may actually encourage investment. Since earlier studies mostly base their arguments on the assumption of geometric Brownian motion, the study extends the assumption to alternative stochastic processes, such as mixed diffusion-jump, mean-reverting process, and jump amplitude process. A general approach of Monte Carlo simulation is developed to derive optimal investment trigger for the situation that the closed-form solution could not be readily obtained under the assumption of alternative process. The main finding is that the overall effect of uncertainty on investment is interpreted by the probability of investing, and the relationship appears to be an invested U-shaped curve between uncertainty and investment. The implication is that uncertainty does not always discourage investment even under several sources of uncertainty. Furthermore, high-risk projects are not always dominated by low-risk projects because the high-risk projects may have a positive realization effect on encouraging investment.

Modeling the Effects of Type and Intensity of Selective Logging on Forests of the Amazon

The aim of the work presented here was to either use existing forest dynamic simulation models or calibrate a new one both within the SYMFOR framework with the purpose of examining changes in stand level basal area and functional composition in response to selective logging considering trees > 10 cm d.b.h for two areas of undisturbed Amazonian non flooded tropical forest in Brazil and one in Peru. Model biological realism was evaluated for forest in the undisturbed and selectively logged state and it was concluded that forest dynamics were realistically represented. Results of the logging simulation experiments showed that in relation to undisturbed forest simulation subject to no form of harvesting intervention there was a significant amount of change over a 90 year simulation period that was positively proportional to the intensity of logging. Areas which had in the dynamic equilibrium of undisturbed forest a greater proportion of a specific ecological guild of trees known as the light hardwoods (LHW’s) seemed to respond more favorably in terms of less deviation but only within a specific range of baseline forest composition beyond which compositional diversity became more important. These finds are in line partially with practical management experience and partiality basic systematics theory respectively.

Method for Determining the Probing Points for Efficient Measurement of Freeform Surface

In inspection and workpiece localization, sampling point data is an important issue. Since the devices for sampling only sample discrete points, not the completely surface, sampling size and location of the points will be taken into consideration. In this paper a method is presented for determining the sampled points size and location for achieving efficient sampling. Firstly, uncertainty analysis of the localization parameters is investigated. A localization uncertainty model is developed to predict the uncertainty of the localization process. Using this model the minimum size of the sampled points is predicted. Secondly, based on the algebra theory an eigenvalue-optimal optimization is proposed. Then a freeform surface is used in the simulation. The proposed optimization is implemented. The simulation result shows its effectivity.

2D Numerical Analysis of Sao Paulo Tunnel

Nonlinear finite element method and Serendipity eight nodes element are used for determining of ground surface settlement due to tunneling. Linear element with elastic behavior is used for modeling of lining. Modified Generalized plasticity model with nonassociated flow rule is applied for analysis of a tunnel in Sao Paulo – Brazil. The tunnel had analyzed by Lades- model with 16 parameters. In this work modified Generalized Plasticity is used with 10 parameters, also Mohr-Coulomb model is used to analysis the tunnel. The results show good agreement with observed results of field data by modified Generalized Plasticity model than other models. The obtained result by Mohr-Coulomb model shows less settlement than other model due to excavation.

Surface Roughness and MRR Effect on Manual Plasma Arc Cutting Machining

Industrial surveys shows that manufacturing companies define the qualities of thermal removing process based on the dimension and physical appearance of the cutting material surface. Therefore, the roughness of the surface area of the material cut by the plasma arc cutting process and the rate of the removed material by the manual plasma arc cutting machine was importantly considered. Plasma arc cutter Selco Genesis 90 was used to cut Standard AISI 1017 Steel of 200 mm x100 mm x 6 mm manually based on the selected parameters setting. The material removal rate (MRR) was measured by determining the weight of the specimens before and after the cutting process. The surface roughness (SR) analysis was conducted using Mitutoyo CS-3100 to determine the average roughness value (Ra). Taguchi method was utilized to achieve optimum condition for both outputs studied. The microstructure analysis in the region of the cutting surface is performed using SEM. The results reveal that the SR values are inversely proportional to the MRR values. The quality of the surface roughness depends on the dross peak that occurred after the cutting process.

Qualitative Modelling for Ferromagnetic Hysteresis Cycle

In determining the electromagnetic properties of magnetic materials, hysteresis modeling is of high importance. Many models are available to investigate those characteristics but they tend to be complex and difficult to implement. In this paper a new qualitative hysteresis model for ferromagnetic core is presented, based on the function approximation capabilities of adaptive neuro fuzzy inference system (ANFIS). The proposed ANFIS model combined the neural network adaptive capabilities and the fuzzy logic qualitative approach can restored the hysteresis curve with a little RMS error. The model accuracy is good and can be easily adapted to the requirements of the application by extending or reducing the network training set and thus the required amount of measurement data.

A Web Text Mining Flexible Architecture

Text Mining is an important step of Knowledge Discovery process. It is used to extract hidden information from notstructured o semi-structured data. This aspect is fundamental because much of the Web information is semi-structured due to the nested structure of HTML code, much of the Web information is linked, much of the Web information is redundant. Web Text Mining helps whole knowledge mining process to mining, extraction and integration of useful data, information and knowledge from Web page contents. In this paper, we present a Web Text Mining process able to discover knowledge in a distributed and heterogeneous multiorganization environment. The Web Text Mining process is based on flexible architecture and is implemented by four steps able to examine web content and to extract useful hidden information through mining techniques. Our Web Text Mining prototype starts from the recovery of Web job offers in which, through a Text Mining process, useful information for fast classification of the same are drawn out, these information are, essentially, job offer place and skills.

Data Mining Using Learning Automata

In this paper a data miner based on the learning automata is proposed and is called LA-miner. The LA-miner extracts classification rules from data sets automatically. The proposed algorithm is established based on the function optimization using learning automata. The experimental results on three benchmarks indicate that the performance of the proposed LA-miner is comparable with (sometimes better than) the Ant-miner (a data miner algorithm based on the Ant Colony optimization algorithm) and CNZ (a well-known data mining algorithm for classification).

Electromagnetic Flow Meter Efficiency

A study of electromagnetic flow meter is presented in the paper. Comparison has been made between the analytical and the numerical results by the use of FEM numerical analysis (Quick Field 5.6) for determining polarization voltage through the circle cross section of the polarization transducer. Exciting and geometrical parameters increasing its effectiveness has been examined. The aim is to obtain maximal output signal. The investigations include different variants of the magnetic flux density distribution around the tube: homogeneous field of magnitude Bm, linear distribution with maximal value Bm and trapezium distribution conserving the same exciting magnetic energy as the homogeneous field.

A Framework for Data Mining Based Multi-Agent: An Application to Spatial Data

Data mining is an extraordinarily demanding field referring to extraction of implicit knowledge and relationships, which are not explicitly stored in databases. A wide variety of methods of data mining have been introduced (classification, characterization, generalization...). Each one of these methods includes more than algorithm. A system of data mining implies different user categories,, which mean that the user-s behavior must be a component of the system. The problem at this level is to know which algorithm of which method to employ for an exploratory end, which one for a decisional end, and how can they collaborate and communicate. Agent paradigm presents a new way of conception and realizing of data mining system. The purpose is to combine different algorithms of data mining to prepare elements for decision-makers, benefiting from the possibilities offered by the multi-agent systems. In this paper the agent framework for data mining is introduced, and its overall architecture and functionality are presented. The validation is made on spatial data. Principal results will be presented.

Predicting the Life Cycle of Complex Technical Systems (CTS)

Complex systems are composed of several plain interacting independent entities. Interaction between these entities creates a unified behavior at the global level that cannot be predicted by examining the behavior of any single individual component of the system. In this paper we consider a welded frame of an automobile trailer as a real example of Complex Technical Systems, The purpose of this paper is to introduce a Statistical method for predicting the life cycle of complex technical systems. To organize gathering of primary data for modeling the life cycle of complex technical systems an “Automobile Trailer Frame" were used as a prototype in this research. The prototype represents a welded structure of several pieces. Both information flows underwent a computerized analysis and classification for the acquisition of final results to reach final recommendations for improving the trailers structure and their operational conditions.

Large-Dimensional Shells under Mining Tremors from Various Mining Regions in Poland

In the paper a detailed analysis of the dynamic response of a cooling tower shell to mining tremors originated from two main regions of mining activity in Poland (Upper Silesian Coal Basin and Legnica-Glogow Copper District) was presented. The representative time histories registered in the both regions were used as ground motion data in calculations of the dynamic response of the structure. It was proved that the dynamic response of the shell is strongly dependent not only on the level of vibration amplitudes but on the dominant frequency range of the mining shock typical for the mining region as well. Also a vertical component of vibrations occurred to have considerable influence on the total dynamic response of the shell. Finally, it turned out that non-uniformity of kinematic excitation resulting from spatial variety of ground motion plays a significant role in dynamic analysis of large-dimensional shells under mining shocks.