Mapping Knowledge Model Onto Java Codes

This paper gives an overview of the mapping mechanism of SEAM-a methodology for the automatic generation of knowledge models and its mapping onto Java codes. It discusses the rules that will be used to map the different components in the knowledge model automatically onto Java classes, properties and methods. The aim of developing this mechanism is to help in the creation of a prototype which will be used to validate the knowledge model which has been generated automatically. It will also help to link the modeling phase with the implementation phase as existing knowledge engineering methodologies do not provide for proper guidelines for the transition from the knowledge modeling phase to development phase. This will decrease the development overheads associated to the development of Knowledge Based Systems.

Use of Cultural Symbols for Transferring House to the Home in the Case of Famagusta

One of the essential requirements for the human beings is the house for living. This is necessary to make the place of satisfaction for contemporary houses residents by attention to their culture. In this article represented the relevant theoretical literature on cultural symbols by use the architecture semiotic to construct the houses as a better place for living. In fact, make a place for everyday life with changing the house to the home is one of the most challengeable subject for architects all around the world. The target of this article is to find Cypriot houses cultural symbols that assist architect to design and build contemporary houses, to make more satisfaction for its residents according to Cypriot life style and their culture. This paper is based on researching the effect of cultural symbols on housing, would require various types of methods. However, this study focuses on two methods, which are quantitative and qualitative. The purpose of the case-specific study is to finding the symbols that used in contemporary houses by attention to the Cypriot cultural symbols in Famagusta houses.

Integrating Fast Karnough Map and Modular Neural Networks for Simplification and Realization of Complex Boolean Functions

In this paper a new fast simplification method is presented. Such method realizes Karnough map with large number of variables. In order to accelerate the operation of the proposed method, a new approach for fast detection of group of ones is presented. Such approach implemented in the frequency domain. The search operation relies on performing cross correlation in the frequency domain rather than time one. It is proved mathematically and practically that the number of computation steps required for the presented method is less than that needed by conventional cross correlation. Simulation results using MATLAB confirm the theoretical computations. Furthermore, a powerful solution for realization of complex functions is given. The simplified functions are implemented by using a new desigen for neural networks. Neural networks are used because they are fault tolerance and as a result they can recognize signals even with noise or distortion. This is very useful for logic functions used in data and computer communications. Moreover, the implemented functions are realized with minimum amount of components. This is done by using modular neural nets (MNNs) that divide the input space into several homogenous regions. Such approach is applied to implement XOR function, 16 logic functions on one bit level, and 2-bit digital multiplier. Compared to previous non- modular designs, a clear reduction in the order of computations and hardware requirements is achieved.

Surface Roughness Optimization in End Milling Operation with Damper Inserted End Milling Cutters

This paper presents a study of the Taguchi design application to optimize surface quality in damper inserted end milling operation. Maintaining good surface quality usually involves additional manufacturing cost or loss of productivity. The Taguchi design is an efficient and effective experimental method in which a response variable can be optimized, given various factors, using fewer resources than a factorial design. This Study included spindle speed, feed rate, and depth of cut as control factors, usage of different tools in the same specification, which introduced tool condition and dimensional variability. An orthogonal array of L9(3^4)was used; ANOVA analyses were carried out to identify the significant factors affecting surface roughness, and the optimal cutting combination was determined by seeking the best surface roughness (response) and signal-to-noise ratio. Finally, confirmation tests verified that the Taguchi design was successful in optimizing milling parameters for surface roughness.

An Analysis of Variation of Ceiling Height and Window Level for Studio Architecture in Malaysia

This paper investigated the impact of ceiling height and window head heights variation on daylighting inside architectural teaching studio with a full width window. In architectural education, using the studio is more than normal classroom in most credit hours. Therefore, window position, size and dimension of studio have direct influence on level of daylighting. Daylighting design is a critical factor that improves student learning, concentration and behavior, in addition to these, it also reduces energy consumption. The methodology of analysis involves using Radiance in IES software under overcast and cloudy sky in Malaysia. It has been established that presentation of daylighting of architecture studio can be enhanced by changing the ceiling heights and window level, because, different ceiling heights and window head heights can contribute to different range of daylight levels.

Complexity Analysis of Some Known Graph Coloring Instances

Graph coloring is an important problem in computer science and many algorithms are known for obtaining reasonably good solutions in polynomial time. One method of comparing different algorithms is to test them on a set of standard graphs where the optimal solution is already known. This investigation analyzes a set of 50 well known graph coloring instances according to a set of complexity measures. These instances come from a variety of sources some representing actual applications of graph coloring (register allocation) and others (mycieleski and leighton graphs) that are theoretically designed to be difficult to solve. The size of the graphs ranged from ranged from a low of 11 variables to a high of 864 variables. The method used to solve the coloring problem was the square of the adjacency (i.e., correlation) matrix. The results show that the most difficult graphs to solve were the leighton and the queen graphs. Complexity measures such as density, mobility, deviation from uniform color class size and number of block diagonal zeros are calculated for each graph. The results showed that the most difficult problems have low mobility (in the range of .2-.5) and relatively little deviation from uniform color class size.

The Effect of Different Nozzle Configurations on Airflow Behaviour and Yarn Quality

Nozzle is the main part of various spinning systems such as air-jet and Murata air vortex systems. Recently, many researchers worked on the usage of the nozzle on different spinning systems such as conventional ring and compact spinning systems. In these applications, primary purpose is to improve the yarn quality. In present study, it was produced the yarns with two different nozzle types and determined the changes in yarn properties. In order to explain the effect of the nozzle, airflow structure in the nozzle was modelled and airflow variables were determined. In numerical simulation, ANSYS 12.1 package program and Fluid Flow (CFX) analysis method was used. As distinct from the literature, Shear Stress Turbulent (SST) model is preferred. And also air pressure at the nozzle inlet was measured by electronic mass flow meter and these values were used for the simulation of the airflow. At last, the yarn was modelled and the area from where the yarn is passing was included to the numerical analysis.

Exploring the Potential of Phase Change Memories as an Alternative to DRAM Technology

Scalability poses a severe threat to the existing DRAM technology. The capacitors that are used for storing and sensing charge in DRAM are generally not scaled beyond 42nm. This is because; the capacitors must be sufficiently large for reliable sensing and charge storage mechanism. This leaves DRAM memory scaling in jeopardy, as charge sensing and storage mechanisms become extremely difficult. In this paper we provide an overview of the potential and the possibilities of using Phase Change Memory (PCM) as an alternative for the existing DRAM technology. The main challenges that we encounter in using PCM are, the limited endurance, high access latencies, and higher dynamic energy consumption than that of the conventional DRAM. We then provide an overview of various methods, which can be employed to overcome these drawbacks. Hybrid memories involving both PCM and DRAM can be used, to achieve good tradeoffs in access latency and storage density. We conclude by presenting, the results of these methods that makes PCM a potential replacement for the current DRAM technology.

Joint Design of MIMO Relay Networks Based on MMSE Criterion

This paper deals with wireless relay communication systems in which multiple sources transmit information to the destination node by the help of multiple relays. We consider a signal forwarding technique based on the minimum mean-square error (MMSE) approach with multiple antennas for each relay. A source-relay-destination joint design strategy is proposed with power constraints at the destination and the source nodes. Simulation results confirm that the proposed joint design method improves the average MSE performance compared with that of conventional MMSE relaying schemes.

A Fast Replica Placement Methodology for Large-scale Distributed Computing Systems

Fine-grained data replication over the Internet allows duplication of frequently accessed data objects, as opposed to entire sites, to certain locations so as to improve the performance of largescale content distribution systems. In a distributed system, agents representing their sites try to maximize their own benefit since they are driven by different goals such as to minimize their communication costs, latency, etc. In this paper, we will use game theoretical techniques and in particular auctions to identify a bidding mechanism that encapsulates the selfishness of the agents, while having a controlling hand over them. In essence, the proposed game theory based mechanism is the study of what happens when independent agents act selfishly and how to control them to maximize the overall performance. A bidding mechanism asks how one can design systems so that agents- selfish behavior results in the desired system-wide goals. Experimental results reveal that this mechanism provides excellent solution quality, while maintaining fast execution time. The comparisons are recorded against some well known techniques such as greedy, branch and bound, game theoretical auctions and genetic algorithms.

Computational Networks for Knowledge Representation

In the artificial intelligence field, knowledge representation and reasoning are important areas for intelligent systems, especially knowledge base systems and expert systems. Knowledge representation Methods has an important role in designing the systems. There have been many models for knowledge such as semantic networks, conceptual graphs, and neural networks. These models are useful tools to design intelligent systems. However, they are not suitable to represent knowledge in the domains of reality applications. In this paper, new models for knowledge representation called computational networks will be presented. They have been used in designing some knowledge base systems in education for solving problems such as the system that supports studying knowledge and solving analytic geometry problems, the program for studying and solving problems in Plane Geometry, the program for solving problems about alternating current in physics.

A Method to Predict Hemorrhage Disease of Grass Carp Tends

Hemorrhage Disease of Grass Carp (HDGC) is a kind of commonly occurring illnesses in summer, and the extremely high death rate result in colossal losses to aquaculture. As the complex connections among each factor which influences aquiculture diseases, there-s no quit reasonable mathematical model to solve the problem at present.A BP neural network which with excellent nonlinear mapping coherence was adopted to establish mathematical model; Environmental factor, which can easily detected, such as breeding density, water temperature, pH and light intensity was set as the main analyzing object. 25 groups of experimental data were used for training and test, and the accuracy of using the model to predict the trend of HDGC was above 80%. It is demonstrated that BP neural network for predicating diseases in HDGC has a particularly objectivity and practicality, thus it can be spread to other aquiculture disease.

The Direct Updating of Damping and Gyroscopic Matrices using Incomplete Complex Test Data

In this paper we develop an efficient numerical method for the finite-element model updating of damped gyroscopic systems based on incomplete complex modal measured data. It is assumed that the analytical mass and stiffness matrices are correct and only the damping and gyroscopic matrices need to be updated. By solving a constrained optimization problem, the optimal corrected symmetric damping matrix and skew-symmetric gyroscopic matrix complied with the required eigenvalue equation are found under a weighted Frobenius norm sense.

A Renovated Cook's Distance Based On The Buckley-James Estimate In Censored Regression

There have been various methods created based on the regression ideas to resolve the problem of data set containing censored observations, i.e. the Buckley-James method, Miller-s method, Cox method, and Koul-Susarla-Van Ryzin estimators. Even though comparison studies show the Buckley-James method performs better than some other methods, it is still rarely used by researchers mainly because of the limited diagnostics analysis developed for the Buckley-James method thus far. Therefore, a diagnostic tool for the Buckley-James method is proposed in this paper. It is called the renovated Cook-s Distance, (RD* i ) and has been developed based on the Cook-s idea. The renovated Cook-s Distance (RD* i ) has advantages (depending on the analyst demand) over (i) the change in the fitted value for a single case, DFIT* i as it measures the influence of case i on all n fitted values Yˆ∗ (not just the fitted value for case i as DFIT* i) (ii) the change in the estimate of the coefficient when the ith case is deleted, DBETA* i since DBETA* i corresponds to the number of variables p so it is usually easier to look at a diagnostic measure such as RD* i since information from p variables can be considered simultaneously. Finally, an example using Stanford Heart Transplant data is provided to illustrate the proposed diagnostic tool.

Study on Specific Energy in Grinding of DRACs: A Response Surface Methodology Approach

In this study, the effects of machining parameters on specific energy during surface grinding of 6061Al-SiC35P composites are investigated. Vol% of SiC, feed and depth of cut were chosen as process variables. The power needed for the calculation of the specific energy is measured from the two watt meter method. Experiments are conducted using standard RSM design called Central composite design (CCD). A second order response surface model was developed for specific energy. The results identify the significant influence factors to minimize the specific energy. The confirmation results demonstrate the practicability and effectiveness of the proposed approach.

Increasing Profitability Supported by Innovative Methods and Designing Monitoring Software in Condition-Based Maintenance: A Case Study

In the present article, a new method has been developed to enhance the application of equipment monitoring, which in turn results in improving condition-based maintenance economic impact in an automobile parts manufacturing factory. This study also describes how an effective software with a simple database can be utilized to achieve cost-effective improvements in maintenance performance. The most important results of this project are indicated here: 1. 63% reduction in direct and indirect maintenance costs. 2. Creating a proper database to analyse failures. 3. Creating a method to control system performance and develop it to similar systems. 4. Designing a software to analyse database and consequently create technical knowledge to face unusual condition of the system. Moreover, the results of this study have shown that the concept and philosophy of maintenance has not been understood in most Iranian industries. Thus, more investment is strongly required to improve maintenance conditions.

A Combined Conventional and Differential Evolution Method for Model Order Reduction

In this paper a mixed method by combining an evolutionary and a conventional technique is proposed for reduction of Single Input Single Output (SISO) continuous systems into Reduced Order Model (ROM). In the conventional technique, the mixed advantages of Mihailov stability criterion and continued Fraction Expansions (CFE) technique is employed where the reduced denominator polynomial is derived using Mihailov stability criterion and the numerator is obtained by matching the quotients of the Cauer second form of Continued fraction expansions. Then, retaining the numerator polynomial, the denominator polynomial is recalculated by an evolutionary technique. In the evolutionary method, the recently proposed Differential Evolution (DE) optimization technique is employed. DE method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. The proposed method is illustrated through a numerical example and compared with ROM where both numerator and denominator polynomials are obtained by conventional method to show its superiority.

Intelligent Agent Approach to the Control of Critical Infrastructure Networks

In this paper we propose an intelligent agent approach to control the electric power grid at a smaller granularity in order to give it self-healing capabilities. We develop a method using the influence model to transform transmission substations into information processing, analyzing and decision making (intelligent behavior) units. We also develop a wireless communication method to deliver real-time uncorrupted information to an intelligent controller in a power system environment. A combined networking and information theoretic approach is adopted in meeting both the delay and error probability requirements. We use a mobile agent approach in optimizing the achievable information rate vector and in the distribution of rates to users (sensors). We developed the concept and the quantitative tools require in the creation of cooperating semiautonomous subsystems which puts the electric grid on the path towards intelligent and self-healing system.

Analysis of the Shielding Effectiveness of Several Magnetic Shields

Today with the rapid growth of telecommunications equipment, electronic and developing more and more networks of power, influence of electromagnetic waves on one another has become hot topic discussions. So in this article, this issue and appropriate mechanisms for EMC operations have been presented. First, a source of alternating current (50 Hz) and a clear victim in a certain distance from the source is placed. With this simple model, the effects of electromagnetic radiation from the source to the victim will be investigated and several methods to reduce these effects have been presented. Therefore passive and active shields have been used. In some steps, shielding effectiveness of proposed shields will be compared. . It should be noted that simulations have been done by the finite element method (FEM).

Social Interventation from Social Maternage to Peer Advocacy

The aim of this paper is to study in depth some methodological aspects of social interventation, focusing on desirable passage from social maternage method to peer advocacy method. For this purpose, we intend analyze social and organizative components, that affect operator-s professional action and that are part of his psychological environment, besides the physical and social one. In fact, operator-s interventation should not be limited to a pure supply of techniques, nor to take shape as improvised action, but “full of good purposes".