Abstract: Quality control charts are very effective in detecting
out of control signals but when a control chart signals an out of
control condition of the process mean, searching for a special cause
in the vicinity of the signal time would not always lead to prompt
identification of the source(s) of the out of control condition as the
change point in the process parameter(s) is usually different from the
signal time. It is very important to manufacturer to determine at what
point and which parameters in the past caused the signal. Early
warning of process change would expedite the search for the special
causes and enhance quality at lower cost. In this paper the quality
variables under investigation are assumed to follow a multivariate
normal distribution with known means and variance-covariance
matrix and the process means after one step change remain at the new
level until the special cause is being identified and removed, also it is
supposed that only one variable could be changed at the same time.
This research applies artificial neural network (ANN) to identify the
time the change occurred and the parameter which caused the change
or shift. The performance of the approach was assessed through a
computer simulation experiment. The results show that neural
network performs effectively and equally well for the whole shift
magnitude which has been considered.
Abstract: This paper is concerned with the delay-distributiondependent
stability criteria for bidirectional associative memory
(BAM) neural networks with time-varying delays. Based on the
Lyapunov-Krasovskii functional and stochastic analysis approach,
a delay-probability-distribution-dependent sufficient condition is derived
to achieve the globally asymptotically mean square stable of
the considered BAM neural networks. The criteria are formulated in
terms of a set of linear matrix inequalities (LMIs), which can be
checked efficiently by use of some standard numerical packages. Finally,
a numerical example and its simulation is given to demonstrate
the usefulness and effectiveness of the proposed results.
Abstract: A complex valued neural network is a neural network
which consists of complex valued input and/or weights and/or thresholds
and/or activation functions. Complex-valued neural networks
have been widening the scope of applications not only in electronics
and informatics, but also in social systems. One of the most important
applications of the complex valued neural network is in signal
processing. In Neural networks, generalized mean neuron model
(GMN) is often discussed and studied. The GMN includes a new
aggregation function based on the concept of generalized mean of all
the inputs to the neuron. This paper aims to present exhaustive results
of using Generalized Mean Neuron model in a complex-valued neural
network model that uses the back-propagation algorithm (called
-Complex-BP-) for learning. Our experiments results demonstrate the
effectiveness of a Generalized Mean Neuron Model in a complex
plane for signal processing over a real valued neural network. We
have studied and stated various observations like effect of learning
rates, ranges of the initial weights randomly selected, error functions
used and number of iterations for the convergence of error required on
a Generalized Mean neural network model. Some inherent properties
of this complex back propagation algorithm are also studied and
discussed.
Abstract: Overcurrent (OC) relays are the major protection
devices in a distribution system. The operating time of the OC relays
are to be coordinated properly to avoid the mal-operation of the
backup relays. The OC relay time coordination in ring fed
distribution networks is a highly constrained optimization problem
which can be stated as a linear programming problem (LPP). The
purpose is to find an optimum relay setting to minimize the time of
operation of relays and at the same time, to keep the relays properly
coordinated to avoid the mal-operation of relays.
This paper presents two phase simplex method for optimum time
coordination of OC relays. The method is based on the simplex
algorithm which is used to find optimum solution of LPP. The
method introduces artificial variables to get an initial basic feasible
solution (IBFS). Artificial variables are removed using iterative
process of first phase which minimizes the auxiliary objective
function. The second phase minimizes the original objective function
and gives the optimum time coordination of OC relays.
Abstract: The reduction in vehicle exhaust emissions achieved
in the last two decades is offset by the growth in traffic, as well as by
changes in the composition of emitted pollutants. The present
investigation illustrates the emissions of in-use gasoline and diesel
passenger cars using the official European driving cycle and the
ARTEMIS real-world driving cycle. It was observed that some of the
vehicles do not comply with the corresponding regulations.
Significant differences in emissions were observed between driving
cycles. Not all pollutants showed a tendency to decrease from Euro 3
to Euro 5.
Abstract: Localized surface plasmon resonance (LSPR) is the
coherent oscillation of conductive electrons confined in noble
metallic nanoparticles excited by electromagnetic radiation, and
nanosphere lithography (NSL) is one of the cost-effective methods to
fabricate metal nanostructures for LSPR. NSL can be categorized
into two major groups: dispersed NSL and closely pack NSL. In
recent years, gold nanocrescents and gold nanoholes with vertical
sidewalls fabricated by dispersed NSL, and silver nanotriangles and
gold nanocaps on silica nanospheres fabricated by closely pack NSL,
have been reported for LSPR biosensing. This paper introduces
several novel gold nanostructures fabricated by NSL in LSPR
applications, including 3D nanostructures obtained by evaporating
gold obliquely on dispersed nanospheres, nanoholes with slant
sidewalls, and patchy nanoparticles on closely packed nanospheres,
all of which render satisfactory sensitivity for LSPR sensing. Since
the LSPR spectrum is very sensitive to the shape of the metal
nanostructures, formulas are derived and software is developed for
calculating the profiles of the obtainable metal nanostructures by
NSL, for different nanosphere masks with different fabrication
conditions. The simulated profiles coincide well with the profiles of
the fabricated gold nanostructures observed under scanning electron
microscope (SEM) and atomic force microscope (AFM), which
proves that the software is a useful tool for the process design of
different LSPR nanostructures.
Abstract: Chua’s circuit is one of the most important electronic devices that are used for Chaos and Bifurcation studies. A central role of secure communication is devoted to it. Since the adaptive control is used vastly in the linear systems control, here we introduce a new trend of application of adaptive method in the chaos controlling field. In this paper, we try to derive a new adaptive control scheme for Chua’s circuit controlling because control of chaos is often very important in practical operations. The novelty of this approach is for sake of its robustness against the external perturbations which is simulated as an additive noise in all measured states and can be generalized to other chaotic systems. Our approach is based on Lyapunov analysis and the adaptation law is considered for the feedback gain. Because of this, we have named it NAFT (Nonlinear Adaptive Feedback Technique). At last, simulations show the capability of the presented technique for Chua’s circuit.
Abstract: Image convolution similar to the receptive fields
found in mammalian visual pathways has long been used in
conventional image processing in the form of Gabor masks.
However, no VLSI implementation of parallel, multi-layered pulsed
processing has been brought forward which would emulate this
property. We present a technical realization of such a pulsed image
processing scheme. The discussed IC also serves as a general testbed
for VLSI-based pulsed information processing, which is of interest
especially with regard to the robustness of representing an analog
signal in the phase or duration of a pulsed, quasi-digital signal, as
well as the possibility of direct digital manipulation of such an
analog signal. The network connectivity and processing properties
are reconfigurable so as to allow adaptation to various processing
tasks.
Abstract: Importance of software quality is increasing leading to development of new sophisticated techniques, which can be used in constructing models for predicting quality attributes. One such technique is Artificial Neural Network (ANN). This paper examined the application of ANN for software quality prediction using Object- Oriented (OO) metrics. Quality estimation includes estimating maintainability of software. The dependent variable in our study was maintenance effort. The independent variables were principal components of eight OO metrics. The results showed that the Mean Absolute Relative Error (MARE) was 0.265 of ANN model. Thus we found that ANN method was useful in constructing software quality model.
Abstract: Independent component analysis (ICA) in the
frequency domain is used for solving the problem of blind source
separation (BSS). However, this method has some problems. For
example, a general ICA algorithm cannot determine the permutation
of signals which is important in the frequency domain ICA. In this
paper, we propose an approach to the solution for a permutation
problem. The idea is to effectively combine two conventional
approaches. This approach improves the signal separation
performance by exploiting features of the conventional approaches.
We show the simulation results using artificial data.
Abstract: It is necessary to evaluate the bridges conditions and
strengthen bridges or parts of them. The reinforcement necessary due
to some reasons can be summarized as: First, a changing in use of
bridge could produce internal forces in a part of structural which
exceed the existing cross-sectional capacity. Second, bridges may
also need reinforcement because damage due to external factors
which reduced the cross-sectional resistance to external loads. One of
other factors could listed here its misdesign in some details, like
safety of bridge or part of its.This article identify the design demands
of Qing Shan bridge located in is in Heilongjiang Province He gang -
Nen Jiang Road 303 provincial highway, Wudalianchi area, China, is
an important bridge in the urban areas. The investigation program
was include the observation and evaluate the damage in T- section
concrete beams , prestressed concrete box girder bridges section in
additional evaluate the whole state of bridge includes the pier ,
abutments , bridge decks, wings , bearing and capping beam, joints,
........etc. The test results show that the bridges in general structural
condition are good. T beam span No 10 were observed, crack
extended upward along the ribbed T beam, and continue to the T
beam flange. Crack width varying between 0.1mm to 0.4mm, the
maximum about 0.4mm. The bridge needs to be improved flexural
bending strength especially at for T beam section.
Abstract: In this paper, to optimize the “Characteristic Straight Line Method" which is used in the soil displacement analysis, a “best estimate" of the geodetic leveling observations has been achieved by taking in account the concept of 'Height systems'. This concept has been discussed in detail and consequently the concept of “height". In landslides dynamic analysis, the soil is considered as a mosaic of rigid blocks. The soil displacement has been monitored and analyzed by using the “Characteristic Straight Line Method". Its characteristic components have been defined constructed from a “best estimate" of the topometric observations. In the measurement of elevation differences, we have used the most modern leveling equipment available. Observational procedures have also been designed to provide the most effective method to acquire data. In addition systematic errors which cannot be sufficiently controlled by instrumentation or observational techniques are minimized by applying appropriate corrections to the observed data: the level collimation correction minimizes the error caused by nonhorizontality of the leveling instrument's line of sight for unequal sight lengths, the refraction correction is modeled to minimize the refraction error caused by temperature (density) variation of air strata, the rod temperature correction accounts for variation in the length of the leveling rod' s Invar/LO-VAR® strip which results from temperature changes, the rod scale correction ensures a uniform scale which conforms to the international length standard and the introduction of the concept of the 'Height systems' where all types of height (orthometric, dynamic, normal, gravity correction, and equipotential surface) have been investigated. The “Characteristic Straight Line Method" is slightly more convenient than the “Characteristic Circle Method". It permits to evaluate a displacement of very small magnitude even when the displacement is of an infinitesimal quantity. The inclination of the landslide is given by the inverse of the distance reference point O to the “Characteristic Straight Line". Its direction is given by the bearing of the normal directed from point O to the Characteristic Straight Line (Fig..6). A “best estimate" of the topometric observations was used to measure the elevation of points carefully selected, before and after the deformation. Gross errors have been eliminated by statistical analyses and by comparing the heights within local neighborhoods. The results of a test using an area where very interesting land surface deformation occurs are reported. Monitoring with different options and qualitative comparison of results based on a sufficient number of check points are presented.
Abstract: The problem of estimating time-varying regression is
inevitably concerned with the necessity to choose the appropriate
level of model volatility - ranging from the full stationarity of instant
regression models to their absolute independence of each other. In the
stationary case the number of regression coefficients to be estimated
equals that of regressors, whereas the absence of any smoothness
assumptions augments the dimension of the unknown vector by the
factor of the time-series length. The Akaike Information Criterion
is a commonly adopted means of adjusting a model to the given
data set within a succession of nested parametric model classes,
but its crucial restriction is that the classes are rigidly defined by
the growing integer-valued dimension of the unknown vector. To
make the Kullback information maximization principle underlying the
classical AIC applicable to the problem of time-varying regression
estimation, we extend it onto a wider class of data models in which
the dimension of the parameter is fixed, but the freedom of its values
is softly constrained by a family of continuously nested a priori
probability distributions.
Abstract: The purpose of this paper is to perform a multidisciplinary design and analysis (MDA) of honeycomb panels used in the satellites structural design. All the analysis is based on clamped-free boundary conditions. In the present work, detailed finite element models for honeycomb panels are developed and analysed. Experimental tests were carried out on a honeycomb specimen of which the goal is to compare the previous modal analysis made by the finite element method as well as the existing equivalent approaches. The obtained results show a good agreement between the finite element analysis, equivalent and tests results; the difference in the first two frequencies is less than 4% and less than 10% for the third frequency. The results of the equivalent model presented in this analysis are obtained with a good accuracy. Moreover, investigations carried out in this research relate to the honeycomb plate modal analysis under several aspects including the structural geometrical variation by studying the various influences of the dimension parameters on the modal frequency, the variation of core and skin material of the honeycomb. The various results obtained in this paper are promising and show that the geometry parameters and the type of material have an effect on the value of the honeycomb plate modal frequency.
Abstract: Automatic segmentation of skin lesions is the first step
towards development of a computer-aided diagnosis of melanoma.
Although numerous segmentation methods have been developed,
few studies have focused on determining the most discriminative
and effective color space for melanoma application. This paper
proposes a novel automatic segmentation algorithm using color space
analysis and clustering-based histogram thresholding, which is able to
determine the optimal color channel for segmentation of skin lesions.
To demonstrate the validity of the algorithm, it is tested on a set of 30
high resolution dermoscopy images and a comprehensive evaluation
of the results is provided, where borders manually drawn by four
dermatologists, are compared to automated borders detected by the
proposed algorithm. The evaluation is carried out by applying three
previously used metrics of accuracy, sensitivity, and specificity and
a new metric of similarity. Through ROC analysis and ranking the
metrics, it is shown that the best results are obtained with the X and
XoYoR color channels which results in an accuracy of approximately
97%. The proposed method is also compared with two state-ofthe-
art skin lesion segmentation methods, which demonstrates the
effectiveness and superiority of the proposed segmentation method.
Abstract: A filter is used to remove undesirable frequency information from a dynamic signal. This paper shows that the Znotch filter filtering technique can be applied to remove the noise nuisance from a machining signal. In machining, the noise components were identified from the sound produced by the operation of machine components itself such as hydraulic system, motor, machine environment and etc. By correlating the noise components with the measured machining signal, the interested components of the measured machining signal which was less interfered by the noise, can be extracted. Thus, the filtered signal is more reliable to be analysed in terms of noise content compared to the unfiltered signal. Significantly, the I-kaz method i.e. comprises of three dimensional graphical representation and I-kaz coefficient, Z∞ could differentiate between the filtered and the unfiltered signal. The bigger space of scattering and the higher value of Z∞ demonstrated that the signal was highly interrupted by noise. This method can be utilised as a proactive tool in evaluating the noise content in a signal. The evaluation of noise content is very important as well as the elimination especially for machining operation fault diagnosis purpose. The Z-notch filtering technique was reliable in extracting noise component from the measured machining signal with high efficiency. Even though the measured signal was exposed to high noise disruption, the signal generated from the interaction between cutting tool and work piece still can be acquired. Therefore, the interruption of noise that could change the original signal feature and consequently can deteriorate the useful sensory information can be eliminated.
Abstract: Several models have been introduced so far for single
electron box, SEB, which all of them were restricted to DC response
and or low temperature limit. In this paper we introduce a new time
dependent, high temperature analytical model for SEB for the first
time. DC behavior of the introduced model will be verified against
SIMON software and its time behavior will be verified against a
newly published paper regarding step response of SEB.
Abstract: This paper presents a useful sub-pixel image
registration method using line segments and a sub-pixel edge detector.
In this approach, straight line segments are first extracted from gray
images at the pixel level before applying the sub-pixel edge detector.
Next, all sub-pixel line edges are mapped onto the orientation-distance
parameter space to solve for line correspondence between images.
Finally, the registration parameters with sub-pixel accuracy are
analytically solved via two linear least-square problems. The present
approach can be applied to various fields where fast registration with
sub-pixel accuracy is required. To illustrate, the present approach is
applied to the inspection of printed circuits on a flat panel. Numerical
example shows that the present approach is effective and accurate
when target images contain a sufficient number of line segments,
which is true in many industrial problems.
Abstract: The paper is concerned with relationships between
SSME and ICTs and focuses on the role of Web 2.0 tools in
the service development process. The research presented aims at
exploring how collaborative technologies can support and improve
service processes, highlighting customer centrality and value coproduction.
The core idea of the paper is the centrality of user
participation and the collaborative technologies as enabling factors;
Wikipedia is analyzed as an example. The result of such analysis is
the identification and description of a pattern characterising specific
services in which users collaborate by means of web tools with value
co-producers during the service process. The pattern of collaborative
co-production concerning several categories of services including
knowledge based services is then discussed.
Abstract: Multiprocessor task scheduling is a NP-hard problem and Genetic Algorithm (GA) has been revealed as an excellent technique for finding an optimal solution. In the past, several methods have been considered for the solution of this problem based on GAs. But, all these methods consider single criteria and in the present work, minimization of the bi-criteria multiprocessor task scheduling problem has been considered which includes weighted sum of makespan & total completion time. Efficiency and effectiveness of genetic algorithm can be achieved by optimization of its different parameters such as crossover, mutation, crossover probability, selection function etc. The effects of GA parameters on minimization of bi-criteria fitness function and subsequent setting of parameters have been accomplished by central composite design (CCD) approach of response surface methodology (RSM) of Design of Experiments. The experiments have been performed with different levels of GA parameters and analysis of variance has been performed for significant parameters for minimisation of makespan and total completion time simultaneously.