Abstract: This paper presents Cost per Equivalent Wafer Out, which we find useful in wafer fab operational cost monitoring and controlling. It removes the loading and product mix effect in the cost variance analysis. The operation heads, therefore, could immediately focus on identifying areas for cost improvement. Without this, they would have to measure the impact of the loading variance and product mix variance between actual and budgeted prior to make any decision on cost improvement. Cost per Equivalent Wafer Out, thereby, increases efficiency in wafer fab operational cost monitoring and controlling.
Abstract: Aligned and random nanofibrous scaffolds of PVA/PCL/nHA were fabricated by electrospinning method. The composite nanofibrous scaffolds were subjected to detailed analysis. Morphological investigations revealed that the prepared nanofibers have uniform morphology and the average fiber diameters of aligned and random scaffolds were 135.5 and 290 nm, respectively. The obtained scaffolds have a porous structure with porosity of 88 and 76% for random and aligned nanofibers, respectively. Furthermore, FTIR analysis demonstrated that there were strong intramolecular interactions between the molecules of PVA/PCL/nHA. On the other hand, mechanical characterizations show that aligning the nanofibers, could significantly improve the rigidity of the resultant biocomposite nanofibrous scaffolds.
Abstract: Urban road network traffic has become one of the
most studied research topics in the last decades. This is mainly due to
the enlargement of the cities and the growing number of motor
vehicles traveling in this road network. One of the most sensitive
problems is to verify if the network is congestion-free. Another
related problem is the automatic reconfiguration of the network
without building new roads to alleviate congestions. These problems
require an accurate model of the traffic to determine the steady state
of the system. An alternative is to simulate the traffic to see if there
are congestions and when and where they occur. One key issue is to
find an adequate model for road intersections. Once the model
established, either a large scale model is built or the intersection is
represented by its performance measures and simulation for analysis.
In both cases, it is important to seek the queueing model to represent
the road intersection. In this paper, we propose to model the road
intersection as a BCMP queueing network and we compare this
analytical model against a simulation model for validation.
Abstract: The Integrated Performance Modelling Environment
(IPME) is a powerful simulation engine for task simulation and
performance analysis. However, it has no high level cognition such
as memory and reasoning for complex simulation. This article
introduces a knowledge representation and reasoning scheme that can
accommodate uncertainty in simulations of military personnel with
IPME. This approach demonstrates how advanced reasoning models
that support similarity-based associative process, rule-based abstract
process, multiple reasoning methods and real-time interaction can be
integrated with conventional task network modelling to provide
greater functionality and flexibility when modelling operator
performance.
Abstract: The counting and analysis of blood cells allows the
evaluation and diagnosis of a vast number of diseases. In particular,
the analysis of white blood cells (WBCs) is a topic of great interest to
hematologists. Nowadays the morphological analysis of blood cells is
performed manually by skilled operators. This involves numerous
drawbacks, such as slowness of the analysis and a nonstandard
accuracy, dependent on the operator skills. In literature there are only
few examples of automated systems in order to analyze the white
blood cells, most of which only partial. This paper presents a
complete and fully automatic method for white blood cells
identification from microscopic images. The proposed method firstly
individuates white blood cells from which, subsequently, nucleus and
cytoplasm are extracted. The whole work has been developed using
MATLAB environment, in particular the Image Processing Toolbox.
Abstract: Background: This investigated the mental health of the elderly survivors six months, ten months and two years after the “5.12 Wenchuan" earthquake. Methods: Two hundred and thirty-two physically healthy older survivors from earthquake-affected Mianyang County were interviewed. The measures included the Revised Impact of Event Scale (IES-R, Chinese version, for PTSD) and a Chinese Mental Health Inventory for the Elderly (MHIE). A repeated measures ANOVA test was used for statistical analysis. Results: The follow-up group had a statistically significant lower IES-R score and lower MHIE score than the initial group ten months after the earthquake. Two years later, the score of IES-R in follow-up group were still lower than that of non-follow-up group, but no differences were significant on the score of MHIE between groups. Furthermore, a negative relationship was found between scores of IES-R and MHIE. Conclusion: The earthquake has had a persistent negative impact on older survivors- mental health within the two-year period and that although the PTSD level declined significantly with time, it did not disappear completely.
Abstract: Deformable active contours are widely used in
computer vision and image processing applications for image
segmentation, especially in biomedical image analysis. The active
contour or “snake" deforms towards a target object by controlling the
internal, image and constraint forces. However, if the contour
initialized with a lesser number of control points, there is a high
probability of surpassing the sharp corners of the object during
deformation of the contour. In this paper, a new technique is
proposed to construct the initial contour by incorporating prior
knowledge of significant corners of the object detected using the
Harris operator. This new reconstructed contour begins to deform, by
attracting the snake towards the targeted object, without missing the
corners. Experimental results with several synthetic images show the
ability of the new technique to deal with sharp corners with a high
accuracy than traditional methods.
Abstract: Integration of system process information obtained
through an image processing system with an evolving knowledge
database to improve the accuracy and predictability of wear particle
analysis is the main focus of the paper. The objective is to automate
intelligently the analysis process of wear particle using classification
via self organizing maps. This is achieved using relationship
measurements among corresponding attributes of various
measurements for wear particle. Finally, visualization technique is
proposed that helps the viewer in understanding and utilizing these
relationships that enable accurate diagnostics.
Abstract: In the paper, the results of sensitivity analysis of the influence of initial imperfections on the web stress state of a thinwalled girder are presented. The results of the study corroborate a very good and effective agreement of experiments with theory. Most input random quantities were found experimentally. The change of sensitivity coefficients in dependence on working load value is analysed. The stress was analysed by means of a geometrically and materially non-linear solution by applying the program ANSYS. This research study offers important background for theoretical studies of stability problems, post-critical effects and limit states of thin-walled steel structures.
Abstract: In the last few years, three multivariate spectral
analysis techniques namely, Principal Component Analysis (PCA),
Independent Component Analysis (ICA) and Non-negative Matrix
Factorization (NMF) have emerged as effective tools for oscillation
detection and isolation. While the first method is used in determining
the number of oscillatory sources, the latter two methods
are used to identify source signatures by formulating the detection
problem as a source identification problem in the spectral domain.
In this paper, we present a critical drawback of the underlying linear
(mixing) model which strongly limits the ability of the associated
source separation methods to determine the number of sources
and/or identify the physical source signatures. It is shown that the
assumed mixing model is only valid if each unit of the process gives
equal weighting (all-pass filter) to all oscillatory components in its
inputs. This is in contrast to the fact that each unit, in general, acts
as a filter with non-uniform frequency response. Thus, the model
can only facilitate correct identification of a source with a single
frequency component, which is again unrealistic. To overcome
this deficiency, an iterative post-processing algorithm that correctly
identifies the physical source(s) is developed. An additional issue
with the existing methods is that they lack a procedure to pre-screen
non-oscillatory/noisy measurements which obscure the identification
of oscillatory sources. In this regard, a pre-screening procedure
is prescribed based on the notion of sparseness index to eliminate
the noisy and non-oscillatory measurements from the data set used
for analysis.