Abstract: A novel file splitting technique for the reduction of the nth-order entropy of text files is proposed. The technique is based on mapping the original text file into a non-ASCII binary file using a new codeword assignment method and then the resulting binary file is split into several subfiles each contains one or more bits from each codeword of the mapped binary file. The statistical properties of the subfiles are studied and it is found that they reflect the statistical properties of the original text file which is not the case when the ASCII code is used as a mapper. The nth-order entropy of these subfiles are determined and it is found that the sum of their entropies is less than that of the original text file for the same values of extensions. These interesting statistical properties of the resulting subfiles can be used to achieve better compression ratios when conventional compression techniques are applied to these subfiles individually and on a bit-wise basis rather than on character-wise basis.
Abstract: This paper proposes a novel architecture for developing decision support systems. Unlike conventional decision support systems, the proposed architecture endeavors to reveal the decision-making process such that humans' subjectivity can be incorporated into a computerized system and, at the same time, to preserve the capability of the computerized system in processing information objectively. A number of techniques used in developing the decision support system are elaborated to make the decisionmarking process transparent. These include procedures for high dimensional data visualization, pattern classification, prediction, and evolutionary computational search. An artificial data set is first employed to compare the proposed approach with other methods. A simulated handwritten data set and a real data set on liver disease diagnosis are then employed to evaluate the efficacy of the proposed approach. The results are analyzed and discussed. The potentials of the proposed architecture as a useful decision support system are demonstrated.
Abstract: This paper presents a speed fuzzy sliding mode
controller for a vector controlled induction machine (IM) fed by a
voltage source inverter (PWM).
The sliding mode based fuzzy control method is developed to
achieve fast response, a best disturbance rejection and to maintain a
good decoupling.
The problem with sliding mode control is that there is high
frequency switching around the sliding mode surface. The FSMC is
the combination of the robustness of Sliding Mode Control (SMC)
and the smoothness of Fuzzy Logic (FL). To reduce the torque
fluctuations (chattering), the sign function used in the conventional
SMC is substituted with a fuzzy logic algorithm.
The proposed algorithm was simulated by Matlab/Simulink
software and simulation results show that the performance of the
control scheme is robust and the chattering problem is solved.
Abstract: Keystroke authentication is a new access control system
to identify legitimate users via their typing behavior. In this paper,
machine learning techniques are adapted for keystroke authentication.
Seven learning methods are used to build models to differentiate user
keystroke patterns. The selected classification methods are Decision
Tree, Naive Bayesian, Instance Based Learning, Decision Table, One
Rule, Random Tree and K-star. Among these methods, three of them
are studied in more details. The results show that machine learning
is a feasible alternative for keystroke authentication. Compared to
the conventional Nearest Neighbour method in the recent research,
learning methods especially Decision Tree can be more accurate. In
addition, the experiment results reveal that 3-Grams is more accurate
than 2-Grams and 4-Grams for feature extraction. Also, combination
of attributes tend to result higher accuracy.
Abstract: In the time of globalisation, growing uncertainty, ambiguity and change, traditional way of doing business are no longer sufficient and it is important to consider non-conventional methods and approaches to release creativity and facilitate innovation and growth. Thus, creative industries, as a natural source of creativity and innovation, draw particular attention. This paper explores feasibility of building creative partnerships between creative industries and business and brings attention to mutual benefits derived from such partnerships. Design/approach - This paper is a theoretical exploration of projects, practices and research findings addressing collaboration between creative industries and business. Thus, it concerns creative industries, arts, business and its representatives in order to define requirements for creative partnerships to work and succeed. Findings – Current practices in engaging into arts-business partnerships are still very few, although most of creative partnerships proved to be highly valuable and mutually beneficial. Certain conditions shall be provided in order to benefit from arts-business creative synergy. Originality/value- By integrating different sources of literature, this article provides a base for conducting empirical research in several dimensions within arts-business partnerships.
Abstract: The anti-lock braking systems installed on vehicles
for safe and effective braking, are high-order nonlinear and timevariant.
Using fuzzy logic controllers increase efficiency of such
systems, but impose a high computational complexity as well. The
main concept introduced by this paper is reducing computational
complexity of fuzzy controllers by deploying problem-solution data
structure. Unlike conventional methods that are based on
calculations, this approach is based on data oriented modeling.
Abstract: This paper proposes a new methodology for the
optimal allocation and sizing of Embedded Generation (EG)
employing Real Coded Genetic Algorithm (RCGA) to minimize the
total power losses and to improve voltage profiles in the radial
distribution networks. RCGA is a method that uses continuous
floating numbers as representation which is different from
conventional binary numbers. The RCGA is used as solution tool,
which can determine the optimal location and size of EG in radial
system simultaneously. This method is developed in MATLAB. The
effect of EG units- installation and their sizing to the distribution
networks are demonstrated using 24 bus system.
Abstract: A state of the art Speaker Identification (SI) system
requires a robust feature extraction unit followed by a speaker
modeling scheme for generalized representation of these features.
Over the years, Mel-Frequency Cepstral Coefficients (MFCC)
modeled on the human auditory system has been used as a standard
acoustic feature set for speech related applications. On a recent
contribution by authors, it has been shown that the Inverted Mel-
Frequency Cepstral Coefficients (IMFCC) is useful feature set for
SI, which contains complementary information present in high
frequency region. This paper introduces the Gaussian shaped filter
(GF) while calculating MFCC and IMFCC in place of typical
triangular shaped bins. The objective is to introduce a higher
amount of correlation between subband outputs. The performances
of both MFCC & IMFCC improve with GF over conventional
triangular filter (TF) based implementation, individually as well as
in combination. With GMM as speaker modeling paradigm, the
performances of proposed GF based MFCC and IMFCC in
individual and fused mode have been verified in two standard
databases YOHO, (Microphone Speech) and POLYCOST
(Telephone Speech) each of which has more than 130 speakers.
Abstract: This study presents the improvement of thermal
performance of heat pipe using copper nanofluid with aqueous
solution of n-Butanol. The nanofluids kept in the suspension of
conventional fluids have the potential of superior heat transfer
capability than the conventional fluids due to their improved thermal
conductivity. In this work, the copper nanofluid which has a 40 nm
size with a concentration of 100 mg/lit is kept in the suspension of
the de-ionized (DI) water and an aqueous solution of n-Butanol and
these fluids are used as a working medium in the heat pipe. The study
discusses about the effect of heat pipe inclination, type of working
fluid and heat input on the thermal efficiency and thermal resistance.
The experimental results are evaluated in terms of its performance
metrics and are compared with that of DI water.
Abstract: This paper presents a systematic approach for the
design of power system stabilizer using genetic algorithm and
investigates the robustness of the GA based PSS. The proposed
approach employs GA search for optimal setting of PSS parameters.
The performance of the proposed GPSS under small and large
disturbances, loading conditions and system parameters is tested.
The eigenvalue analysis and nonlinear simulation results show the
effectiveness of the GPSS to damp out the system oscillations. It is
found tat the dynamic performance with the GPSS shows improved
results, over conventionally tuned PSS over a wide range of
operating conditions.
Abstract: Nanostructured materials have attracted many
researchers due to their outstanding mechanical and physical
properties. For example, carbon nanotubes (CNTs) or carbon
nanofibres (CNFs) are considered to be attractive reinforcement
materials for light weight and high strength metal matrix composites.
These composites are being projected for use in structural
applications for their high specific strength as well as functional
materials for their exciting thermal and electrical characteristics. The
critical issues of CNT-reinforced MMCs include processing
techniques, nanotube dispersion, interface, strengthening mechanisms
and mechanical properties. One of the major obstacles to the effective
use of carbon nanotubes as reinforcements in metal matrix
composites is their agglomeration and poor distribution/dispersion
within the metallic matrix. In order to tap into the advantages of the
properties of CNTs (or CNFs) in composites, the high dispersion of
CNTs (or CNFs) and strong interfacial bonding are the key issues
which are still challenging. Processing techniques used for synthesis
of the composites have been studied with an objective to achieve
homogeneous distribution of carbon nanotubes in the matrix.
Modified mechanical alloying (ball milling) techniques have emerged
as promising routes for the fabrication of carbon nanotube (CNT)
reinforced metal matrix composites. In order to obtain a
homogeneous product, good control of the milling process, in
particular control of the ball movement, is essential. The control of
the ball motion during the milling leads to a reduction in grinding
energy and a more homogeneous product. Also, the critical inner
diameter of the milling container at a particular rotational speed can
be calculated. In the present work, we use conventional and modified
mechanical alloying to generate a homogenous distribution of 2 wt.
% CNT within Al powders. 99% purity Aluminium powder (Acros,
200mesh) was used along with two different types of multiwall
carbon nanotube (MWCNTs) having different aspect ratios to
produce Al-CNT composites. The composite powders were processed
into bulk material by compaction, and sintering using a cylindrical
compaction and tube furnace. Field Emission Scanning electron
microscopy (FESEM), X-Ray diffraction (XRD), Raman
spectroscopy and Vickers macro hardness tester were used to
evaluate CNT dispersion, powder morphology, CNT damage, phase
analysis, mechanical properties and crystal size determination.
Despite the success of ball milling in dispersing CNTs in Al powder,
it is often accompanied with considerable strain hardening of the Al
powder, which may have implications on the final properties of the
composite. The results show that particle size and morphology vary
with milling time. Also, by using the mixing process and sonication
before mechanical alloying and modified ball mill, dispersion of the
CNTs in Al matrix improves.
Abstract: We here propose improved version of elastic graph matching (EGM) as a face detector, called the multi-scale EGM (MS-EGM). In this improvement, Gabor wavelet-based pyramid reduces computational complexity for the feature representation often used in the conventional EGM, but preserving a critical amount of information about an image. The MS-EGM gives us higher detection performance than Viola-Jones object detection algorithm of the AdaBoost Haar-like feature cascade. We also show rapid detection speeds of the MS-EGM, comparable to the Viola-Jones method. We find fruitful benefits in the MS-EGM, in terms of topological feature representation for a face.
Abstract: This paper examines predictability in stock return in
developed and emergingmarkets by testing long memory in stock
returns using wavelet approach. Wavelet-based maximum likelihood
estimator of the fractional integration estimator is superior to the
conventional Hurst exponent and Geweke and Porter-Hudak
estimator in terms of asymptotic properties and mean squared error.
We use 4-year moving windows to estimate the fractional integration
parameter. Evidence suggests that stock return may not be predictable
indeveloped countries of the Asia-Pacificregion. However,
predictability of stock return insome developing countries in this
region such as Indonesia, Malaysia and Philippines may not be ruled
out. Stock return in the Thailand stock market appears to be not
predictable after the political crisis in 2008.
Abstract: In this paper, a pipelined version of genetic algorithm,
called PLGA, and a corresponding hardware platform are described.
The basic operations of conventional GA (CGA) are made pipelined
using an appropriate selection scheme. The selection operator, used
here, is stochastic in nature and is called SA-selection. This helps
maintaining the basic generational nature of the proposed pipelined
GA (PLGA). A number of benchmark problems are used to compare
the performances of conventional roulette-wheel selection and the
SA-selection. These include unimodal and multimodal functions with
dimensionality varying from very small to very large. It is seen that
the SA-selection scheme is giving comparable performances with
respect to the classical roulette-wheel selection scheme, for all the
instances, when quality of solutions and rate of convergence are considered.
The speedups obtained by PLGA for different benchmarks
are found to be significant. It is shown that a complete hardware
pipeline can be developed using the proposed scheme, if parallel
evaluation of the fitness expression is possible. In this connection
a low-cost but very fast hardware evaluation unit is described.
Results of simulation experiments show that in a pipelined hardware
environment, PLGA will be much faster than CGA. In terms of
efficiency, PLGA is found to outperform parallel GA (PGA) also.
Abstract: As the Computed Tomography(CT) requires normally
hundreds of projections to reconstruct the image, patients are exposed
to more X-ray energy, which may cause side effects such as cancer.
Even when the variability of the particles in the object is very less,
Computed Tomography requires many projections for good quality
reconstruction. In this paper, less variability of the particles in an
object has been exploited to obtain good quality reconstruction.
Though the reconstructed image and the original image have same
projections, in general, they need not be the same. In addition
to projections, if a priori information about the image is known,
it is possible to obtain good quality reconstructed image. In this
paper, it has been shown by experimental results why conventional
algorithms fail to reconstruct from a few projections, and an efficient
polynomial time algorithm has been given to reconstruct a bi-level
image from its projections along row and column, and a known sub
image of unknown image with smoothness constraints by reducing the
reconstruction problem to integral max flow problem. This paper also
discusses the necessary and sufficient conditions for uniqueness and
extension of 2D-bi-level image reconstruction to 3D-bi-level image
reconstruction.
Abstract: Computed tomography and laminography are heavily investigated in a compressive sensing based image reconstruction framework to reduce the dose to the patients as well as to the radiosensitive devices such as multilayer microelectronic circuit boards. Nowadays researchers are actively working on optimizing the compressive sensing based iterative image reconstruction algorithm to obtain better quality images. However, the effects of the sampled data’s properties on reconstructed the image’s quality, particularly in an insufficient sampled data conditions have not been explored in computed laminography. In this paper, we investigated the effects of two data properties i.e. sampling density and data incoherence on the reconstructed image obtained by conventional computed laminography and a recently proposed method called spherical sinusoidal scanning scheme. We have found that in a compressive sensing based image reconstruction framework, the image quality mainly depends upon the data incoherence when the data is uniformly sampled.
Abstract: Conventional materials like glass, wood or metals
replacement with polymer materials is still continuing. More simple
thus cheaper production is the main reason. However due to high
energy and petrochemical prices are polymer prices increasing too.
That´s why various kinds of fillers are used to make polymers
cheaper. Of course target is to maintain or improve properties of
these compounds. In this paper are solved rheology issues of
polymers compounded with vegetal origin fibers.
Abstract: LabVIEW and SIMULINK are two most widely used
graphical programming environments for designing digital signal
processing and control systems. Unlike conventional text-based
programming languages such as C, Cµ and MATLAB, graphical
programming involves block-based code developments, allowing a
more efficient mechanism to build and analyze control systems. In
this paper a LabVIEW environment has been employed as a
graphical user interface for monitoring the operation of a controlled
distillation column, by visualizing both the closed loop performance
and the user selected control conditions, while the column dynamics
has been modeled under the SIMULINK environment. This tool has
been applied to the PID based decoupled control of a binary
distillation column. By means of such integrated environments the
control designer is able to monitor and control the plant behavior and
optimize the response when both, the quality improvement of
distillation products and the operation efficiency tasks, are
considered.
Abstract: This research presents a fuzzy multi-objective model
for a machine selection problem in a flexible manufacturing system
of a tire company. Two main objectives are minimization of an
average machine error and minimization of the total setup time.
Conventionally, the working team uses trial and error in selecting a
pressing machine for each task due to the complexity and constraints
of the problem. So, both objectives may not satisfy. Moreover, trial
and error takes a lot of time to get the final decision. Therefore, in
this research preemptive fuzzy goal programming model is developed
for solving this multi-objective problem. The proposed model can
obtain the appropriate results that the Decision Making (DM) is
satisfied for both objectives. Besides, alternative choice can be easily
generated by varying the satisfaction level. Additionally, decision
time can be reduced by using the model, which includes all
constraints of the system to generate the solutions. A numerical
example is also illustrated to show the effectiveness of the proposed
model.
Abstract: This paper proposes, for the first time, how the
challenges facing the guard-band designs including the margin
assist-circuits scheme for the screening-test in the coming process
generations should be addressed. The increased screening error
impacts are discussed based on the proposed statistical analysis
models. It has been shown that the yield-loss caused by the
misjudgment on the screening test would become 5-orders of
magnitude larger than that for the conventional one when the
amplitude of random telegraph noise (RTN) caused variations
approaches to that of random dopant fluctuation. Three fitting methods
to approximate the RTN caused complex Gamma mixtures
distributions by the simple Gaussian mixtures model (GMM) are
proposed and compared. It has been verified that the proposed
methods can reduce the error of the fail-bit predictions by 4-orders of
magnitude.