Abstract: Searching similar documents and document
management subjects have important place in text mining. One of the
most important parts of similar document research studies is the
process of classifying or clustering the documents. In this study, a
similar document search approach that includes discussion of out the
case of belonging to multiple categories (multiple categories
problem) has been carried. The proposed method that based on Fuzzy
Similarity Classification (FSC) has been compared with Rocchio
algorithm and naive Bayes method which are widely used in text
mining. Empirical results show that the proposed method is quite
successful and can be applied effectively. For the second stage,
multiple categories vector method based on information of categories
regarding to frequency of being seen together has been used.
Empirical results show that achievement is increased almost two
times, when proposed method is compared with classical approach.
Abstract: Nowadays, fluidized bed plays an important part in industry. The design of this kind of reactor requires knowing the interfacial area between two phases and this interfacial area leads to calculate the solid holdup in the bed. Consequently achieving interfacial area between gas and solid in the bed experimentally is so significant. On interfacial area measurement in fluidized bed with gas has been worked, but light transmission technique has been used less. Therefore, in the current research the possibility of using of this technique and its accuracy are investigated. Measuring, a fluidized bed was designed and the problems were averted as far as possible. By using fine solid with equal shape and diameter and installing an optical system, the absorption of light during the time of fluidization has been measured. Results indicate that this method that its validity has been proved in the gas-liquid system, by different reasons have less application in gas-solid system. One important reason could be non-uniformity in such systems.
Abstract: Traditional optical networks are gradually evolving towards intelligent optical networks due to the need for faster bandwidth provisioning, protection and restoration of the network that can be accomplished with devices like optical switch, add drop multiplexer and cross connects. Since dense wavelength multiplexing forms the physical layer for intelligent optical networking, the roll of high speed all optical switch is important. This paper analyzes such an ultra-high speed polymer electro-optic switch. The performances of the 2x2 optical waveguide switch with rectangular, triangular and trapezoidal grating profiles on various device parameters are analyzed. The simulation result shows that trapezoidal grating is the optimized structure which has the coupling length of 81μm and switching voltage of 11V for the operating wavelength of 1550nm. The switching time for this proposed switch is 0.47 picosecond. This makes the proposed switch to be an important element in the intelligent optical network.
Abstract: One of the major problems in liberalized power
markets is loss allocation. In this paper, a different method for
allocating transmission losses to pool market participants is
proposed. The proposed method is fundamentally based on
decomposition of loss function and current projection concept. The
method has been implemented and tested on several networks and
one sample summarized in the paper. The results show that the
method is comprehensive and fair to allocating the energy losses of a
power market to its participants.
Abstract: The problems associated with wind predictions of
WAsP model in complex terrain are already the target of several
studies in the last decade. In this paper, the influence of surrounding
orography on accuracy of wind data analysis of a train is
investigated. For the case study, a site with complex surrounding
orography is considered. This site is located in Manjil, one of the
windiest cities of Iran. For having precise evaluation of wind regime
in the site, one-year wind data measurements from two metrological
masts are used. To validate the obtained results from WAsP, the
cross prediction between each mast is performed. The analysis
reveals that WAsP model can estimate the wind speed behavior
accurately. In addition, results show that this software can be used
for predicting the wind regime in flat sites with complex surrounding
orography.
Abstract: The study examined the influence of pay differentials on employee retention in the State Colleges of Education in the South-South Region of Nigeria. 275 subjects drawn from members of the wage negotiating teams in the Colleges were administered questionnaires constructed for study. Analysis of Variance revealed that the observed pay differentials significantly influenced retainership, f(5,269 = 6.223, P< 0.05). However, the Multiple Classification Analysis and Post-Hoc test indicated that employees in two of the Colleges with slightly lower and higher pay levels may probably remain with their employers while employees in other Colleges with the least and highest pay levels suggested quitting. Based on these observations, the influence of pay on employee retention seems inconclusive. Generally, employees in the colleges studied are dissatisfied with current pay levels. Management should confront these challenges by improving pay packages to encourage employees to remain and be dedicated to duty.
Abstract: The chemistry of sulphone hydrazide has gained increase interest in both synthetic organic chemistry and biological fields and has considerable value. The therapeutic importance of these compounds is the attractive force to continue research in such a point. The present review covers the literature up to date for the synthesis, reactions and applications of such compounds.
Abstract: The Minimum Vertex Cover (MVC) problem is a classic
graph optimization NP - complete problem. In this paper a competent
algorithm, called Vertex Support Algorithm (VSA), is designed to
find the smallest vertex cover of a graph. The VSA is tested on a
large number of random graphs and DIMACS benchmark graphs.
Comparative study of this algorithm with the other existing methods
has been carried out. Extensive simulation results show that the VSA
can yield better solutions than other existing algorithms found in the
literature for solving the minimum vertex cover problem.
Abstract: Monitored 3-Dimensional (3D) video experience can be utilized as “feedback information” to fine tune the service parameters for providing a better service to the demanding 3D service customers. The 3D video experience which includes both video quality and depth perception is influenced by several contextual and content related factors (e.g., ambient illumination condition, content characteristics, etc) due to the complex nature of the 3D video. Therefore, effective factors on this experience should be utilized while assessing it. In this paper, structural information of the depth map sequences of the 3D video is considered as content related factor effective on the depth perception assessment. Cartoon-like filter is utilized to abstract the significant depth levels in the depth map sequences to determine the structural information. Moreover, subjective experiments are conducted using 3D videos associated with cartoon-like depth map sequences to investigate the effectiveness of ambient illumination condition, which is a contextual factor, on depth perception. Using the knowledge gained through this study, 3D video experience metrics can be developed to deliver better service to the 3D video service users.
Abstract: Learning programming is difficult for many learners. Some researches have found that the main difficulty relates to cognitive load. Cognitive overload happens in programming due to the nature of the subject which is intrinisicly over-bearing on the working memory. It happens due to the complexity of the subject itself. The problem is made worse by the poor instructional design methodology used in the teaching and learning process. Various efforts have been proposed to reduce the cognitive load, e.g. visualization softwares, part-program method etc. Use of many computer based systems have also been tried to tackle the problem. However, little success has been made to alleviate the problem. More has to be done to overcome this hurdle. This research attempts at understanding how cognitive load can be managed so as to reduce the problem of overloading. We propose a mechanism to measure the cognitive load during pre instruction, post instruction and in instructional stages of learning. This mechanism is used to help the instruction. As the load changes the instruction is made to adapt itself to ensure cognitive viability. This mechanism could be incorporated as a sub domain in the student model of various computer based instructional systems to facilitate the learning of programming.
Abstract: A synchronous network-on-chip using wormhole packet switching
and supporting guaranteed-completion best-effort with low-priority (LP)
and high-priority (HP) wormhole packet delivery service is presented in
this paper. Both our proposed LP and HP message services deliver a good
quality of service in term of lossless packet completion and in-order message
data delivery. However, the LP message service does not guarantee minimal
completion bound. The HP packets will absolutely use 100% bandwidth of
their reserved links if the HP packets are injected from the source node with
maximum injection. Hence, the service are suitable for small size messages
(less than hundred bytes). Otherwise the other HP and LP messages, which
require also the links, will experience relatively high latency depending on the
size of the HP message. The LP packets are routed using a minimal adaptive
routing, while the HP packets are routed using a non-minimal adaptive routing
algorithm. Therefore, an additional 3-bit field, identifying the packet type,
is introduced in their packet headers to classify and to determine the type
of service committed to the packet. Our NoC prototypes have been also
synthesized using a 180-nm CMOS standard-cell technology to evaluate the
cost of implementing the combination of both services.
Abstract: Linear stability analysis of wake-shear layers in twophase
shallow flows is performed in the present paper. Twodimensional
shallow water equations are used in the analysis. It is
assumed that the fluid contains uniformly distributed solid particles.
No dynamic interaction between the carrier fluid and particles is
expected in the initial moment. The stability calculations are
performed for different values of the particle loading parameter and
two other parameters which characterize the velocity ratio and the
velocity deficit. The results show that the particle loading parameter
has a stabilizing effect on the flow while the increase in the velocity
ratio or in the velocity deficit destabilizes the flow.
Abstract: The spectral action balance equation is an equation that
used to simulate short-crested wind-generated waves in shallow water
areas such as coastal regions and inland waters. This equation consists
of two spatial dimensions, wave direction, and wave frequency which
can be solved by finite difference method. When this equation with
dominating convection term are discretized using central differences,
stability problems occur when the grid spacing is chosen too coarse.
In this paper, we introduce the splitting upwind schemes for avoiding
stability problems and prove that it is consistent to the upwind scheme
with same accuracy. The splitting upwind schemes was adopted
to split the wave spectral action balance equation into four onedimensional
problems, which for each small problem obtains the
independently tridiagonal linear systems. For each smaller system
can be solved by direct or iterative methods at the same time which
is very fast when performed by a multi-processor computer.
Abstract: In the last decade, energy based control theory has undergone a significant breakthrough in dealing with underactated mechanical systems with two successful and similar tools, controlled Lagrangians and controlled Hamiltanians (IDA-PBC). However, because of the complexity of these tools, successful case studies are lacking, in particular, MIMO cases. The seminal theoretical paper of controlled Lagrangians proposed by Bloch and his colleagues presented a benchmark example–a 4 d.o.f underactuated pendulum on a cart but a detailed and completed design is neglected. To compensate this ignorance, the note revisit their design idea by addressing explicit control functions for a similar device motivated by a vector thrust body hovering in the air. To the best of our knowledge, this system is the first MIMO, underactuated example that is stabilized by using energy based tools at the courtesy of the original design idea. Some observations are given based on computer simulation.
Abstract: Software-as-a-Service (SaaS) is a form of cloud
computing that relieves the user of the burden of hardware and
software installation and management. SaaS can be used at the course
level to enhance curricula and student experience. When cloud
computing and SaaS are included in educational literature, the focus
is typically on implementing administrative functions. Yet, SaaS can
make more immediate and substantial contributions to the technical
course content in educational offerings. This paper explores cloud
computing and SaaS, provides examples, reports on experiences
using SaaS to offer specialized software in courses, and analyzes the
advantages and disadvantages of using SaaS at the course level. The
paper contributes to the literature in higher education by analyzing
the major technical concepts, potential, and constraints for using
SaaS to deliver specialized software at the course level. Further it
may enable more educators and students to benefit from this
emerging technology.
Abstract: Microarrays have become the effective, broadly used tools in biological and medical research to address a wide range of problems, including classification of disease subtypes and tumors. Many statistical methods are available for analyzing and systematizing these complex data into meaningful information, and one of the main goals in analyzing gene expression data is the detection of samples or genes with similar expression patterns. In this paper, we express and compare the performance of several clustering methods based on data preprocessing including strategies of normalization or noise clearness. We also evaluate each of these clustering methods with validation measures for both simulated data and real gene expression data. Consequently, clustering methods which are common used in microarray data analysis are affected by normalization and degree of noise and clearness for datasets.
Abstract: In this paper, a two-dimensional (2D) numerical
model for the tidal currents simulation in Persian Gulf is presented.
The model is based on the depth averaged equations of shallow water
which consider hydrostatic pressure distribution. The continuity
equation and two momentum equations including the effects of bed
friction, the Coriolis effects and wind stress have been solved. To
integrate the 2D equations, the Alternative Direction Implicit (ADI)
technique has been used. The base of equations discritization was
finite volume method applied on rectangular mesh. To evaluate the
model validation, a dam break case study including analytical
solution is selected and the comparison is done. After that, the
capability of the model in simulation of tidal current in a real field is
represented by modeling the current behavior in Persian Gulf. The
tidal fluctuations in Hormuz Strait have caused the tidal currents in
the area of study. Therefore, the water surface oscillations data at
Hengam Island on Hormoz Strait are used as the model input data.
The check point of the model is measured water surface elevations at
Assaluye port. The comparison between the results and the
acceptable agreement of them showed the model ability for modeling
marine hydrodynamic.
Abstract: This paper deals with heterogeneous autoregressive
models of realized volatility (HAR-RV models) on high-frequency
data of stock indices in the USA. Its aim is to capture the behavior of
three groups of market participants trading on a daily, weekly and
monthly basis and assess their role in predicting the daily realized
volatility. The benefits of this work lies mainly in the application of
heterogeneous autoregressive models of realized volatility on stock
indices in the USA with a special aim to analyze an impact of the
global financial crisis on applied models forecasting performance.
We use three data sets, the first one from the period before the global
financial crisis occurred in the years 2006-2007, the second one from
the period when the global financial crisis fully hit the U.S. financial
market in 2008-2009 years, and the last period was defined over
2010-2011 years. The model output indicates that estimated realized
volatility in the market is very much determined by daily traders and
in some cases excludes the impact of those market participants who
trade on monthly basis.
Abstract: As a part of evaluation system for R&D program, the
Korean government has applied feasibility analysis since 2008.
Various professionals put forth a great effort in order to catch up the
high degree of freedom of R&D programs, and make contributions to
evolving the feasibility analysis. We analyze diverse R&D programs
from various viewpoints, such as technology, policy, and Economics,
integrate the separate analysis, and finally arrive at a definite result;
whether a program is feasible or unfeasible. This paper describes the
concept and method of the feasibility analysis as a decision making
tool. The analysis unit and content of each criterion, which are key
elements in a comprehensive decision making structure, are examined
Abstract: In this paper, we propose an effective relay
communication for layered video transmission as an alternative to
make the most of limited resources in a wireless communication
network where loss often occurs. Relaying brings stable multimedia
services to end clients, compared to multiple description coding
(MDC). Also, retransmission of only parity data about one or more
video layer using channel coder to the end client of the relay device is
paramount to the robustness of the loss situation. Using these
methods in resource-constrained environments, such as real-time user
created content (UCC) with layered video transmission, can provide
high-quality services even in a poor communication environment.
Minimal services are also possible. The mathematical analysis shows
that the proposed method reduced the probability of GOP loss rate
compared to MDC and raptor code without relay. The GOP loss rate
is about zero, while MDC and raptor code without relay have a GOP
loss rate of 36% and 70% in case of 10% frame loss rate.