Abstract: The effect of extraction solvent upon properties
of carrageenan from Eucheuma cottonii was studied. The
distilled water and KOH solution (concentration 0.1- 0.5N) were
used as the solvent. Extraction process was carried out in water
bath equipped by stirrer with constant speed of 275 rpm with a
constant ratio of seaweed weight to solvent volume ( 1:50 g/mL)
at 86oC for 45 minutes. The extract was then precipitated in 3
volume of 90% ethanol, oven dried at 60oC. Based on
experimental data, alkali significantly influenced yield and
properties of extracted carrageenan. The extracted carrageenan
was found to have essentially identical FTIR spectra to the
reference samples of kappa-carrageenan. Increasing the KOH
concentration led to carrageenan containing less sulfate content
and intrinsic viscosity. The gel strength increased along with the
increasing of KOH concentration. The decreasing of intrinsic
viscosity value indicates that a polymer degradation occurs
during alkali extraction.
Abstract: For the first incumbent operator it is very important
to understand how to react when the second operator comes to the
market. In this paper which is prepared for preliminary study of
GSM market in Iran, we have studied five MENA markets
according to the similarity point of view. This paper aims at
analyzing the impact of second entrants in selected markets on
certain marketing key performance indicators (KPI) such as:
Market shares (by operator), prepaid share, minutes of use (MoU),
Price and average revenue per user (ARPU) (for total market
each).
Abstract: This paper explores the effectiveness of machine
learning techniques in detecting firms that issue fraudulent financial
statements (FFS) and deals with the identification of factors
associated to FFS. To this end, a number of experiments have been
conducted using representative learning algorithms, which were
trained using a data set of 164 fraud and non-fraud Greek firms in the
recent period 2001-2002. The decision of which particular method to
choose is a complicated problem. A good alternative to choosing
only one method is to create a hybrid forecasting system
incorporating a number of possible solution methods as components
(an ensemble of classifiers). For this purpose, we have implemented
a hybrid decision support system that combines the representative
algorithms using a stacking variant methodology and achieves better
performance than any examined simple and ensemble method. To
sum up, this study indicates that the investigation of financial
information can be used in the identification of FFS and underline the
importance of financial ratios.
Abstract: In single trial analysis, when using Principal
Component Analysis (PCA) to extract Visual Evoked Potential
(VEP) signals, the selection of principal components (PCs) is an
important issue. We propose a new method here that selects only
the appropriate PCs. We denote the method as selective eigen-rate
(SER). In the method, the VEP is reconstructed based on the rate
of the eigen-values of the PCs. When this technique is applied on
emulated VEP signals added with background
electroencephalogram (EEG), with a focus on extracting the
evoked P3 parameter, it is found to be feasible. The improvement
in signal to noise ratio (SNR) is superior to two other existing
methods of PC selection: Kaiser (KSR) and Residual Power (RP).
Though another PC selection method, Spectral Power Ratio (SPR)
gives a comparable SNR with high noise factors (i.e. EEGs), SER
give more impressive results in such cases. Next, we applied SER
method to real VEP signals to analyse the P3 responses for
matched and non-matched stimuli. The P3 parameters extracted
through our proposed SER method showed higher P3 response for
matched stimulus, which confirms to the existing neuroscience
knowledge. Single trial PCA using KSR and RP methods failed to
indicate any difference for the stimuli.
Abstract: The P-Bigram method is a string comparison methods
base on an internal two characters-based similarity measure. The edit
distance between two strings is the minimal number of elementary
editing operations required to transform one string into the other. The
elementary editing operations include deletion, insertion, substitution
two characters. In this paper, we address the P-Bigram method to
sole the similarity problem in DNA sequence. This method provided
an efficient algorithm that locates all minimum operation in a string.
We have been implemented algorithm and found that our program
calculated that smaller distance than one string. We develop PBigram
edit distance and show that edit distance or the similarity and
implementation using dynamic programming. The performance of
the proposed approach is evaluated using number edit and percentage
similarity measures.
Abstract: Chronic conditions carry with them strong emotions
and often lead to charged relationships between patients and their
health providers and, by extension, patients and health researchers.
Persons are both autonomous and relational and a purely cognitive
model of autonomy neglects the social and relational basis of chronic
illness. Ensuring genuine informed consent in research requires a
thorough understanding of how participants perceive a study and
their reasons for participation. Surveys may not capture the
complexities of reasoning that underlies study participation.
Contradictory reasons for participation, for instance an initial claim
of altruism as rationale and a subsequent claim of personal benefit
(therapeutic misconception), affect the quality of informed consent.
Individuals apply principles through the filter of personal values and
lived experience. Authentic autonomy, and hence authentic consent
to research, occurs within the context of patients- unique life
narratives and illness experiences.
Abstract: In this paper we propose a family of algorithms based
on 3rd and 4th order cumulants for blind single-input single-output
(SISO) Non-Minimum Phase (NMP) Finite Impulse Response (FIR)
channel estimation driven by non-Gaussian signal. The input signal
represents the signal used in 10GBASE-T (or IEEE 802.3an-2006)
as a Tomlinson-Harashima Precoded (THP) version of random
Pulse-Amplitude Modulation with 16 discrete levels (PAM-16). The
proposed algorithms are tested using three non-minimum phase
channel for different Signal-to-Noise Ratios (SNR) and for different
data input length. Numerical simulation results are presented to
illustrate the performance of the proposed algorithms.
Abstract: Grid composite structures have many applications in aerospace industry in which deal with transverse loadings abundantly. In present paper a stiffened composite cylindrical shell with clamped-free boundary condition under transverse end load experimentally and numerically was studied. Some electrical strain gauges were employed to measure the strains. Also a finite element analysis was done for validation of experimental result. The FEM software used was ANSYS11. In addition, the results between stiffened composite shell and unstiffened composite shell were compared. It was observed that intersection of two stiffeners has an important effect in decrease of stress in the shell. Fairly good agreements were observed between the numerical and the measured results. According to recent studies about grid composite structures, it should be noted that any investigation like this research has not been reported.
Abstract: In any distributed systems, process scheduling plays a
vital role in determining the efficiency of the system. Process scheduling algorithms are used to ensure that the components of the
system would be able to maximize its utilization and able to complete all the processes assigned in a specified period of time.
This paper focuses on the development of comparative simulator for distributed process scheduling algorithms. The objectives of the works that have been carried out include the development of the
comparative simulator, as well as to implement a comparative study
between three distributed process scheduling algorithms; senderinitiated,
receiver-initiated and hybrid sender-receiver-initiated
algorithms. The comparative study was done based on the Average Waiting Time (AWT) and Average Turnaround Time (ATT) of the
processes involved. The simulation results show that the performance of the algorithms depends on the number of nodes in the system.
Abstract: System MEMORI automatically detects and recognizes rotated and/or rescaled versions of the objects of a database within digital color images with cluttered background. This task is accomplished by means of a region grouping algorithm guided by heuristic rules, whose parameters concern some geometrical properties and the recognition score of the database objects. This paper focuses on the strategies implemented in MEMORI for the estimation of the heuristic rule parameters. This estimation, being automatic, makes the system a highly user-friendly tool.
Abstract: In Korea, the technology of a load fo nuclear power plant has been being developed.
automatic controller which is able to control temperature and axial power distribution was developed. identification algorithm and a model predictive contact former transforms the nuclear reactor status into
numerically. And the latter uses them and ge
manipulated values such as two kinds of control ro
this automatic controller, the performance of a coperation was evaluated. As a result, the automatic generated model parameters of a nuclear react to nuclear reactor average temperature and axial power the desired targets during a daily load follow.
Abstract: The flash memory has many advantages such as low power consumption, strong shock resistance, fast I/O and non-volatility. And it is increasingly used in the mobile storage device. The YAFFS, one of the NAND flash file system, is widely used in the embedded device. However, the existing YAFFS takes long time to mount the file system because it scans whole spare areas in all pages of NAND flash memory. In order to solve this problem, we propose a new content-based flash file system using a mounting time reduction technique. The proposed method only scans partial spare areas of some special pages by using content-based block management. The experimental results show that the proposed method reduces the average mounting time by 87.2% comparing with JFFS2 and 69.9% comparing with YAFFS.
Abstract: The separation of dissolved gas including dissolved oxygen can be used in breathing for a human under water. When one is suddenly wrecked or meets a tsunami, one is instantly drowned and cannot breathe under water. To avoid this crisis, when we meet waves, the dissolved gas separated from water by wave is used, while air can be used to breathe when we are about to escape from water. In this thesis, we investigated the separation characteristics of dissolved gas using the pipe type of hollow fiber membrane with polypropylene and the nude type of one with polysulfone. The hollow fiber membranes with good characteristics under water are used to separate the dissolved gas. The hollow fiber membranes with good characteristics in an air are used to transfer air. The combination of membranes with good separation characteristics under water and good transferring one in an air is used to breathe instantly under water to be alive at crisis. These results showed that polypropylene represented better performance than polysulfone under both of air and water conditions.
Abstract: All around the world pulp and paper industries are the
biggest plant production with the environmental pollution as the
biggest challenge facing the pulp manufacturing operations. The
concern among these industries is to produce a high volume of papers
with the high quality standard and of low cost without affecting the
environment. This result obtained from this bleaching study show
that the activation of peroxide was an effective method of reducing
the total applied charge of chlorine dioxide which is harmful to our
environment and also show that softwood and hardwood Kraft pulps
responded linearly to the peroxide treatments. During the bleaching
process the production plant produce chlorines. Under the trial stages
chloride dioxide has been reduced by 3 kg/ton to reduce the
brightness from 65% ISO to 60% ISO of pulp and the dosing point
returned to the E stage charges by pre-treating Kraft pulps with
hydrogen peroxide. The pulp and paper industry has developed
elemental chlorine free (ECF) and totally chlorine free (TCF)
bleaching, in their quest for being environmental friendly, they have
been looking at ways to turn their ECF process into a TCF process
while still being competitive. This prompted the research to
investigate the capability of the hydrogen peroxide as catalyst to
reduce chloride dioxide.
Abstract: The previous study of new metal gasket that contact
width and contact stress an important design parameter for optimizing
metal gasket performance. The optimum design based on an elastic
and plastic contact stress was founded. However, the influence of
flange surface roughness had not been investigated thoroughly. The
flange has many kinds of surface roughness. In this study, we
conducted a gasket model include a flange surface roughness effect. A
finite element method was employed to develop simulation solution. A
uniform quadratic mesh used for meshing the gasket material and a
gradually quadrilateral mesh used for meshing the flange. The gasket
model was simulated by using two simulation stages which is forming
and tightening simulation. A simulation result shows that a smoother
of surface roughness has higher slope for force per unit length. This
mean a squeezed against between flange and gasket will be strong. The
slope of force per unit length for gasket 400-MPa mode was higher
than the gasket 0-MPa mode.
Abstract: In this competitive age, one of the key tools of most successful organizations is knowledge management. Today some organizations measure their current knowledge and use it as an indicator for rating the organization on their reports. Noting that the universities and colleges of medical science have a great role in public health of societies, their access to newest scientific research and the establishment of organizational knowledge management systems is very important. In order to explore the Application of Knowledge Management Factors, a national study was undertaken. The main purpose of this study was to find the rate of the application of knowledge management factors and some ways to establish more application of knowledge management system in Esfahan University-s Medical College (EUMC). Esfahan is the second largest city after Tehran, the capital city of Iran, and the EUMC is the biggest medical college in Esfahan. To rate the application of knowledge management, this study uses a quantitative research methodology based on Probst, Raub and Romhardt model of knowledge management. A group of 267 faculty members and staff of the EUMC were asked via questionnaire. Finding showed that the rate of the application of knowledge management factors in EUMC have been lower than average. As a result, an interview with ten faculty members conducted to find the guidelines to establish more applications of knowledge management system in EUMC.
Abstract: This work considered the thermodynamic feasibility
of scrubbing volatile organic compounds into biodiesel in view of
designing a gas treatment process with this absorbent. A detailed
vapour – liquid equilibrium investigation was performed using the
original UNIFAC group contribution method. The four biodiesels
studied in this work are methyl oleate, methyl palmitate, methyl
linolenate and ethyl stearate. The original UNIFAC procedure was
used to estimate the infinite dilution activity coefficients of 13
selected volatile organic compounds in the biodiesels. The
calculations were done at the VOC mole fraction of 9.213x10-8. Ethyl
stearate gave the most favourable phase equilibrium. A close
agreement was found between the infinite dilution activity coefficient
of toluene found in this work and those reported in literature.
Thermodynamic models can efficiently be used to calculate vast
amount of phase equilibrium behaviour using limited number of
experimental data.
Abstract: IEEE has recently incorporated CCMP protocol to provide robust security to IEEE 802.11 wireless LANs. It is found that CCMP has been designed with a weak nonce construction and transmission mechanism, which leads to the exposure of initial counter value. This weak construction of nonce renders the protocol vulnerable to attacks by intruders. This paper presents how the initial counter can be pre-computed by the intruder. This vulnerability of counter block value leads to pre-computation attack on the counter mode encryption of CCMP. The failure of the counter mode will result in the collapse of the whole security mechanism of 802.11 WLAN.
Abstract: Recent years have seen a growing trend towards the
integration of multiple information sources to support large-scale
prediction of protein-protein interaction (PPI) networks in model
organisms. Despite advances in computational approaches, the
combination of multiple “omic" datasets representing the same type
of data, e.g. different gene expression datasets, has not been
rigorously studied. Furthermore, there is a need to further investigate
the inference capability of powerful approaches, such as fullyconnected
Bayesian networks, in the context of the prediction of PPI
networks. This paper addresses these limitations by proposing a
Bayesian approach to integrate multiple datasets, some of which
encode the same type of “omic" data to support the identification of
PPI networks. The case study reported involved the combination of
three gene expression datasets relevant to human heart failure (HF).
In comparison with two traditional methods, Naive Bayesian and
maximum likelihood ratio approaches, the proposed technique can
accurately identify known PPI and can be applied to infer potentially
novel interactions.
Abstract: Based on a non-linear single track model which
describes the dynamics of vehicle, an optimal path planning strategy
is developed. Real time optimization is used to generate reference
control values to allow leading the vehicle alongside a calculated lane
which is optimal for different objectives such as energy consumption,
run time, safety or comfort characteristics. Strict mathematic
formulation of the autonomous driving allows taking decision on
undefined situation such as lane change or obstacle avoidance. Based
on position of the vehicle, lane situation and obstacle position, the
optimization problem is reformulated in real-time to avoid the
obstacle and any car crash.