Abstract: In recent years, a new numerical method has been
developed, the extended finite element method (X-FEM). The
objective of this work is to exploit the (X-FEM) for the treatment of
the fracture mechanics problems on 3D geometries, where we
showed the ability of this method to simulate the fatigue crack
growth into two cases: edge and central crack. In the results we
compared the six first natural frequencies of mode shapes uncracking
with the cracking initiation in the structure, and showed the stress
intensity factor (SIF) evolution function as crack size propagation
into structure, the analytical validation of (SIF) is presented. For to
evidence the aspects of this method, all result is compared between
FEA and X-FEM.
Abstract: Knowledge sharing enables the information or
knowledge to be transmitted from one source to another. This paper
demonstrates the needs of having the online book catalogue which
can be used to facilitate disseminating information on textbook used
in the university. This project is aimed to give access to the students
and lecturers to the list of books in the bookstore and at the same
time to allow book reviewing without having to visit the bookstore
physically. Research is carried out according to the boundaries which
accounts to current process of new book purchasing, current system
used by the bookstore and current process the lecturers go through
for reviewing textbooks. The questionnaire is used to gather the
requirements and it is distributed to 100 students and 40 lecturers.
This project has enabled the improvement of a manual process to be
carried out automatically, through a web based platform. It is shown
based on the user acceptance survey carried out that target groups
found that this web service is feasible to be implemented in
Universiti Teknologi PETRONAS (UTP), and they have shown
positive signs of interest in utilizing it in the future.
Abstract: In this paper, we propose a texture feature-based
language identification using wavelet-domain BDIP (block difference
of inverse probabilities) and BVLC (block variance of local
correlation coefficients) features and FFT (fast Fourier transform)
feature. In the proposed method, wavelet subbands are first obtained
by wavelet transform from a test image and denoised by Donoho-s
soft-thresholding. BDIP and BVLC operators are next applied to the
wavelet subbands. FFT blocks are also obtained by 2D (twodimensional)
FFT from the blocks into which the test image is
partitioned. Some significant FFT coefficients in each block are
selected and magnitude operator is applied to them. Moments for each
subband of BDIP and BVLC and for each magnitude of significant
FFT coefficients are then computed and fused into a feature vector. In
classification, a stabilized Bayesian classifier, which adopts variance
thresholding, searches the training feature vector most similar to the
test feature vector. Experimental results show that the proposed
method with the three operations yields excellent language
identification even with rather low feature dimension.
Abstract: This paper is concerned with the design and implementation of MICOSim, an event-driven simulator written in Java for evaluating the performance of Grid entities (users, brokers and resources) under different scenarios such as varying the numbers of users, resources and brokers and varying their specifications and employed strategies.
Abstract: This article discusses the questions concerning of creating small packet networks for energy companies with application of high voltage power line carrier equipment (PLC) with functionality of IP traffic transmission. The main idea is to create converged PLC links between substations and dispatching centers where packet data and voice are transmitted in one data flow. The article contents description of basic conception of the network, evaluation of voice traffic transmission parameters, and discussion of header compression techniques in relation to PLC links. The results of exploration show us, that convergent packet PLC links can be very useful in the construction of small packet networks between substations in remote locations, such as deposits or low populated areas.
Abstract: In this paper, 3X3 routing nodes are proposed to
provide speedup and parallel processing capability in Data Vortex
network architectures. The new design not only significantly
improves network throughput and latency, but also eliminates the
need for distributive traffic control mechanism originally embedded
among nodes and the need for nodal buffering. The cost effectiveness
is studied by a comparison study with the previously proposed 2-
input buffered networks, and considerable performance enhancement
can be achieved with similar or lower cost of hardware. Unlike
previous implementation, the network leaves small probability of
contention, therefore, the packet drop rate must be kept low for such
implementation to be feasible and attractive, and it can be achieved
with proper choice of operation conditions.
Abstract: Today-s Voltage Regulator Modules (VRMs) face increasing design challenges as the number of transistors in microprocessors increases per Moore-s Law. These challenges have recently become even more demanding as microprocessors operate at sub voltage range at significantly high current. This paper presents a new multiphase topology with cell configuration for improved performance in low voltage and high current applications. A lab scale hardware prototype of the new topology was design and constructed. Laboratory tests were performed on the proposed converter and compared with a commercially available VRM. Results from the proposed topology exhibit improved performance compared to the commercially available counterpart.
Abstract: This paper presents a novel methodology for Maximum Power Point Tracking (MPPT) of a grid-connected 20 kW Photovoltaic (PV) system using neuro-fuzzy network. The proposed method predicts the reference PV voltage guarantying optimal power transfer between the PV generator and the main utility grid. The neuro-fuzzy network is composed of a fuzzy rule-based classifier and three Radial Basis Function Neural Networks (RBFNN). Inputs of the network (irradiance and temperature) are classified before they are fed into the appropriated RBFNN for either training or estimation process while the output is the reference voltage. The main advantage of the proposed methodology, comparing to a conventional single neural network-based approach, is the distinct generalization ability regarding to the nonlinear and dynamic behavior of a PV generator. In fact, the neuro-fuzzy network is a neural network based multi-model machine learning that defines a set of local models emulating the complex and non-linear behavior of a PV generator under a wide range of operating conditions. Simulation results under several rapid irradiance variations proved that the proposed MPPT method fulfilled the highest efficiency comparing to a conventional single neural network.
Abstract: The majority of existing predictors for time series are
model-dependent and therefore require some prior knowledge for the
identification of complex systems, usually involving system
identification, extensive training, or online adaptation in the case of
time-varying systems. Additionally, since a time series is usually
generated by complex processes such as the stock market or other
chaotic systems, identification, modeling or the online updating of
parameters can be problematic. In this paper a model-free predictor
(MFP) for a time series produced by an unknown nonlinear system or
process is derived using tracking theory. An identical derivation of the
MFP using the property of the Newton form of the interpolating
polynomial is also presented. The MFP is able to accurately predict
future values of a time series, is stable, has few tuning parameters and
is desirable for engineering applications due to its simplicity, fast
prediction speed and extremely low computational load. The
performance of the proposed MFP is demonstrated using the
prediction of the Dow Jones Industrial Average stock index.
Abstract: Nanomaterials have attracted considerable attention
during the last two decades, due to their unusual electrical, mechanical
and other physical properties as compared with their bulky
counterparts. The mechanical properties of nanostructured materials
show strong size dependency, which has been explained within the
framework of continuum mechanics by including the effects of surface
stress. The size-dependent deformations of two-dimensional
nanosized structures with surface effects are investigated in the paper
by the finite element method. Truss element is used to evaluate the
contribution of surface stress to the total potential energy and the
Gurtin and Murdoch surface stress model is implemented with
ANSYS through its user programmable features. The proposed
approach is used to investigate size-dependent stress concentration
around a nanosized circular hole and the size-dependent effective
moduli of nanoporous materials. Numerical results are compared with
available analytical results to validate the proposed modeling
approach.
Abstract: Software crisis refers to the situation in which the developers are not able to complete the projects within time and budget constraints and moreover these overscheduled and over budget projects are of low quality as well. Several methodologies have been adopted form time to time to overcome this situation and now in the focus is component based software engineering. In this approach, emphasis is on reuse of already existing software artifacts. But the results can not be achieved just by preaching the principles; they need to be practiced as well. This paper highlights some of the very basic elements of this approach, which has to be in place to get the desired goals of high quality, low cost with shorter time-to-market software products.
Abstract: The purpose of this investigation is to relate the rain
power and the overland flow power to soil erodibility to assess the
effects of both parameters on soil erosion using variable rainfall
intensity on remoulded agricultural soil. Six rainfall intensities were
used to simulate the natural rainfall and are as follows: 12.4mm/h,
20.3mm/h, 28.6mm/h, 52mm/h, 73.5mm/h and 103mm/h. The results
have shown that the relationship between overland flow power and
rain power is best represented by a linear function (R2=0.99). As
regards the relationships between soil erodibility factor and rain and
overland flow powers, the evolution of both parameters with the
erodibility factor follow a polynomial function with high coefficient
of determination. From their coefficients of determination (R2=0.95)
for rain power and (R2=0.96) for overland flow power, we can
conclude that the flow has more power to detach particles than rain.
This could be explained by the fact that the presence of particles,
already detached by rain and transported by the flow, give the flow
more weight and then contribute to the detachment of particles by
collision.
Abstract: The main goal of this work is to propose a way for
combined use of two nontraditional algorithms by solving topological
problems on telecommunications concentrator networks. The
algorithms suggested are the Simulated Annealing algorithm and the
Genetic Algorithm. The Algorithm of Simulated Annealing unifies
the well known local search algorithms. In addition - Simulated
Annealing allows acceptation of moves in the search space witch lead
to decisions with higher cost in order to attempt to overcome any
local minima obtained. The Genetic Algorithm is a heuristic approach
witch is being used in wide areas of optimization works. In the last
years this approach is also widely implemented in
Telecommunications Networks Planning. In order to solve less or
more complex planning problem it is important to find the most
appropriate parameters for initializing the function of the algorithm.
Abstract: Nowadays, the rapid development of multimedia
and internet allows for wide distribution of digital media data.
It becomes much easier to edit, modify and duplicate digital
information Besides that, digital documents are also easy to
copy and distribute, therefore it will be faced by many
threatens. It-s a big security and privacy issue with the large
flood of information and the development of the digital
format, it become necessary to find appropriate protection
because of the significance, accuracy and sensitivity of the
information. Nowadays protection system classified with more
specific as hiding information, encryption information, and
combination between hiding and encryption to increase information
security, the strength of the information hiding science is due to the
non-existence of standard algorithms to be used in hiding secret
messages. Also there is randomness in hiding methods such as
combining several media (covers) with different methods to pass a
secret message. In addition, there are no formal methods to be
followed to discover the hidden data. For this reason, the task of this
research becomes difficult. In this paper, a new system of information
hiding is presented. The proposed system aim to hidden information
(data file) in any execution file (EXE) and to detect the hidden file
and we will see implementation of steganography system which
embeds information in an execution file. (EXE) files have been
investigated. The system tries to find a solution to the size of the
cover file and making it undetectable by anti-virus software. The
system includes two main functions; first is the hiding of the
information in a Portable Executable File (EXE), through the
execution of four process (specify the cover file, specify the
information file, encryption of the information, and hiding the
information) and the second function is the extraction of the hiding
information through three process (specify the steno file, extract the
information, and decryption of the information). The system has
achieved the main goals, such as make the relation of the size of the
cover file and the size of information independent and the result file
does not make any conflict with anti-virus software.
Abstract: According to conjugate gradient algorithm, a new consensus protocol algorithm of discrete-time multi-agent systems is presented, which can achieve finite-time consensus. Finally, a numerical example is given to illustrate our theoretical result.
Abstract: Paper presents knowledge about types of test in area
of materials properties of selected methods of rapid prototyping
technologies. In today used rapid prototyping technologies for
production of models and final parts are used materials in initial state
as solid, liquid or powder material structure. In solid state are used
various forms such as pellets, wire or laminates. Basic range
materials include paper, nylon, wax, resins, metals and ceramics. In
Fused Deposition Modeling (FDM) rapid prototyping technology are
mainly used as basic materials ABS (Acrylonitrile Butadiene
Styrene), polyamide, polycarbonate, polyethylene and polypropylene.
For advanced FDM applications are used special materials as silicon
nitrate, PZT (Piezoceramic Material - Lead Zirconate Titanate),
aluminium oxide, hydroxypatite and stainless steel.
Abstract: The aim of this article is to narrate the utility of novel simulation approach i.e. convolution method to predict blood concentration of drug utilizing dissolution data of salbutamol sulphate microparticulate formulations with different release patterns (1:1, 1:2 and 1:3, drug:polymer). Dissolution apparatus II USP 2007 and 900 ml double distilled water stirrd at 50 rpm was employed for dissolution analysis. From dissolution data, blood drug concentration was determined, and in return predicted blood drug concentration data was used to calculate the pharmacokinetic parameters i.e. Cmax, Tmax, and AUC. Convolution is a good biwaiver technique; however its better utility needs it application in the conditions where biorelevant dissolution media are used.
Abstract: The world is moving rapidly toward the deployment
of information and communication systems. Nowadays, computing
systems with their fast growth are found everywhere and one of the main challenges for these systems is increasing attacks and security threats against them. Thus, capturing, analyzing and verifying security requirements becomes a very important activity in
development process of computing systems, specially in developing
systems such as banking, military and e-business systems. For
developing every system, a process model which includes a process,
methods and tools is chosen. The Rational Unified Process (RUP) is
one of the most popular and complete process models which is used
by developers in recent years. This process model should be extended to be used in developing secure software systems. In this
paper, the Requirement Discipline of RUP is extended to improve RUP for developing secure software systems. These proposed extensions are adding and integrating a number of Activities, Roles,
and Artifacts to RUP in order to capture, document and model threats
and security requirements of system. These extensions introduce a
group of clear and stepwise activities to developers. By following these activities, developers assure that security requirements are
captured and modeled. These models are used in design, implementation and test activitie
Abstract: This paper compares planning results of the electricity and water generation inventory up to year 2030 in the State of
Kuwait. Currently, the generation inventory consists of oil and gas fired technologies only. The planning study considers two main cases. The first case, Reference case, examines a generation inventory based on oil and gas fired generation technologies only.
The second case examines the inclusion of renewables as part of the generation inventory under two scenarios. In the first scenario, Ref-RE, renewable build-out is based on optimum economic performance
of overall generation system. Result shows that the optimum installed
renewable capacity with electric energy generation of 11% . In the second scenario, Ref-RE20, the renewable capacity build-out is
forced to provide 20% of electric energy by 2030. The respective energy systems costs of Reference, Ref-RE and Ref-RE20 case
scenarios reach US dollar 24, 10 and 14 billion annually in 2030.
Abstract: The need of high frame-rate imaging has been triggered by the new applications of ultrasound imaging to transient elastography and real-time 3D ultrasound. Using plane wave excitation (PWE) is one of the methods to achieve very high frame-rate imaging since an image can be formed with a single insonification. However, due to the lack of transmit focusing, the image quality with PWE is lower compared with those using conventional focused transmission. To solve this problem, we propose a filter-retrieved transmit focusing (FRF) technique combined with cross-correlation weighting (FRF+CC weighting) for high frame-rate imaging with PWE. A restrospective focusing filter is designed to simultaneously minimize the predefined sidelobe energy associated with single PWE and the filter energy related to the signal-to-noise-ratio (SNR). This filter attempts to maintain the mainlobe signals and to reduce the sidelobe ones, which gives similar mainlobe signals and different sidelobes between the original PWE and the FRF baseband data. Normalized cross-correlation coefficient at zero lag is calculated to quantify the degree of similarity at each imaging point and used as a weighting matrix to the FRF baseband data to further suppress sidelobes, thus improving the filter-retrieved focusing quality.