Abstract: Islamic banking is one the most blossoming doctrine in
economic system of the world. The Fast growing awareness about
Islamic financial system has brought strong feeling to Muslims to
confront the western interest-based economic cycle. The Islamic
economic system is emerging as a reliable alternative to the interest
based system. This study is proposed to ascertain the motivational
factors encouraging people to go for Islamic banking in Pakistan.
These pulsing factors are determined by generation of hypothesis that
there are certain factors which are urging people to opt Islamic
banking system and to see the differences in their ranking by applying
Friedman test. These factors include: Economically derived factors
such as stability of Islamic banks in crisis, profit and loss sharing
doctrine and equity sharing etc. This study also highlights the
religiously derived factors such as interest free banking, Shariah
tenets and supervisory of Islamic Shariah board and sociopsychological
factors.
Abstract: This research paper deals with the implementation of face recognition using neural network (recognition classifier) on low-resolution images. The proposed system contains two parts, preprocessing and face classification. The preprocessing part converts original images into blurry image using average filter and equalizes the histogram of those image (lighting normalization). The bi-cubic interpolation function is applied onto equalized image to get resized image. The resized image is actually low-resolution image providing faster processing for training and testing. The preprocessed image becomes the input to neural network classifier, which uses back-propagation algorithm to recognize the familiar faces. The crux of proposed algorithm is its beauty to use single neural network as classifier, which produces straightforward approach towards face recognition. The single neural network consists of three layers with Log sigmoid, Hyperbolic tangent sigmoid and Linear transfer function respectively. The training function, which is incorporated in our work, is Gradient descent with momentum (adaptive learning rate) back propagation. The proposed algorithm was trained on ORL (Olivetti Research Laboratory) database with 5 training images. The empirical results provide the accuracy of 94.50%, 93.00% and 90.25% for 20, 30 and 40 subjects respectively, with time delay of 0.0934 sec per image.
Abstract: Many single or multispan arch bridges are
strengthened with the addition of some kind of structural support
between adjacent arches of multispan or beside the arch barrel of a
single span to increase the strength of the overall structure. It was
traditionally formed by either placing loose rubble masonry blocks
between the arches and beside the arches or using mortar or concrete
to construct a more substantial structural bond between the spans. On
the other hand backing materials are present in some existing bridges.
Existing arch assessment procedures generally ignore the effects of
backing materials. In this paper an investigation of the effects of
backing on ratings for masonry arch bridges is carried out. It is
observed that increasing the overall lateral stability of the arch
system through the inclusion of structural backing results in an
enhanced failure load by reducing the likelihood of any tension
occurring at the top of the arch.
Abstract: In a complex project environment, project teams face
multi-dimensional communication problems that can ultimately lead
to project breakdown. Team Performance varies in Face-to-Face
(FTF) environment versus groups working remotely in a computermediated
communication (CMC) environment. A brief review of the
Input_Process_Output model suggested by James E. Driskell, Paul H.
Radtke and Eduardo Salas in “Virtual Teams: Effects of
Technological Mediation on Team Performance (2003)", has been
done to develop the basis of this research. This model theoretically
analyzes the effects of technological mediation on team processes,
such as, cohesiveness, status and authority relations, counternormative
behavior and communication. An empirical study
described in this paper has been undertaken to test the
“cohesiveness" of diverse project teams in a multi-national
organization. This study uses both quantitative and qualitative
techniques for data gathering and analysis. These techniques include
interviews, questionnaires for data collection and graphical data
representation for analyzing the collected data. Computer-mediated
technology may impact team performance because of difference in
cohesiveness among teams and this difference may be moderated by
factors, such as, the type of communication environment, the type of
task and the temporal context of the team. Based on the reviewed
model, sets of hypotheses are devised and tested. This research,
reports on a study that compared team cohesiveness among virtual
teams using CMC and non-CMC communication mediums. The
findings suggest that CMC can help virtual teams increase team
cohesiveness among their members, making CMC an effective
medium for increasing productivity and team performance.
Abstract: Removal of Methylene Blue (MB) from aqueous
solution by adsorbing it on Gypsum was investigated by batch
method. The studies were conducted at 25°C and included the effects
of pH and initial concentration of Methylene Blue. The adsorption
data was analyzed by using the Langmuir, Freundlich and Tempkin
isotherm models. The maximum monolayer adsorption capacity was
found to be 36 mg of the dye per gram of gypsum. The data were
also analyzed in terms of their kinetic behavior and was found to
obey the pseudo second order equation.
Abstract: During recent years, attention in 'Green Computing'
has moved research into energy-saving techniques for home
computers to enterprise systems' Client and Server machines. Saving
energy or reduction of carbon footprints is one of the aspects of
Green Computing. The research in the direction of Green Computing
is more than just saving energy and reducing carbon foot prints. This
study provides a brief account of Green Computing. The emphasis of
this study is on current trends in Green Computing; challenges in the
field of Green Computing and the future trends of Green Computing.
Abstract: This paper presents the development of low cost Nano membrane fabrication system. The system is specially designed for anodic aluminum oxide membrane. This system is capable to perform the processes such as anodization and electro-polishing. The designed machine was successfully tested for 'mild anodization' (MA) for 48 hours and 'hard anodization' (HA) for 3 hours at constant 0oC. The system is digitally controlled and guided for temperature maintenance during anodization and electro-polishing. The total cost of the developed machine is 20 times less than the multi-cooling systems available in the market which are generally used for this purpose.
Abstract: Magnetic and semiconductor nanomaterials exhibit
novel magnetic and optical properties owing to their unique size and
shape-dependent effects. With shrinking the size down to nanoscale
region, various anomalous properties that normally not present in bulk
start to dominate. Ability in harnessing of these anomalous properties
for the design of various advance electronic devices is strictly
dependent on synthetic strategies. Hence, current research has focused
on developing a rational synthetic control to produce high quality
nanocrystals by using organometallic approach to tune both size and
shape of the nanomaterials. In order to elucidate the growth
mechanism, transmission electron microscopy was employed as a
powerful tool in performing real time-resolved morphologies and
structural characterization of magnetic (Fe3O4) and semiconductor
(ZnO) nanocrystals. The current synthetic approach is found able to
produce nanostructures with well-defined shapes. We have found that
oleic acid is an effective capping ligand in preparing oxide-based
nanostructures without any agglomerations, even at high temperature.
The oleate-based precursors and capping ligands are fatty acid
compounds, which are respectively originated from natural palm oil
with low toxicity. In comparison with other synthetic approaches in
producing nanostructures, current synthetic method offers an effective
route to produce oxide-based nanomaterials with well-defined shapes
and good monodispersity. The nanocystals are well-separated with
each other without any stacking effect. In addition, the as-synthesized
nanopellets are stable in terms of chemically and physically if
compared to those nanomaterials that are previous reported. Further
development and extension of current synthetic strategy are being
pursued to combine both of these materials into nanocomposite form
that will be used as “smart magnetic nanophotocatalyst" for industry
waste water treatment.
Abstract: Natural pozzolan (NP) is one of the potential
prehistoric alternative binders in the construction industry. It has
been investigated as cement replacement in ordinary concrete by
several researchers for many purposes. Various supplementary
cementitious materials (SCMs) such as fly ash, limestone dust and
silica fume are widely used in the production of SCC; however,
limited studies to address the effect of NP on the properties of SCC
are documented. The current research is composed of different SCC
paste and concrete mixtures containing different replacement levels
of local NP as an alternative SCM. The effect of volume of paste
containing different amounts of local NP related to W/B ratio and
cement content on SCC fresh properties was assessed. The variations
in the fresh properties of SCC paste and concrete represented by
slump flow (flowability) and the flow rate were determined and
discussed. The results indicated that the flow properties of SCC paste
and concrete mixtures, at their optimized superplasticizer dosages,
were affected by the binder content of local NP and the total volume
fraction of SCC paste.
Abstract: The research on the effectiveness of environmental
assessment (EA) is a milestone effort to evaluate the state of the field,
including many contributors related with a lot of countries since more
than two decades. In the 1960s, there was a surge of interest between
modern industrialized countries over unexpected opposite effects of
technical invention. The interest led to choice of approaches for
assessing and prediction the impressions of technology and
advancement for social and economic, state health and safety, solidity
and the circumstances. These are consisting of risk assessment,
technology assessment, environmental impact assessment and costbenefit
analysis. In this research contribution, the authors have
described the research status for environmental assessment in
cumulative environmental system. This article discusses the methods
for cumulative effect assessment (CEA).
Abstract: Real-time object tracking is a problem which involves extraction of critical information from complex and uncertain imagedata. In this paper, we present a comprehensive methodology to design an artificial neural network (ANN) for a real-time object tracking application. The object, which is tracked for the purpose of demonstration, is a specific airplane. However, the proposed ANN can be trained to track any other object of interest. The ANN has been simulated and tested on the training and testing datasets, as well as on a real-time streaming video. The tracking error is analyzed with post-regression analysis tool, which finds the correlation among the calculated coordinates and the correct coordinates of the object in the image. The encouraging results from the computer simulation and analysis show that the proposed ANN architecture is a good candidate solution to a real-time object tracking problem.
Abstract: The study investigates the causal link between trade
openness and economic growth for four South Asian countries for
period 1972-1985 and 1986-2007 to examine the scenario before and
after the implementation of SAARC. Panel cointegration and
FMOLS techniques are employed for short run and long run
estimates. In 1972-85 short run unidirectional causality from GDP to
openness is found whereas, in 1986-2007 there exists bi-directional
causality between GDP and openness. The long run elasticity
magnitude between GDP and openness contains negative sign in
1972-85 which shows that there exists long run negative relationship.
While in time period 1986-2007 the elasticity magnitude has positive
sign that indicates positive causation between GDP and openness. So
it can be concluded that after the implementation of SAARC overall
situation of selected countries got better. Also long run coefficient of
error term suggests that short term equilibrium adjustments are driven
by adjustment back to long run equilibrium.
Abstract: Very Large and/or computationally complex optimization problems sometimes require parallel or highperformance computing for achieving a reasonable time for computation. One of the most popular and most complicate problems of this family is “Traveling Salesman Problem". In this paper we have introduced a Branch & Bound based algorithm for the solution of such complicated problems. The main focus of the algorithm is to solve the “symmetric traveling salesman problem". We reviewed some of already available algorithms and felt that there is need of new algorithm which should give optimal solution or near to the optimal solution. On the basis of the use of logarithmic sampling, it was found that the proposed algorithm produced a relatively optimal solution for the problem and results excellent performance as compared with the traditional algorithms of this series.
Abstract: Surface roughness (Ra) is one of the most important requirements in machining process. In order to obtain better surface roughness, the proper setting of cutting parameters is crucial before the process take place. This research presents the development of mathematical model for surface roughness prediction before milling process in order to evaluate the fitness of machining parameters; spindle speed, feed rate and depth of cut. 84 samples were run in this study by using FANUC CNC Milling α-Τ14ιE. Those samples were randomly divided into two data sets- the training sets (m=60) and testing sets(m=24). ANOVA analysis showed that at least one of the population regression coefficients was not zero. Multiple Regression Method was used to determine the correlation between a criterion variable and a combination of predictor variables. It was established that the surface roughness is most influenced by the feed rate. By using Multiple Regression Method equation, the average percentage deviation of the testing set was 9.8% and 9.7% for training data set. This showed that the statistical model could predict the surface roughness with about 90.2% accuracy of the testing data set and 90.3% accuracy of the training data set.
Abstract: World has entered in 21st century. The technology of
computer graphics and digital cameras is prevalent. High resolution
display and printer are available. Therefore high resolution images
are needed in order to produce high quality display images and high
quality prints. However, since high resolution images are not usually
provided, there is a need to magnify the original images. One
common difficulty in the previous magnification techniques is that of
preserving details, i.e. edges and at the same time smoothing the data
for not introducing the spurious artefacts. A definitive solution to this
is still an open issue. In this paper an image magnification using
adaptive interpolation by pixel level data-dependent geometrical
shapes is proposed that tries to take into account information about
the edges (sharp luminance variations) and smoothness of the image.
It calculate threshold, classify interpolation region in the form of
geometrical shapes and then assign suitable values inside
interpolation region to the undefined pixels while preserving the
sharp luminance variations and smoothness at the same time.
The results of proposed technique has been compared qualitatively
and quantitatively with five other techniques. In which the qualitative
results show that the proposed method beats completely the Nearest
Neighbouring (NN), bilinear(BL) and bicubic(BC) interpolation. The
quantitative results are competitive and consistent with NN, BL, BC
and others.
Abstract: Aspect Oriented Programming promises many
advantages at programming level by incorporating the cross cutting
concerns into separate units, called aspects. Join Points are
distinguishing features of Aspect Oriented Programming as they
define the points where core requirements and crosscutting concerns
are (inter)connected. Currently, there is a problem of multiple
aspects- composition at the same join point, which introduces the
issues like ordering and controlling of these superimposed aspects.
Dynamic strategies are required to handle these issues as early as
possible. State chart is an effective modeling tool to capture dynamic
behavior at high level design. This paper provides methodology to
formulate the strategies for multiple aspect composition at high level,
which helps to better implement these strategies at coding level. It
also highlights the need of designing shared join point at high level,
by providing the solutions of these issues using state chart diagrams
in UML 2.0. High level design representation of shared join points
also helps to implement the designed strategy in systematic way.
Abstract: Using neural network we try to model the unknown function f for given input-output data pairs. The connection strength of each neuron is updated through learning. Repeated simulations of crisp neural network produce different values of weight factors that are directly affected by the change of different parameters. We propose the idea that for each neuron in the network, we can obtain quasi-fuzzy weight sets (QFWS) using repeated simulation of the crisp neural network. Such type of fuzzy weight functions may be applied where we have multivariate crisp input that needs to be adjusted after iterative learning, like claim amount distribution analysis. As real data is subjected to noise and uncertainty, therefore, QFWS may be helpful in the simplification of such complex problems. Secondly, these QFWS provide good initial solution for training of fuzzy neural networks with reduced computational complexity.
Abstract: The need in cognitive radio system for a simple, fast, and independent technique to sense the spectrum occupancy has led to the energy detection approach. Energy detector is known by its dependency on noise variation in the system which is one of its major drawbacks. In this paper, we are aiming to improve its performance by utilizing a weighted collaborative spectrum sensing, it is similar to the collaborative spectrum sensing methods introduced previously in the literature. These weighting methods give more improvement for collaborative spectrum sensing as compared to no weighting case. There is two method proposed in this paper: the first one depends on the channel status between each sensor and the primary user while the second depends on the value of the energy measured in each sensor.
Abstract: This paper proposes a novel solution for optimizing
the size and communication overhead of a distributed multiagent
system without compromising the performance. The proposed approach
addresses the challenges of scalability especially when the
multiagent system is large. A modified spectral clustering technique
is used to partition a large network into logically related clusters.
Agents are assigned to monitor dedicated clusters rather than monitor
each device or node. The proposed scalable multiagent system is
implemented using JADE (Java Agent Development Environment)
for a large power system. The performance of the proposed topologyindependent
decentralized multiagent system and the scalable multiagent
system is compared by comprehensively simulating different
fault scenarios. The time taken for reconfiguration, the overall computational
complexity, and the communication overhead incurred are
computed. The results of these simulations show that the proposed
scalable multiagent system uses fewer agents efficiently, makes faster
decisions to reconfigure when a fault occurs, and incurs significantly
less communication overhead.
Abstract: In this paper, periodic force operation of a wastewater treatment process has been studied for the improved process performance. A previously developed dynamic model for the process is used to conduct the performance analysis. The static version of the model was utilized first to determine the optimal productivity conditions for the process. Then, feed flow rate in terms of dilution rate i.e. (D) is transformed into sinusoidal function. Nonlinear model predictive control algorithm is utilized to regulate the amplitude and period of the sinusoidal function. The parameters of the feed cyclic functions are determined which resulted in improved productivity than the optimal productivity under steady state conditions. The improvement in productivity is found to be marginal and is satisfactory in substrate conversion compared to that of the optimal condition and to the steady state condition, which corresponds to the average value of the periodic function. Successful results were also obtained in the presence of modeling errors and external disturbances.