Abstract: Recently, there have been considerable efforts towards the convergence between P2P and Grid computing in order to reach a solution that takes the best of both worlds by exploiting the advantages that each offers. Augmenting the peer-to-peer model to the services of the Grid promises to eliminate bottlenecks and ensure greater scalability, availability, and fault-tolerance. The Grid Information Service (GIS) directly influences quality of service for grid platforms. Most of the proposed solutions for decentralizing the GIS are based on completely flat overlays. The main contributions for this paper are: the investigation of a novel resource discovery framework for Grid implementations based on a hierarchy of structured peer-to-peer overlay networks, and introducing a discovery algorithm utilizing the proposed framework. Validation of the framework-s performance is done via simulation. Experimental results show that the proposed organization has the advantage of being scalable while providing fault-isolation, effective bandwidth utilization, and hierarchical access control. In addition, it will lead to a reliable, guaranteed sub-linear search which returns results within a bounded interval of time and with a smaller amount of generated traffic within each domain.
Abstract: The speculative locking (SL) protocol extends the twophase locking (2PL) protocol to allow for parallelism among conflicting transactions. The adaptive speculative locking (ASL) protocol provided further enhancements and outperformed SL protocols under most conditions. Neither of these protocols consider the impact of network latency on the performance of the distributed database systems. We have studied the performance of ASL protocol taking into account the communication overhead. The results indicate that though system load can counter network latency, it can still become a bottleneck in many situations. The impact of latency on performance depends on many factors including the system resources. A flexible discrete event simulator was used as the testbed for this study.
Abstract: In-core memory requirement is a bottleneck in solving
large three dimensional Navier-Stokes finite element problem
formulations using sparse direct solvers. Out-of-core solution
strategy is a viable alternative to reduce the in-core memory
requirements while solving large scale problems. This study
evaluates the performance of various out-of-core sequential solvers
based on multifrontal or supernodal techniques in the context of
finite element formulations for three dimensional problems on a
Windows platform. Here three different solvers, HSL_MA78,
MUMPS and PARDISO are compared. The performance of these
solvers is evaluated on a 64-bit machine with 16GB RAM for finite
element formulation of flow through a rectangular channel. It is
observed that using out-of-core PARDISO solver, relatively large
problems can be solved. The implementation of Newton and
modified Newton's iteration is also discussed.
Abstract: PT XYZ is a bottled drinking water company. To preserve production resources owned by the company so that the resources could be utilized well, it has implemented maintenance management system, which has important role in company's profitability, and is one of the factors influenced overall company's performance. Yet, up to now the company has never measured maintenance activities' contribution to company's performance. Performance evaluation is done according to adapted Balanced Scorecard model fitted to maintenance function context. This model includes six perspectives: innovation and growth, production, maintenance, environment, costumer, and finance. Actual performance measurement is done through Analytic Hierarchy Process and Objective Matrix. From the research done, we can conclude that the company's maintenance function is categorized in moderate performance. But, there are some indicators which has high priority but low performance, which are: costumers' complain rate, work lateness rate, and Return on Investment.
Abstract: With the increasing number of on-chip components and the critical requirement for processing power, Chip Multiprocessor (CMP) has gained wide acceptance in both academia and industry during the last decade. However, the conventional bus-based onchip communication schemes suffer from very high communication delay and low scalability in large scale systems. Network-on-Chip (NoC) has been proposed to solve the bottleneck of parallel onchip communications by applying different network topologies which separate the communication phase from the computation phase. Observing that the memory bandwidth of the communication between on-chip components and off-chip memory has become a critical problem even in NoC based systems, in this paper, we propose a novel 3D NoC with on-chip Dynamic Random Access Memory (DRAM) in which different layers are dedicated to different functionalities such as processors, cache or memory. Results show that, by using our proposed architecture, average link utilization has reduced by 10.25% for SPLASH-2 workloads. Our proposed design costs 1.12% less execution cycles than the traditional design on average.
Abstract: A procedure for the preparation of clarified Pawpaw
Juice was developed. About 750ml Pawpaw pulp was measured into
2 measuring cylinders A & B of capacity 1 litre heated to 400C,
cooled to 200C. 30mls pectinase was added into cylinder A, while
30mls distilled water was added into cylinder B. Enzyme treated
sample (A) was allowed to digest for 5hours after which it was heated
to 900C for 15 minutes to inactivate the enzyme. The heated sample
was cooled and with the aid of a mucillin cloth the pulp was filtered
to obtain the clarified pawpaw juice. The juice was filled into 100ml
plastic bottles, pasteurized at 950C for 45 minutes, cooled and stored
at room temperature. The sample treated with 30mls distilled water
also underwent the same process. Freshly pasteurized sample was
analyzed for specific gravity, titratable acidity, pH, sugars and
ascorbic acid. The remaining sample was then stored for 2 weeks and
the above analyses repeated. There were differences in the results of
the freshly pasteurized samples and stored sample in pH and ascorbic
acid levels, also sample treated with pectinase yielded higher
volumes of juice than that treated with distilled water.
Abstract: Due to the complex network architecture, the mobile
adhoc network-s multihop feature gives additional problems to the
users. When the traffic load at each node gets increased, the
additional contention due its traffic pattern might cause the nodes
which are close to destination to starve the nodes more away from the
destination and also the capacity of network is unable to satisfy the
total user-s demand which results in an unfairness problem. In this
paper, we propose to create an algorithm to compute the optimal
MAC-layer bandwidth assigned to each flow in the network. The
bottleneck links contention area determines the fair time share which
is necessary to calculate the maximum allowed transmission rate used
by each flow. To completely utilize the network resources, we
compute two optimal rates namely, the maximum fair share and
minimum fair share. We use the maximum fair share achieved in
order to limit the input rate of those flows which crosses the
bottleneck links contention area when the flows that are not allocated
to the optimal transmission rate and calculate the following highest
fair share. Through simulation results, we show that the proposed
protocol achieves improved fair share and throughput with reduced
delay.
Abstract: Knowledge is indispensable but voluminous knowledge becomes a bottleneck for efficient processing. A great challenge for data mining activity is the generation of large number of potential rules as a result of mining process. In fact sometimes result size is comparable to the original data. Traditional data mining pruning activities such as support do not sufficiently reduce the huge rule space. Moreover, many practical applications are characterized by continual change of data and knowledge, thereby making knowledge voluminous with each change. The most predominant representation of the discovered knowledge is the standard Production Rules (PRs) in the form If P Then D. Michalski & Winston proposed Censored Production Rules (CPRs), as an extension of production rules, that exhibit variable precision and supports an efficient mechanism for handling exceptions. A CPR is an augmented production rule of the form: If P Then D Unless C, where C (Censor) is an exception to the rule. Such rules are employed in situations in which the conditional statement 'If P Then D' holds frequently and the assertion C holds rarely. By using a rule of this type we are free to ignore the exception conditions, when the resources needed to establish its presence, are tight or there is simply no information available as to whether it holds or not. Thus the 'If P Then D' part of the CPR expresses important information while the Unless C part acts only as a switch changes the polarity of D to ~D. In this paper a scheme based on Dempster-Shafer Theory (DST) interpretation of a CPR is suggested for discovering CPRs from the discovered flat PRs. The discovery of CPRs from flat rules would result in considerable reduction of the already discovered rules. The proposed scheme incrementally incorporates new knowledge and also reduces the size of knowledge base considerably with each episode. Examples are given to demonstrate the behaviour of the proposed scheme. The suggested cumulative learning scheme would be useful in mining data streams.
Abstract: Model-based approaches have been applied successfully
to a wide range of tasks such as specification, simulation, testing, and
diagnosis. But one bottleneck often prevents the introduction of these
ideas: Manual modeling is a non-trivial, time-consuming task.
Automatically deriving models by observing and analyzing running
systems is one possible way to amend this bottleneck. To
derive a model automatically, some a-priori knowledge about the
model structure–i.e. about the system–must exist. Such a model
formalism would be used as follows: (i) By observing the network
traffic, a model of the long-term system behavior could be generated
automatically, (ii) Test vectors can be generated from the model,
(iii) While the system is running, the model could be used to diagnose
non-normal system behavior.
The main contribution of this paper is the introduction of a model
formalism called 'probabilistic regression automaton' suitable for the
tasks mentioned above.
Abstract: This study investigated the number of Aedes larvae,
the key breeding sites of Aedes sp., and the relationship between
climatic factors and the incidence of DHF in Samui Islands. We
conducted our questionnaire and larval surveys from randomly
selected 105 households in Samui Islands in July-September 2006.
Pearson-s correlation coefficient was used to explore the primary
association between the DHF incidence and all climatic factors.
Multiple stepwise regression technique was then used to fit the
statistical model. The results showed that the positive indoor
containers were small jars, cement tanks, and plastic tanks. The
positive outdoor containers were small jars, cement tanks, plastic
tanks, used cans, tires, plastic bottles, discarded objects, pot saucers,
plant pots, and areca husks. All Ae. albopictus larval indices (i.e., CI,
HI, and BI) were higher than Ae. aegypti larval indices in this area.
These larval indices were higher than WHO standard. This indicated
a high risk of DHF transmission at Samui Islands. The multiple
stepwise regression model was y = –288.80 + 11.024xmean temp. The
mean temperature was positively associated with the DHF incidence
in this area.