Abstract: This article demonstrated development of
controlled release system of an NSAID drug, Diclofenac
sodium employing different ratios of Ethyl cellulose.
Diclofenac sodium and ethyl cellulose in different proportions
were processed by microencapsulation based on phase
separation technique to formulate microcapsules. The
prepared microcapsules were then compressed into tablets to
obtain controlled release oral formulations. In-vitro evaluation
was performed by dissolution test of each preparation was
conducted in 900 ml of phosphate buffer solution of pH 7.2
maintained at 37 ± 0.5 °C and stirred at 50 rpm. At predetermined
time intervals (0, 0.5, 1.0, 1.5, 2, 3, 4, 6, 8, 10, 12,
16, 20 and 24 hrs). The drug concentration in the collected
samples was determined by UV spectrophotometer at 276 nm.
The physical characteristics of diclofenac sodium
microcapsules were according to accepted range. These were
off-white, free flowing and spherical in shape. The release
profile of diclofenac sodium from microcapsules was found to
be directly proportional to the proportion of ethylcellulose and
coat thickness. The in-vitro release pattern showed that with
ratio of 1:1 and 1:2 (drug: polymer), the percentage release of
drug at first hour was 16.91 and 11.52 %, respectively as
compared to 1:3 which is only 6.87 % with in this time. The
release mechanism followed higuchi model for its release
pattern. Tablet Formulation (F2) of present study was found
comparable in release profile the marketed brand Phlogin-SR,
microcapsules showed an extended release beyond 24 h.
Further, a good correlation was found between drug release
and proportion of ethylcellulose in the microcapsules.
Microencapsulation based on coacervation found as good
technique to control release of diclofenac sodium for making
the controlled release formulations.
Abstract: Program slicing is the task of finding all statements in
a program that directly or indirectly influence the value of a variable
occurrence. The set of statements that can affect the value of a
variable at some point in a program is called a program backward
slice. In several software engineering applications, such as program
debugging and measuring program cohesion and parallelism, several
slices are computed at different program points. The existing
algorithms for computing program slices are introduced to compute a
slice at a program point. In these algorithms, the program, or the
model that represents the program, is traversed completely or
partially once. To compute more than one slice, the same algorithm
is applied for every point of interest in the program. Thus, the same
program, or program representation, is traversed several times.
In this paper, an algorithm is introduced to compute all forward
static slices of a computer program by traversing the program
representation graph once. Therefore, the introduced algorithm is
useful for software engineering applications that require computing
program slices at different points of a program. The program
representation graph used in this paper is called Program Dependence
Graph (PDG).
Abstract: one of the significant factors for improving the
accuracy of Land Surface Temperature (LST) retrieval is the correct
understanding of the directional anisotropy for thermal radiance. In
this paper, the multiple scattering effect between heterogeneous
non-isothermal surfaces is described rigorously according to the
concept of configuration factor, based on which a directional thermal
radiance model is built, and the directional radiant character for urban
canopy is analyzed. The model is applied to a simple urban canopy
with row structure to simulate the change of Directional Brightness
Temperature (DBT). The results show that the DBT is aggrandized
because of the multiple scattering effects, whereas the change range of
DBT is smoothed. The temperature difference, spatial distribution,
emissivity of the components can all lead to the change of DBT. The
“hot spot" phenomenon occurs when the proportion of high
temperature component in the vision field came to a head. On the other
hand, the “cool spot" phenomena occur when low temperature
proportion came to the head. The “spot" effect disappears only when
the proportion of every component keeps invariability. The model
built in this paper can be used for the study of directional effect on
emissivity, the LST retrieval over urban areas and the adjacency effect
of thermal remote sensing pixels.
Abstract: In this study, an inland metropolitan area, Gwangju, in Korea was selected to assess the amplification potential of earthquake motion and provide the information for regional seismic countermeasure. A geographic information system-based expert system was implemented for reliably predicting the spatial geotechnical layers in the entire region of interesting by building a geo-knowledge database. Particularly, the database consists of the existing boring data gathered from the prior geotechnical projects and the surface geo-knowledge data acquired from the site visit. For practical application of the geo-knowledge database to estimate the earthquake hazard potential related to site amplification effects at the study area, seismic zoning maps on geotechnical parameters, such as the bedrock depth and the site period, were created within GIS framework. In addition, seismic zonation of site classification was also performed to determine the site amplification coefficients for seismic design at any site in the study area. KeywordsEarthquake hazard, geo-knowledge, geographic information system, seismic zonation, site period.
Abstract: Problem-based learning (PBL) is one of the student
centered approaches and has been considered by a number of higher
educational institutions in many parts of the world as a method of
delivery. This paper presents a creative thinking approach for
implementing Problem-based Learning in Mechanics of Structure
within a Malaysian Polytechnics environment. In the learning
process, students learn how to analyze the problem given among the
students and sharing classroom knowledge into practice. Further,
through this course-s emphasis on problem-based learning, students
acquire creative thinking skills and professional skills as they tackle
complex, interdisciplinary and real-situation problems. Once the
creative ideas are generated, there are useful additional techniques
for tender ideas that will grow into a productive concept or solution.
The combination of creative skills and technical abilities will enable
the students to be ready to “hit-the-ground-running" and produce in
industry when they graduate.
Abstract: Structured catalysts formed from the growth of
zeolites on substrates is an area of increasing interest due to the
increased efficiency of the catalytic process, and the ability to
provide superior heat transfer and thermal conductivity for both
exothermic and endothermic processes.
However, the generation of structured catalysts represents a
significant challenge when balancing the relationship variables
between materials properties and catalytic performance, with the
Na2O, H2O and Al2O3 gel composition paying a significant role in
this dynamic, thereby affecting the both the type and range of
application.
The structured catalyst films generated as part of this
investigation have been characterised using a range of techniques,
including X-ray diffraction (XRD), Electron microscopy (SEM),
Energy Dispersive X-ray analysis (EDX) and Thermogravimetric
Analysis (TGA), with the transition from oxide-on-alloy wires to
hydrothermally synthesised uniformly zeolite coated surfaces being
demonstrated using both SEM and XRD. The robustness of the
coatings has been ascertained by subjecting these to thermal cycling
(ambient to 550oC), with the results indicating that the synthesis time
and gel compositions have a crucial effect on the quality of zeolite
growth on the FeCrAlloy wires.
Finally, the activity of the structured catalyst was verified by a
series of comparison experiments with standard zeolite Y catalysts in
powdered pelleted forms.
Abstract: This paper deals with modeling and parameter
identification of nonlinear systems described by Hammerstein model
having Piecewise nonlinear characteristics such as Dead-zone
nonlinearity characteristic. The simultaneous use of both an easy
decomposition technique and the triangular basis functions leads to a
particular form of Hammerstein model. The approximation by using
Triangular basis functions for the description of the static nonlinear
block conducts to a linear regressor model, so that least squares
techniques can be used for the parameter estimation. Singular Values
Decomposition (SVD) technique has been applied to separate the
coupled parameters. The proposed approach has been efficiently
tested on academic examples of simulation.
Abstract: This paper explores the sense of place in the Vredefort Dome World Heritage site, South Africa, as an essential input for the formulation of spatial planning proposals for the area. Intangible aspects such as personal and symbolic meanings of sites are currently not integrated in spatial planning in South Africa. This may have a detrimental effect on local inhabitants who have a long history with the site and built up a strong place identity. Involving local inhabitants at an early stage of the planning process and incorporating their attitudes and opinions in future intervention in the area, may also contribute to the acceptance of the legitimacy of future policy. An interdisciplinary and mixed-method research approach was followed in this study in order to identify possible ways to anchor spatial planning proposals in the identity of the place. In essence, the qualitative study revealed that inhabitants reflect a deep and personal relationship with and within the area, which contributes significantly to their sense of emotional security and selfidentity. Results include a strong conservation-orientated attitude with regard to the natural rural character of the site, especially in the inner core.
Abstract: This paper presents the comparative study of coded
data methods for finding the benefit of concealing the natural data
which is the mercantile secret. Influential parameters of the number
of replicates (rep), treatment effects (τ) and standard deviation (σ)
against the efficiency of each transformation method are investigated.
The experimental data are generated via computer simulations under
the specified condition of the process with the completely
randomized design (CRD). Three ways of data transformation consist
of Box-Cox, arcsine and logit methods. The difference values of F
statistic between coded data and natural data (Fc-Fn) and hypothesis
testing results were determined. The experimental results indicate
that the Box-Cox results are significantly different from natural data
in cases of smaller levels of replicates and seem to be improper when
the parameter of minus lambda has been assigned. On the other hand,
arcsine and logit transformations are more robust and obviously,
provide more precise numerical results. In addition, the alternate
ways to select the lambda in the power transformation are also
offered to achieve much more appropriate outcomes.
Abstract: Solar power plants(SPPs) have shown a lot of good outcomes
in providing a various functions depending on industrial expectations by
deploying ad-hoc networking with helps of light loaded and battery powered
sensor nodes. In particular, it is strongly requested to develop an algorithm to
deriver the sensing data from the end node of solar power plants to the sink node
on time. In this paper, based on the above observation we have proposed an
IEEE802.15.4 based self routing scheme for solar power plants. The proposed
beacon based priority routing Algorithm (BPRA) scheme utilizes beacon
periods in sending message with embedding the high priority data and thus
provides high quality of service(QoS) in the given criteria. The performance
measures are the packet Throughput, delivery, latency, total energy
consumption. Simulation results under TinyOS Simulator(TOSSIM) have
shown the proposed scheme outcome the conventional Ad hoc On-Demand
Distance Vector(AODV) Routing in solar power plants.
Abstract: An overview of the important aspects of managing
and controlling industrial effluent discharges to public sewers namely
sampling, characterization, quantification and legislative controls has
been presented. The findings have been validated by means of a case
study covering three industrial sectors namely, tanning, textile
finishing and food processing industries. Industrial effluents
discharges were found to be best monitored by systematic and
automatic sampling and quantified using water meter readings
corrected for evaporative and consumptive losses. Based on the
treatment processes employed in the public owned treatment works
and the chemical oxygen demand and biochemical oxygen demand
levels obtained, the effluent from all the three industrial sectors
studied were found to lie in the toxic zone. Thus, physico-chemical
treatment of these effluents is required to bring them into the
biodegradable zone. KL values (quoted to base e) were greater than
0.50 day-1 compared to 0.39 day-1 for typical municipality
wastewater.
Abstract: In the highly competitive and rapidly changing global
marketplace, independent organizations and enterprises often come
together and form a temporary alignment of virtual enterprise in a
supply chain to better provide products or service. As firms adopt the
systems approach implicit in supply chain management, they must
manage the quality from both internal process control and external
control of supplier quality and customer requirements. How to
incorporate quality management of upstream and downstream supply
chain partners into their own quality management system has recently
received a great deal of attention from both academic and practice.
This paper investigate the collaborative feature and the entities-
relationship in a supply chain, and presents an ontology of
collaborative supply chain from an approach of aligning
service-oriented framework with service-dominant logic. This
perspective facilitates the segregation of material flow management
from manufacturing capability management, which provides a
foundation for the coordination and integration of the business process
to measure, analyze, and continually improve the quality of products,
services, and process. Further, this approach characterizes the different
interests of supply chain partners, providing an innovative approach to
analyze the collaborative features of supply chain. Furthermore, this
ontology is the foundation to develop quality management system
which internalizes the quality management in upstream and
downstream supply chain partners and manages the quality in supply
chain systematically.
Abstract: In this paper we will develop a sequential life test approach applied to a modified low alloy-high strength steel part used in highway overpasses in Brazil.We will consider two possible underlying sampling distributions: the Normal and theInverse Weibull models. The minimum life will be considered equal to zero. We will use the two underlying models to analyze a fatigue life test situation, comparing the results obtained from both.Since a major chemical component of this low alloy-high strength steel part has been changed, there is little information available about the possible values that the parameters of the corresponding Normal and Inverse Weibull underlying sampling distributions could have. To estimate the shape and the scale parameters of these two sampling models we will use a maximum likelihood approach for censored failure data. We will also develop a truncation mechanism for the Inverse Weibull and Normal models. We will provide rules to truncate a sequential life testing situation making one of the two possible decisions at the moment of truncation; that is, accept or reject the null hypothesis H0. An example will develop the proposed truncated sequential life testing approach for the Inverse Weibull and Normal models.
Abstract: Over the years, there is a growing trend towards
quality-based specifications in highway construction. In many
Quality Control/Quality Assurance (QC/QA) specifications, the
contractor is primarily responsible for quality control of the process,
whereas the highway agency is responsible for testing the acceptance
of the product. A cooperative investigation was conducted in Illinois
over several years to develop a prototype End-Result Specification
(ERS) for asphalt pavement construction. The final characteristics of
the product are stipulated in the ERS and the contractor is given
considerable freedom in achieving those characteristics. The risk for
the contractor or agency depends on how the acceptance limits and
processes are specified. Stochastic simulation models are very useful
in estimating and analyzing payment risk in ERS systems and these
form an integral part of the Illinois-s prototype ERS system. This
paper describes the development of an innovative methodology to
estimate the variability components in in-situ density, air voids and
asphalt content data from ERS projects. The information gained from
this would be crucial in simulating these ERS projects for estimation
and analysis of payment risks associated with asphalt pavement
construction. However, these methods require at least two parties to
conduct tests on all the split samples obtained according to the
sampling scheme prescribed in present ERS implemented in Illinois.
Abstract: A key element of many distribution systems is the
routing and scheduling of vehicles servicing a set of customers. A
wide variety of exact and approximate algorithms have been
proposed for solving the vehicle routing problems (VRP). Exact
algorithms can only solve relatively small problems of VRP, which is
classified as NP-Hard. Several approximate algorithms have proven
successful in finding a feasible solution not necessarily optimum.
Although different parts of the problem are stochastic in nature; yet,
limited work relevant to the application of discrete event system
simulation has addressed the problem. Presented here is optimization
using simulation of VRP; where, a simplified problem has been
developed in the ExtendSimTM simulation environment; where,
ExtendSimTM evolutionary optimizer is used to minimize the total
transportation cost of the problem. Results obtained from the model
are very satisfactory. Further complexities of the problem are
proposed for consideration in the future.
Abstract: Aim of this paper is to explore the prospect of a new approach of mobile phone banking in Libya. This study evaluates customer knowledge on commercial mobile banking in Libya. To examine the relationship between age, occupation and intention for using mobile banking for commercial purpose, a survey was conducted to gather information from one hundred Libyan bank clients. The results indicate that Libyan customers have accepted the new technology and they are ready to use it. There is no significant joint relationship between age and occupation found in intention to use mobile banking in Libya. On the other hand, the customers’ knowledge about mobile banking has a greater relationship with the intention. This study has implications for demographic researches and consumer behaviour disciplines. It also has profitable implications for banks and managers in Libya, as it will assist in better understanding of the Libyan consumers and their activities, when they develop their market strategies and new service.
Abstract: Deoxyribonucleic Acid or DNA computing has
emerged as an interdisciplinary field that draws together chemistry,
molecular biology, computer science and mathematics. Thus, in this
paper, the possibility of DNA-based computing to solve an absolute
1-center problem by molecular manipulations is presented. This is
truly the first attempt to solve such a problem by DNA-based
computing approach. Since, part of the procedures involve with
shortest path computation, research works on DNA computing for
shortest path Traveling Salesman Problem, in short, TSP are reviewed.
These approaches are studied and only the appropriate one is adapted
in designing the computation procedures. This DNA-based
computation is designed in such a way that every path is encoded by
oligonucleotides and the path-s length is directly proportional to the
length of oligonucleotides. Using these properties, gel electrophoresis
is performed in order to separate the respective DNA molecules
according to their length. One expectation arise from this paper is that
it is possible to verify the instance absolute 1-center problem using
DNA computing by laboratory experiments.
Abstract: Sonogram images of normal and lymphocyte thyroid tissues have considerable overlap which makes it difficult to interpret and distinguish. Classification from sonogram images of thyroid gland is tackled in semiautomatic way. While making manual diagnosis from images, some relevant information need not to be recognized by human visual system. Quantitative image analysis could be helpful to manual diagnostic process so far done by physician. Two classes are considered: normal tissue and chronic lymphocyte thyroid (Hashimoto's Thyroid). Data structure is analyzed using K-nearest-neighbors classification. This paper is mentioned that unlike the wavelet sub bands' energy, histograms and Haralick features are not appropriate to distinguish between normal tissue and Hashimoto's thyroid.
Abstract: The Internet and the ever growing applications enable
communities to share and collaborate through common platforms.
However, this growing pattern is not witnessed yet even for elearning.
This paper is based on a doctoral research which aimed at
researching the ways students interact in an online campus and the
supports that they look for and require. Content analysis, based on the
Panchoo/Jaillet methodology, was done on four synchronous
meetings between a tutor and his ten students. The UNIV-Rct ecampus,
analogical to a physical campus, was found to be user
friendly and the students enrolled in a master-s course faced no
difficulties in using it. In addition to the environmental aspects, the
pedagogical implementation of the course has driven the students to
interact and collaborate significantly and this has contributed to
overcome the problems faced by the distance learners. This
completely online model was found to be fruitful in helping distant
learners fight their loneliness and brave their difficulties in a socioconstructivism
approach.
Abstract: This paper aims to present a framework for the
organizational knowledge management, which seeks to deploy a
standardized structure for the integrated management of knowledge is
a common language based on domains, processes and global
indicators inspired by the COBIT framework 5 (ISACA, 2012),
which supports the integration of three technologies, enterprise
information architecture (EIA), the business process modeling (BPM)
and service-oriented architecture (SOA). The Gomak Framework is a
management platform that seeks to integrate the information
technology infrastructure, the structure of applications, information
infrastructure, and business logic and business model to support a
sound strategy of organizational knowledge management, low
process-based approach and concurrent engineering. Concurrent
engineering (CE) is a systematic approach to integrated product
development that respond to customer expectations, involving all
perspectives in parallel, from the beginning of the product life cycle.
(European Space Agency, 2000).