Optimized Facial Features-based Age Classification

The evaluation and measurement of human body dimensions are achieved by physical anthropometry. This research was conducted in view of the importance of anthropometric indices of the face in forensic medicine, surgery, and medical imaging. The main goal of this research is to optimization of facial feature point by establishing a mathematical relationship among facial features and used optimize feature points for age classification. Since selected facial feature points are located to the area of mouth, nose, eyes and eyebrow on facial images, all desire facial feature points are extracted accurately. According this proposes method; sixteen Euclidean distances are calculated from the eighteen selected facial feature points vertically as well as horizontally. The mathematical relationships among horizontal and vertical distances are established. Moreover, it is also discovered that distances of the facial feature follows a constant ratio due to age progression. The distances between the specified features points increase with respect the age progression of a human from his or her childhood but the ratio of the distances does not change (d = 1 .618 ) . Finally, according to the proposed mathematical relationship four independent feature distances related to eight feature points are selected from sixteen distances and eighteen feature point-s respectively. These four feature distances are used for classification of age using Support Vector Machine (SVM)-Sequential Minimal Optimization (SMO) algorithm and shown around 96 % accuracy. Experiment result shows the proposed system is effective and accurate for age classification.

Microalbuminuria in Human Immunodeficiency Virus Infection and Acquired Immunodeficiency Syndrome

Human immunodeficiency virus infection and acquired immunodeficiency syndrome is a global pandemic with cases reporting from virtually every country and continues to be a common infection in developing country like India. Microalbuminuria is a manifestation of human immunodeficiency virus associated nephropathy. Therefore, microalbuminuria may be an early marker of human immunodeficiency virus associated nephropathy, and screening for its presence may be beneficial. A strikingly high prevalence of microalbuminuria among human immunodeficiency virus infected patients has been described in various studies. Risk factors for clinically significant proteinuria include African - American race, higher human immunodeficiency virus ribonucleic acid level and lower CD4 lymphocyte count. The cardiovascular risk factors of increased systolic blood pressure and increase fasting blood sugar level are strongly associated with microalbuminuria in human immunodeficiency virus patient. These results suggest that microalbuminuria may be a sign of current endothelial dysfunction and micro-vascular disease and there is substantial risk of future cardiovascular disease events. Positive contributing factors include early kidney disease such as human immunodeficiency virus associated nephropathy, a marker of end organ damage related to co morbidities of diabetes or hypertension, or more diffuse endothelial cells dysfunction. Nevertheless after adjustment for non human immunodeficiency virus factors, human immunodeficiency virus itself is a major risk factor. The presence of human immunodeficiency virus infection is independent risk to develop microalbuminuria in human immunodeficiency virus patient. Cardiovascular risk factors appeared to be stronger predictors of microalbuminuria than markers of human immunodeficiency virus severity person with human immunodeficiency virus infection and microalbuminuria therefore appear to potentially bear the burden of two separate damage related to known vascular end organ damage related to know vascular risk factors, and human immunodeficiency virus specific processes such as the direct viral infection of kidney cells.The higher prevalence of microalbuminuria among the human immunodeficiency virus infected could be harbinger of future increased risks of both kidney and cardiovascular disease. Further study defining the prognostic significance of microalbuminuria among human immunodeficiency virus infected persons will be essential. Microalbuminuria seems to be a predictor of cardiovascular disease in diabetic and non diabetic subjects, hence it can also be used for early detection of micro vascular disease in human immunodeficiency virus positive patients, thus can help to diagnose the disease at the earliest.

Structural and Electronic Characterization of Supported Ni and Au Catalysts used in Environment Protection Determined by XRD,XAS and XPS methods

The nickel and gold nanoclusters as supported catalysts were analyzed by XAS, XRD and XPS in order to determine their local, global and electronic structure. The present study has pointed out a strong deformation of the local structure of the metal, due to its interaction with oxide supports. The average particle size, the mean squares of the microstrain, the particle size distribution and microstrain functions of the supported Ni and Au catalysts were determined by XRD method using Generalized Fermi Function for the X-ray line profiles approximation. Based on EXAFS analysis we consider that the local structure of the investigated systems is strongly distorted concerning the atomic number pairs. Metal-support interaction is confirmed by the shape changes of the probability densities of electron transitions: Ni K edge (1s → continuum and 2p), Au LIII-edge (2p3/2 → continuum, 6s, 6d5/2 and 6d3/2). XPS investigations confirm the metal-support interaction at their interface.

Turbo-Coded Mobile Terrestrial Communication Systems in Urban and Suburban Areas for Wireless Multimedia Applications

With the rapid popularization of internet services, it is apparent that the next generation terrestrial communication systems must be capable of supporting various applications like voice, video, and data. This paper presents the performance evaluation of turbo- coded mobile terrestrial communication systems, which are capable of providing high quality services for delay sensitive (voice or video) and delay tolerant (text transmission) multimedia applications in urban and suburban areas. Different types of multimedia information require different service qualities, which are generally expressed in terms of a maximum acceptable bit-error-rate (BER) and maximum tolerable latency. The breakthrough discovery of turbo codes allows us to significantly reduce the probability of bit errors with feasible latency. In a turbo-coded system, a trade-off between latency and BER results from the choice of convolutional component codes, interleaver type and size, decoding algorithm, and the number of decoding iterations. This trade-off can be exploited for multimedia applications by using optimal and suboptimal performance parameter amalgamations to achieve different service qualities. The results are therefore proposing an adaptive framework for turbo-coded wireless multimedia communications which incorporate a set of performance parameters that achieve an appropriate set of service qualities, depending on the application's requirements.

The Effects of Software Size on Development Effort and Software Quality

Effective evaluation of software development effort is an important issue during project plan. This study provides a model to predict development effort based on the software size estimated with function points. We generalize the average amount of effort spent on each phase of the development, and give the estimates for the effort used in software building, testing, and implementation. Finally, this paper finds a strong correlation between software defects and software size. As the size of software constantly increases, the quality remains to be a matter which requires major concern.

Dynamic-Stochastic Influence Diagrams: Integrating Time-Slices IDs and Discrete Event Systems Modeling

The Influence Diagrams (IDs) is a kind of Probabilistic Belief Networks for graphic modeling. The usage of IDs can improve the communication among field experts, modelers, and decision makers, by showing the issue frame discussed from a high-level point of view. This paper enhances the Time-Sliced Influence Diagrams (TSIDs, or called Dynamic IDs) based formalism from a Discrete Event Systems Modeling and Simulation (DES M&S) perspective, for Exploring Analysis (EA) modeling. The enhancements enable a modeler to specify times occurred of endogenous events dynamically with stochastic sampling as model running and to describe the inter- influences among them with variable nodes in a dynamic situation that the existing TSIDs fails to capture. The new class of model is named Dynamic-Stochastic Influence Diagrams (DSIDs). The paper includes a description of the modeling formalism and the hiberarchy simulators implementing its simulation algorithm, and shows a case study to illustrate its enhancements.

Effects of Thread Dimensions of Functionally Graded Dental Implants on Stress Distribution

In this study, stress distributions on dental implants made of functionally graded biomaterials (FGBM) are investigated numerically. The implant body is considered to be subjected to axial compression loads. Numerical problem is assumed to be 2D, and ANSYS commercial software is used for the analysis. The cross section of the implant thread varies as varying the height (H) and the width (t) of the thread. According to thread dimensions of implant and material properties of FGBM, equivalent stress distribution on the implant is determined and presented with contour plots along with the maximum equivalent stress values. As a result, with increasing material gradient parameter (n), the equivalent stress decreases, but the minimum stress distribution increases. Maximum stress values decrease with decreasing implant radius (r). Maximum von Mises stresses increases with decreasing H when t is constant. On the other hand, the stress values are not affected by variation of t in the case of H = constant.

A New Objective Weight on Interval Type-2 Fuzzy Sets

The design of weight is one of the important parts in fuzzy decision making, as it would have a deep effect on the evaluation results. Entropy is one of the weight measure based on objective evaluation. Non--probabilistic-type entropy measures for fuzzy set and interval type-2 fuzzy sets (IT2FS) have been developed and applied to weight measure. Since the entropy for (IT2FS) for decision making yet to be explored, this paper proposes a new objective weight method by using entropy weight method for multiple attribute decision making (MADM). This paper utilizes the nature of IT2FS concept in the evaluation process to assess the attribute weight based on the credibility of data. An example was presented to demonstrate the feasibility of the new method in decision making. The entropy measure of interval type-2 fuzzy sets yield flexible judgment and could be applied in decision making environment.

Development of a Porous Silica Film by Sol-gel Process

In the present work homogeneous silica film on silicon was fabricated by colloidal silica sol. The silica sol precursor with uniformly granular particle was derived by the alkaline hydrolysis of tetraethoxyorthosilicate (TEOS) in presence of glycerol template. The film was prepared by dip coating process. The templated hetero-structured silica film was annealed at elevated temperatures to generate nano- and meso porosity in the film. The film was subsequently annealed at different temperatures to make it defect free and abrasion resistant. The sol and the film were characterized by the measurement of particle size distribution, scanning electron microscopy, XRD, FTIR spectroscopy, transmission electron microscopy, atomic force microscopy, measurement of the refractive index, thermal conductivity and abrasion resistance. The porosity of the films decreased whereas refractive index and dielectric constant of it `increased with the increase in the annealing temperature. The thermal conductivity of the films increased with the increase in the film thickness. The developed porous silica film holds strong potential for use in different areas.

Feature Point Reduction for Video Stabilization

Corner detection and optical flow are common techniques for feature-based video stabilization. However, these algorithms are computationally expensive and should be performed at a reasonable rate. This paper presents an algorithm for discarding irrelevant feature points and maintaining them for future use so as to improve the computational cost. The algorithm starts by initializing a maintained set. The feature points in the maintained set are examined against its accuracy for modeling. Corner detection is required only when the feature points are insufficiently accurate for future modeling. Then, optical flows are computed from the maintained feature points toward the consecutive frame. After that, a motion model is estimated based on the simplified affine motion model and least square method, with outliers belonging to moving objects presented. Studentized residuals are used to eliminate such outliers. The model estimation and elimination processes repeat until no more outliers are identified. Finally, the entire algorithm repeats along the video sequence with the points remaining from the previous iteration used as the maintained set. As a practical application, an efficient video stabilization can be achieved by exploiting the computed motion models. Our study shows that the number of times corner detection needs to perform is greatly reduced, thus significantly improving the computational cost. Moreover, optical flow vectors are computed for only the maintained feature points, not for outliers, thus also reducing the computational cost. In addition, the feature points after reduction can sufficiently be used for background objects tracking as demonstrated in the simple video stabilizer based on our proposed algorithm.

Learning Objects: A New Paradigm for ELearning Resource Development for Secondary Schools in Tanzania

The Information and Communication Technologies (ICTs), and the Wide World Web (WWW) have fundamentally altered the practice of teaching and learning world wide. Many universities, organizations, colleges and schools are trying to apply the benefits of the emerging ICT. In the early nineties the term learning object was introduced into the instructional technology vernacular; the idea being that educational resources could be broken into modular components for later combination by instructors, learners, and eventually computes into larger structures that would support learning [1]. However in many developing countries, the use of ICT is still in its infancy stage and the concept of learning object is quite new. This paper outlines the learning object design considerations for developing countries depending on learning environment.

Wetting Behavior of Reactive and Non–Reactive Wetting of Liquids on Metallic Substrates

Wetting characteristics of reactive (Sn–0.7Cu solder) and non– reactive (castor oil) wetting of liquids on Cu and Ag plated Al substrates have been investigated. Solder spreading exhibited capillary, gravity and viscous regimes. Oils did not exhibit noticeable spreading regimes. Solder alloy showed better wettability on Ag coated Al substrate compared to Cu plating. In the case of castor oil, Cu coated Al substrate exhibited good wettability as compared to Ag coated Al substrates. The difference in wettability during reactive wetting of solder and non–reactive wetting of oils is attributed to the change in the surface energies of Al substrates brought about by the formation of intermetallic compounds (IMCs).

Optimization of Conditions for Xanthan Gum Production from Waste Date in Submerged Fermantation

Xanthan gum is one of the major commercial biopolymers. Due to its excellent rheological properties xanthan gum is used in many applications, mainly in food industry. Commercial production of xanthan gum uses glucose as the carbon substrate; consequently the price of xanthan production is high. One of the ways to decrease xanthan price, is using cheaper substrate like agricultural wastes. Iran is one of the biggest date producer countries. However approximately 50% of date production is wasted annually. The goal of this study is to produce xanthan gum from waste date using Xanthomonas campestris PTCC1473 by submerged fermentation. In this study the effect of three variables including phosphor and nitrogen amount and agitation rate in three levels using response surface methodology (RSM) has been studied. Results achieved from statistical analysis Design Expert 7.0.0 software showed that xanthan increased with increasing level of phosphor. Low level of nitrogen leaded to higher xanthan production. Xanthan amount, increasing agitation had positive influence. The statistical model identified the optimum conditions nitrogen amount=3.15g/l, phosphor amount=5.03 g/l and agitation=394.8 rpm for xanthan. To model validation, experiments in optimum conditions for xanthan gum were carried out. The mean of result for xanthan was 6.72±0.26. The result was closed to the predicted value by using RSM.

Structural Integrity Management for Fixed Offshore Platforms in Malaysia

Structural Integrity Management (SIM) is important for the protection of offshore crew, environment, business assets and company and industry reputation. API RP 2A contained guidelines for assessment of existing platforms mostly for the Gulf of Mexico (GOM). ISO 19902 SIM framework also does not specifically cater for Malaysia. There are about 200 platforms in Malaysia with 90 exceeding their design life. The Petronas Carigali Sdn Bhd (PCSB) uses the Asset Integrity Management System and the very subjective Risk based Inspection Program for these platforms. Petronas currently doesn-t have a standalone Petronas Technical Standard PTS-SIM. This study proposes a recommended practice for the SIM process for offshore structures in Malaysia, including studies by API and ISO and local elements such as the number of platforms, types of facilities, age and risk ranking. Case study on SMG-A platform in Sabah shows missing or scattered platform data and a gap in inspection history. It is to undergo a level 3 underwater inspection in year 2015.

Fragile Watermarking for Color Images Using Thresholding Technique

In this paper, we propose ablock-wise watermarking scheme for color image authentication to resist malicious tampering of digital media. The thresholding technique is incorporated into the scheme such that the tampered region of the color image can be recovered with high quality while the proofing result is obtained. The watermark for each block consists of its dual authentication data and the corresponding feature information. The feature information for recovery iscomputed bythe thresholding technique. In the proofing process, we propose a dual-option parity check method to proof the validity of image blocks. In the recovery process, the feature information of each block embedded into the color image is rebuilt for high quality recovery. The simulation results show that the proposed watermarking scheme can effectively proof the tempered region with high detection rate and can recover the tempered region with high quality.

Process-Oriented Learning Requirements for Employees and for Organizations

Using activity theory, organisational theory and didactics as theoretical foundations, a comprehensive model of the organisational dimensions relevant for learning and knowledge transfer will be developed. In a second step, a Learning Assessment Guideline will be elaborated. This guideline will be designed to permit a targeted analysis of organisations to identify the status quo in those areas crucial to the implementation of learning and knowledge transfer. In addition, this self-analysis tool will enable learning managers to select adequate didactic models for e- and blended learning. As part of the European Integrated Project "Process-oriented Learning and Information Exchange" (PROLIX), this model of organisational prerequisites for learning and knowledge transfer will be empirically tested in four profit and non-profit organisations in Great Britain, Germany and France (to be finalized in autumn 2006). The findings concern not only the capability of the model of organisational dimensions, but also the predominant perceptions of and obstacles to learning in organisations.

Fingerprint Compression Using Multiwavelets

Large volumes of fingerprints are collected and stored every day in a wide range of applications, including forensics, access control etc. It is evident from the database of Federal Bureau of Investigation (FBI) which contains more than 70 million finger prints. Compression of this database is very important because of this high Volume. The performance of existing image coding standards generally degrades at low bit-rates because of the underlying block based Discrete Cosine Transform (DCT) scheme. Over the past decade, the success of wavelets in solving many different problems has contributed to its unprecedented popularity. Due to implementation constraints scalar wavelets do not posses all the properties which are needed for better performance in compression. New class of wavelets called 'Multiwavelets' which posses more than one scaling filters overcomes this problem. The objective of this paper is to develop an efficient compression scheme and to obtain better quality and higher compression ratio through multiwavelet transform and embedded coding of multiwavelet coefficients through Set Partitioning In Hierarchical Trees algorithm (SPIHT) algorithm. A comparison of the best known multiwavelets is made to the best known scalar wavelets. Both quantitative and qualitative measures of performance are examined for Fingerprints.

SVM-based Multiview Face Recognition by Generalization of Discriminant Analysis

Identity verification of authentic persons by their multiview faces is a real valued problem in machine vision. Multiview faces are having difficulties due to non-linear representation in the feature space. This paper illustrates the usability of the generalization of LDA in the form of canonical covariate for face recognition to multiview faces. In the proposed work, the Gabor filter bank is used to extract facial features that characterized by spatial frequency, spatial locality and orientation. Gabor face representation captures substantial amount of variations of the face instances that often occurs due to illumination, pose and facial expression changes. Convolution of Gabor filter bank to face images of rotated profile views produce Gabor faces with high dimensional features vectors. Canonical covariate is then used to Gabor faces to reduce the high dimensional feature spaces into low dimensional subspaces. Finally, support vector machines are trained with canonical sub-spaces that contain reduced set of features and perform recognition task. The proposed system is evaluated with UMIST face database. The experiment results demonstrate the efficiency and robustness of the proposed system with high recognition rates.

Strategies of Entrepreneurs to Collaborate with Alliances for Commercializing Technology and New Product Innovation: A Practical Learning in Thailand

This paper provides a key driver-based conceptual framework that can be used to improve a firm-s success in commercializing technology and in new product innovation resulting from collaboration with other organizations through strategic alliances. Based on a qualitative study using an interview approach, strategic alliances of entrepreneurs in the food processing industry in Thailand are explored. This paper describes factors affecting decisions to collaborate through alliances. It identifies four issues: maintaining the efficiency of the value chain for production capability, adapting to present and future competition, careful assessment of value of outcomes, and management of innovation. We consider five driving factors: resource orientation, assessment of risk, business opportunity, sharing of benefits and confidence in alliance partners. These factors will be of interest to entrepreneurs and policy makers with regard to further understanding of the direction of business strategies.

Study of Heat Transfer in the Poly Ethylene Fluidized Bed Reactor Numerically and Experimentally

In this research, heat transfer of a poly Ethylene fluidized bed reactor without reaction were studied experimentally and computationally at different superficial gas velocities. A multifluid Eulerian computational model incorporating the kinetic theory for solid particles was developed and used to simulate the heat conducting gas–solid flows in a fluidized bed configuration. Momentum exchange coefficients were evaluated using the Syamlal– O-Brien drag functions. Temperature distributions of different phases in the reactor were also computed. Good agreement was found between the model predictions and the experimentally obtained data for the bed expansion ratio as well as the qualitative gas–solid flow patterns. The simulation and experimental results showed that the gas temperature decreases as it moves upward in the reactor, while the solid particle temperature increases. Pressure drop and temperature distribution predicted by the simulations were in good agreement with the experimental measurements at superficial gas velocities higher than the minimum fluidization velocity. Also, the predicted time-average local voidage profiles were in reasonable agreement with the experimental results. The study showed that the computational model was capable of predicting the heat transfer and the hydrodynamic behavior of gas-solid fluidized bed flows with reasonable accuracy.