Face Texture Reconstruction for Illumination Variant Face Recognition

In illumination variant face recognition, existing methods extracting face albedo as light normalized image may lead to loss of extensive facial details, with light template discarded. To improve that, a novel approach for realistic facial texture reconstruction by combining original image and albedo image is proposed. First, light subspaces of different identities are established from the given reference face images; then by projecting the original and albedo image into each light subspace respectively, texture reference images with corresponding lighting are reconstructed and two texture subspaces are formed. According to the projections in texture subspaces, facial texture with normal light can be synthesized. Due to the combination of original image, facial details can be preserved with face albedo. In addition, image partition is applied to improve the synthesization performance. Experiments on Yale B and CMUPIE databases demonstrate that this algorithm outperforms the others both in image representation and in face recognition.

Power and Delay Optimized Graph Representation for Combinational Logic Circuits

Structural representation and technology mapping of a Boolean function is an important problem in the design of nonregenerative digital logic circuits (also called combinational logic circuits). Library aware function manipulation offers a solution to this problem. Compact multi-level representation of binary networks, based on simple circuit structures, such as AND-Inverter Graphs (AIG) [1] [5], NAND Graphs, OR-Inverter Graphs (OIG), AND-OR Graphs (AOG), AND-OR-Inverter Graphs (AOIG), AND-XORInverter Graphs, Reduced Boolean Circuits [8] does exist in literature. In this work, we discuss a novel and efficient graph realization for combinational logic circuits, represented using a NAND-NOR-Inverter Graph (NNIG), which is composed of only two-input NAND (NAND2), NOR (NOR2) and inverter (INV) cells. The networks are constructed on the basis of irredundant disjunctive and conjunctive normal forms, after factoring, comprising terms with minimum support. Construction of a NNIG for a non-regenerative function in normal form would be straightforward, whereas for the complementary phase, it would be developed by considering a virtual instance of the function. However, the choice of best NNIG for a given function would be based upon literal count, cell count and DAG node count of the implementation at the technology independent stage. In case of a tie, the final decision would be made after extracting the physical design parameters. We have considered AIG representation for reduced disjunctive normal form and the best of OIG/AOG/AOIG for the minimized conjunctive normal forms. This is necessitated due to the nature of certain functions, such as Achilles- heel functions. NNIGs are found to exhibit 3.97% lesser node count compared to AIGs and OIG/AOG/AOIGs; consume 23.74% and 10.79% lesser library cells than AIGs and OIG/AOG/AOIGs for the various samples considered. We compare the power efficiency and delay improvement achieved by optimal NNIGs over minimal AIGs and OIG/AOG/AOIGs for various case studies. In comparison with functionally equivalent, irredundant and compact AIGs, NNIGs report mean savings in power and delay of 43.71% and 25.85% respectively, after technology mapping with a 0.35 micron TSMC CMOS process. For a comparison with OIG/AOG/AOIGs, NNIGs demonstrate average savings in power and delay by 47.51% and 24.83%. With respect to device count needed for implementation with static CMOS logic style, NNIGs utilize 37.85% and 33.95% lesser transistors than their AIG and OIG/AOG/AOIG counterparts.

A Maximum Parsimony Model to Reconstruct Phylogenetic Network in Honey Bee Evolution

Phylogenies ; The evolutionary histories of groups of species are one of the most widely used tools throughout the life sciences, as well as objects of research with in systematic, evolutionary biology. In every phylogenetic analysis reconstruction produces trees. These trees represent the evolutionary histories of many groups of organisms, bacteria due to horizontal gene transfer and plants due to process of hybridization. The process of gene transfer in bacteria and hybridization in plants lead to reticulate networks, therefore, the methods of constructing trees fail in constructing reticulate networks. In this paper a model has been employed to reconstruct phylogenetic network in honey bee. This network represents reticulate evolution in honey bee. The maximum parsimony approach has been used to obtain this reticulate network.

Strengthening the HCI Approaches in the Software Development Process

User-Centered Design (UCD), Usability Engineering (UE) and Participatory Design (PD) are the common Human- Computer Interaction (HCI) approaches that are practiced in the software development process, focusing towards issues and matters concerning user involvement. It overlooks the organizational perspective of HCI integration within the software development organization. The Management Information Systems (MIS) perspective of HCI takes a managerial and organizational context to view the effectiveness of integrating HCI in the software development process. The Human-Centered Design (HCD) which encompasses all of the human aspects including aesthetic and ergonomic, is claimed as to provide a better approach in strengthening the HCI approaches to strengthen the software development process. In determining the effectiveness of HCD in the software development process, this paper presents the findings of a content analysis of HCI approaches by viewing those approaches as a technology which integrates user requirements, ranging from the top management to other stake holder in the software development process. The findings obtained show that HCD approach is a technology that emphasizes on human, tools and knowledge in strengthening the HCI approaches to strengthen the software development process in the quest to produce a sustainable, usable and useful software product.

Impact of Customer Management System in Improving Customer Retention: Optimization of Negative Customer Feedback

Complaints today have the ability to retain customer loyalty using state of the art systems and strategies in customer relationship management to analyze and respond to a plethora of customer perception. The Majority of companies are not aware of the beneficiary utilization of customer complaints for the sake of quality improvements. Also, some companies have problems determining how resolution of complaints can be profitable. In this study, we will define the problems and ascertain the importance of customer management system on the companies. Furthermore, we will determine the impact of such a system on efficiency, confidence, profitability and customer complaints. Eventually, we will develop methods and address the issues. In this paper, we used an open-ended questionnaire and distributed that to 30 randomly chosen respondents which were the passengers in an airport. We also define three hypotheses for our study and we will validate each of them. Then using frequency, Chi- Square and quality control method we optimized the size of customers- negative feedback and improved the process of customer retention.

Estimation of Systolic and Diastolic Pressure using the Pulse Transit Time

In this paper, algorithm estimating the blood pressure was proposed using the pulse transit time (PTT) as a more convenient method of measuring the blood pressure. After measuring ECG and pressure pulse, and photoplethysmography, the PTT was calculated from the acquired signals. Thereafter, the system to indirectly measure the systolic pressure and the diastolic pressure was composed using the statistic method. In comparison between the blood pressure indirectly measured by proposed algorithm estimating the blood pressure and real blood pressure measured by conventional sphygmomanometer, the systolic pressure indicates the mean error of ±3.24mmHg and the standard deviation of 2.53mmHg, while the diastolic pressure indicates the satisfactory result, that is, the mean error of ±1.80mmHg and the standard deviation of 1.39mmHg. These results are satisfied with the regulation of ANSI/AAMI for certification of sphygmomanometer that real measurement error value should be within the mean error of ±5mmHg and the standard deviation of 8mmHg. These results are suggest the possibility of applying to portable and long time blood pressure monitoring system hereafter.

Harmful Effect of Ambient Ozone on Growth and Productivity of Two Legume Crops Visia Faba, and Pisum sativum in Riyadh City, K.S.A.

Ozone (O3) is considered as one of the most phytotoxic pollutants with deleterious effects on living and non living components of Ecosystems. It reduces growth and yield of many crops as well as alters the physiology and crop quality. The present study described series of experiments to investigate the effects of ambient O3 at different locations with different ambient levels of O3 depending on proximity to pollutant source and ranged between 17 ppb/h in control experiment to 112 ppb/h in industrial area respectively. The ambient levels in other three locations (King Saud University botanical garden, King Fahd Rd, and Almanakh Garden) were 61,61,77 ppb/h respectively. Tow legume crops species (vicia vaba L ; and Pisum sativum) differ in their phenology and sensitivity were used. The results showed a significant negative effect to ozone on morphology, number of injured leaves, growth and productivity with a difference in the degree of response depending on the plant type. Visia Faba showed sensitivity to ozone to number and leaf area and the degree of injury leaves 3, pisum sativum show higher sensitivity for the gas for degree of injury 1,The relative growth rate and seed weight, it turns out there is no significant difference between the two plants in plant height and number of seeds.

Knowledge Sharing: A Survey, Assessment and Directions for Future Research: Individual Behavior Perspective

One of the most important areas of knowledge management studies is knowledge sharing. Measured in terms of number of scientific articles and organization-s applications, knowledge sharing stands as an example of success in the field. This paper reviews the related papers in the context of the underlying individual behavioral variables to providea direction framework for future research and writing.

Solving Partially Monotone Problems with Neural Networks

In many applications, it is a priori known that the target function should satisfy certain constraints imposed by, for example, economic theory or a human-decision maker. Here we consider partially monotone problems, where the target variable depends monotonically on some of the predictor variables but not all. We propose an approach to build partially monotone models based on the convolution of monotone neural networks and kernel functions. The results from simulations and a real case study on house pricing show that our approach has significantly better performance than partially monotone linear models. Furthermore, the incorporation of partial monotonicity constraints not only leads to models that are in accordance with the decision maker's expertise, but also reduces considerably the model variance in comparison to standard neural networks with weight decay.

FRC – A New Sustainable Option for Construction to Mitigate Earthquakes

Ten simply supported grossly underreinforced tapered concrete beams of full size were tested upto complete collapse under flexural effect .Out of 10 beams, 5 beams were nonfibrous and the remaining beams contained fibres. The beams had a variation in the tapered angle as 2°, 4°, 6°, 8° and 10°. The concrete mix, conventional steel and the type of fibre used were held constant. Flat corrugated steel fibres were utilized as secondary reinforcement. The strength and stability parameters were measured. It is established that the fibrous tapered beams can be used economically in earthquake prone areas.

Effect of Calcination Temperature and MgO Crystallite Size on MgO/TiO2 Catalyst System for Soybean Transesterification

The effect of calcination temperature and MgO crystallite sizes on the structure and catalytic performance of TiO2 supported nano-MgO catalyst for the trans-esterification of soybean oil has been studied. The catalyst has been prepared by deposition precipitation method, characterised by XRD and FTIR and tested in an autoclave at 225oC. The soybean oil conversion after 15 minutes of the trans-esterification reaction increased when the calcination temperature was increased from 500 to 600oC and decreased with further increase in calcination temperature. Some glycerolysis activity was also detected on catalysts calcined at 600 and 700oC after 45 minutes of reaction. The trans-esterification reaction rate increased with the decrease in MgO crystallite size for the first 30 min.

Software Effort Estimation Using Soft Computing Techniques

Various models have been derived by studying large number of completed software projects from various organizations and applications to explore how project sizes mapped into project effort. But, still there is a need to prediction accuracy of the models. As Neuro-fuzzy based system is able to approximate the non-linear function with more precision. So, Neuro-Fuzzy system is used as a soft computing approach to generate model by formulating the relationship based on its training. In this paper, Neuro-Fuzzy technique is used for software estimation modeling of on NASA software project data and performance of the developed models are compared with the Halstead, Walston-Felix, Bailey-Basili and Doty Models mentioned in the literature.

Modelling Indoor Air Carbon Dioxide (CO2)Concentration using Neural Network

The use of neural networks is popular in various building applications such as prediction of heating load, ventilation rate and indoor temperature. Significant is, that only few papers deal with indoor carbon dioxide (CO2) prediction which is a very good indicator of indoor air quality (IAQ). In this study, a data-driven modelling method based on multilayer perceptron network for indoor air carbon dioxide in an apartment building is developed. Temperature and humidity measurements are used as input variables to the network. Motivation for this study derives from the following issues. First, measuring carbon dioxide is expensive and sensors power consumptions is high and secondly, this leads to short operating times of battery-powered sensors. The results show that predicting CO2 concentration based on relative humidity and temperature measurements, is difficult. Therefore, more additional information is needed.

Security Enhanced RFID Middleware System

Recently, the RFID (Radio Frequency Identification) technology attracts the world market attention as essential technology for ubiquitous environment. The RFID market has focused on transponders and reader development. But that concern has shifted to RFID software like as high-valued e-business applications, RFID middleware and related development tools. However, due to the high sensitivity of data and service transaction within the RFID network, security consideration must be addressed. In order to guarantee trusted e-business based on RFID technology, we propose a security enhanced RFID middleware system. Our proposal is compliant with EPCglobal ALE (Application Level Events), which is standard interface for middleware and its clients. We show how to provide strengthened security and trust by protecting transported data between middleware and its client, and stored data in middleware. Moreover, we achieve the identification and service access control against illegal service abuse. Our system enables secure RFID middleware service and trusted e-business service.

Generating Frequent Patterns through Intersection between Transactions

The problem of frequent itemset mining is considered in this paper. One new technique proposed to generate frequent patterns in large databases without time-consuming candidate generation. This technique is based on focusing on transaction instead of concentrating on itemset. This algorithm based on take intersection between one transaction and others transaction and the maximum shared items between transactions computed instead of creating itemset and computing their frequency. With applying real life transactions and some consumption is taken from real life data, the significant efficiency acquire from databases in generation association rules mining.

A Study on using N-Pattern Chains of Design Patterns based on Software Quality Metrics

Design patterns describe good solutions to common and reoccurring problems in program design. Applying design patterns in software design and implementation have significant effects on software quality metrics such as flexibility, usability, reusability, scalability and robustness. There is no standard rule for using design patterns. There are some situations that a pattern is applied for a specific problem and this pattern uses another pattern. In this paper, we study the effect of using chain of patterns on software quality metrics.

A CBR System to New Product Development: An Application for Hearing Devices Design

Nowadays, quick technological changes force companies to develop innovative products in an increasingly competitive environment. Therefore, how to enhance the time of new product development is very important. This design problem often lacks the exact formula for getting it, and highly depends upon human designers- past experiences. For these reasons, in this work, a Casebased reasoning (CBR) system to assist in new product development is proposed. When a case is recovered from the case base, the system will take into account not only the attribute-s specific value and how important it is. It will also take into account if the attribute has a positive influence over the product development. Hence the manufacturing time will be improved. This information will be introduced as a new concept called “adaptability". An application to this method for hearing instrument new design illustrates the proposed approach.

Efficient Power-Delay Product Modulo 2n+1 Adder Design

As embedded and portable systems were emerged power consumption of circuits had been major challenge. On the other hand latency as determines frequency of circuits is also vital task. Therefore, trade off between both of them will be desirable. Modulo 2n+1 adders are important part of the residue number system (RNS) based arithmetic units with the interesting moduli set (2n-1,2n, 2n+1). In this manuscript we have introduced novel binary representation to the design of modulo 2n+1 adder. VLSI realization of proposed architecture under 180 nm full static CMOS technology reveals its superiority in terms of area, power consumption and power-delay product (PDP) against several peer existing structures.

A Self-Consistent Scheme for Elastic-Plastic Asperity Contact

In this paper, a generalized self-consistent scheme, or “three phase model", is used to set up a micro-mechanics model for rough surface contact with randomly distributed asperities. The dimensionless average real pressure p is obtained as function of the ratio of the real contact area to the apparent contact area, 0 A / A r . Both elastic and plastic materials are considered, and the influence of the plasticity of material on p is discussed. Both two-dimensional and three-dimensional rough surface contact problems are considered.

Dempster-Shafer's Approach for Autonomous Virtual Agent Navigation in Virtual Environments

This paper presents a solution for the behavioural animation of autonomous virtual agent navigation in virtual environments. We focus on using Dempster-Shafer-s Theory of Evidence in developing visual sensor for virtual agent. The role of the visual sensor is to capture the information about the virtual environment or identifie which part of an obstacle can be seen from the position of the virtual agent. This information is require for vitual agent to coordinate navigation in virtual environment. The virual agent uses fuzzy controller as a navigation system and Fuzzy α - level for the action selection method. The result clearly demonstrates the path produced is reasonably smooth even though there is some sharp turn and also still not diverted too far from the potential shortest path. This had indicated the benefit of our method, where more reliable and accurate paths produced during navigation task.