Pay Differentials and Employee Retention in the State Colleges of Education in the South-South Zone, Nigeria

The study examined the influence of pay differentials on employee retention in the State Colleges of Education in the South-South Region of Nigeria. 275 subjects drawn from members of the wage negotiating teams in the Colleges were administered questionnaires constructed for study. Analysis of Variance revealed that the observed pay differentials significantly influenced retainership, f(5,269 = 6.223, P< 0.05). However, the Multiple Classification Analysis and Post-Hoc test indicated that employees in two of the Colleges with slightly lower and higher pay levels may probably remain with their employers while employees in other Colleges with the least and highest pay levels suggested quitting. Based on these observations, the influence of pay on employee retention seems inconclusive. Generally, employees in the colleges studied are dissatisfied with current pay levels. Management should confront these challenges by improving pay packages to encourage employees to remain and be dedicated to duty.

Cartoon Effect and Ambient Illumination Based Depth Perception Assessment of 3D Video

Monitored 3-Dimensional (3D) video experience can be utilized as “feedback information” to fine tune the service parameters for providing a better service to the demanding 3D service customers. The 3D video experience which includes both video quality and depth perception is influenced by several contextual and content related factors (e.g., ambient illumination condition, content characteristics, etc) due to the complex nature of the 3D video. Therefore, effective factors on this experience should be utilized while assessing it. In this paper, structural information of the depth map sequences of the 3D video is considered as content related factor effective on the depth perception assessment. Cartoon-like filter is utilized to abstract the significant depth levels in the depth map sequences to determine the structural information. Moreover, subjective experiments are conducted using 3D videos associated with cartoon-like depth map sequences to investigate the effectiveness of ambient illumination condition, which is a contextual factor, on depth perception. Using the knowledge gained through this study, 3D video experience metrics can be developed to deliver better service to the 3D video service users. 

Flexible Wormhole-Switched Network-on-chip with Two-Level Priority Data Delivery Service

A synchronous network-on-chip using wormhole packet switching and supporting guaranteed-completion best-effort with low-priority (LP) and high-priority (HP) wormhole packet delivery service is presented in this paper. Both our proposed LP and HP message services deliver a good quality of service in term of lossless packet completion and in-order message data delivery. However, the LP message service does not guarantee minimal completion bound. The HP packets will absolutely use 100% bandwidth of their reserved links if the HP packets are injected from the source node with maximum injection. Hence, the service are suitable for small size messages (less than hundred bytes). Otherwise the other HP and LP messages, which require also the links, will experience relatively high latency depending on the size of the HP message. The LP packets are routed using a minimal adaptive routing, while the HP packets are routed using a non-minimal adaptive routing algorithm. Therefore, an additional 3-bit field, identifying the packet type, is introduced in their packet headers to classify and to determine the type of service committed to the packet. Our NoC prototypes have been also synthesized using a 180-nm CMOS standard-cell technology to evaluate the cost of implementing the combination of both services.

Linear Instability of Wake-Shear Layers in Two-Phase Shallow Flows

Linear stability analysis of wake-shear layers in twophase shallow flows is performed in the present paper. Twodimensional shallow water equations are used in the analysis. It is assumed that the fluid contains uniformly distributed solid particles. No dynamic interaction between the carrier fluid and particles is expected in the initial moment. The stability calculations are performed for different values of the particle loading parameter and two other parameters which characterize the velocity ratio and the velocity deficit. The results show that the particle loading parameter has a stabilizing effect on the flow while the increase in the velocity ratio or in the velocity deficit destabilizes the flow.

Remarks on Energy Based Control of a Nonlinear, Underactuated, MIMO and Unstable Benchmark

In the last decade, energy based control theory has undergone a significant breakthrough in dealing with underactated mechanical systems with two successful and similar tools, controlled Lagrangians and controlled Hamiltanians (IDA-PBC). However, because of the complexity of these tools, successful case studies are lacking, in particular, MIMO cases. The seminal theoretical paper of controlled Lagrangians proposed by Bloch and his colleagues presented a benchmark example–a 4 d.o.f underactuated pendulum on a cart but a detailed and completed design is neglected. To compensate this ignorance, the note revisit their design idea by addressing explicit control functions for a similar device motivated by a vector thrust body hovering in the air. To the best of our knowledge, this system is the first MIMO, underactuated example that is stabilized by using energy based tools at the courtesy of the original design idea. Some observations are given based on computer simulation.

Towards Cloud Computing Anatomy

Cloud Computing has recently emerged as a compelling paradigm for managing and delivering services over the internet. The rise of Cloud Computing is rapidly changing the landscape of information technology, and ultimately turning the longheld promise of utility computing into a reality. As the development of Cloud Computing paradigm is speedily progressing, concepts, and terminologies are becoming imprecise and ambiguous, as well as different technologies are interfering. Thus, it becomes crucial to clarify the key concepts and definitions. In this paper, we present the anatomy of Cloud Computing, covering its essential concepts, prominent characteristics, its affects, architectural design and key technologies. We differentiate various service and deployment models. Also, significant challenges and risks need are tackled in order to guarantee the long-term success of Cloud Computing. The aim of this paper is to provide a better understanding of the anatomy of Cloud Computing and pave the way for further research in this area.

Development of Transmission Line Sleeve Inspection Robot

The line sleeves on power transmission line connects two conductors while the transmission line is constructing. However, the line sleeves sometimes cause transmission line break down, because the line sleeves are deteriorated and decayed by acid rain. When the transmission line is broken, the economical loss is huge. Therefore the line sleeves on power transmission lines should be inspected periodically to prevent power failure. In this paper, Korea Electric Power Research Institute reviewed several robots to inspect line status and proposes a robot to inspect line sleeve by measuring magnetic field on line sleeve. The developed inspection tool can reliable to move along transmission line and overcome several obstacles on transmission line. The developed system is also applied on power transmission line and verified the efficiency of the robot.

Evaluation of Clustering Based on Preprocessing in Gene Expression Data

Microarrays have become the effective, broadly used tools in biological and medical research to address a wide range of problems, including classification of disease subtypes and tumors. Many statistical methods are available for analyzing and systematizing these complex data into meaningful information, and one of the main goals in analyzing gene expression data is the detection of samples or genes with similar expression patterns. In this paper, we express and compare the performance of several clustering methods based on data preprocessing including strategies of normalization or noise clearness. We also evaluate each of these clustering methods with validation measures for both simulated data and real gene expression data. Consequently, clustering methods which are common used in microarray data analysis are affected by normalization and degree of noise and clearness for datasets.

Numerical Simulation of Tidal Currents in Persian Gulf

In this paper, a two-dimensional (2D) numerical model for the tidal currents simulation in Persian Gulf is presented. The model is based on the depth averaged equations of shallow water which consider hydrostatic pressure distribution. The continuity equation and two momentum equations including the effects of bed friction, the Coriolis effects and wind stress have been solved. To integrate the 2D equations, the Alternative Direction Implicit (ADI) technique has been used. The base of equations discritization was finite volume method applied on rectangular mesh. To evaluate the model validation, a dam break case study including analytical solution is selected and the comparison is done. After that, the capability of the model in simulation of tidal current in a real field is represented by modeling the current behavior in Persian Gulf. The tidal fluctuations in Hormuz Strait have caused the tidal currents in the area of study. Therefore, the water surface oscillations data at Hengam Island on Hormoz Strait are used as the model input data. The check point of the model is measured water surface elevations at Assaluye port. The comparison between the results and the acceptable agreement of them showed the model ability for modeling marine hydrodynamic.

Dynamical Analysis of Circadian Gene Expression

Microarrays technique allows the simultaneous measurements of the expression levels of thousands of mRNAs. By mining this data one can identify the dynamics of the gene expression time series. By recourse of principal component analysis, we uncover the circadian rhythmic patterns underlying the gene expression profiles from Cyanobacterium Synechocystis. We applied PCA to reduce the dimensionality of the data set. Examination of the components also provides insight into the underlying factors measured in the experiments. Our results suggest that all rhythmic content of data can be reduced to three main components.

A Numerical Strategy to Design Maneuverable Micro-Biomedical Swimming Robots Based on Biomimetic Flagellar Propulsion

Medical applications are among the most impactful areas of microrobotics. The ultimate goal of medical microrobots is to reach currently inaccessible areas of the human body and carry out a host of complex operations such as minimally invasive surgery (MIS), highly localized drug delivery, and screening for diseases at their very early stages. Miniature, safe and efficient propulsion systems hold the key to maturing this technology but they pose significant challenges. A new type of propulsion developed recently, uses multi-flagella architecture inspired by the motility mechanism of prokaryotic microorganisms. There is a lack of efficient methods for designing this type of propulsion system. The goal of this paper is to overcome the lack and this way, a numerical strategy is proposed to design multi-flagella propulsion systems. The strategy is based on the implementation of the regularized stokeslet and rotlet theory, RFT theory and new approach of “local corrected velocity". The effects of shape parameters and angular velocities of each flagellum on overall flow field and on the robot net forces and moments are considered. Then a multi-layer perceptron artificial neural network is designed and employed to adjust the angular velocities of the motors for propulsion control. The proposed method applied successfully on a sample configuration and useful demonstrative results is obtained.

A Novel Methodology for Synthesis of Fault Trees from MATLAB-Simulink Model

Fault tree analysis is a well-known method for reliability and safety assessment of engineering systems. In the last 3 decades, a number of methods have been introduced, in the literature, for automatic construction of fault trees. The main difference between these methods is the starting model from which the tree is constructed. This paper presents a new methodology for the construction of static and dynamic fault trees from a system Simulink model. The method is introduced and explained in detail, and its correctness and completeness is experimentally validated by using an example, taken from literature. Advantages of the method are also mentioned.

Efficiency of Post-Tensioning Method for Seismic Retrofitting of Pre-Cast Cylindrical Concrete Reservoirs

Cylindrical concrete reservoirs are appropriate choice for storing liquids as water, oil and etc. By using of the pre-cast concrete reservoirs instead of the in-situ constructed reservoirs, the speed and precision of the construction would considerably increase. In this construction method, wall and roof panels would make in factory with high quality materials and precise controlling. Then, pre-cast wall and roof panels would carry out to the construction site for assembling. This method has a few faults such as: the existing weeks in connection of wall panels together and wall panels to foundation. Therefore, these have to be resisted under applied loads such as seismic load. One of the innovative methods which was successfully applied for seismic retrofitting of numerous pre-cast cylindrical water reservoirs in New Zealand, using of the high tensile cables around the reservoirs and post-tensioning them. In this paper, analytical modeling of wall and roof panels and post-tensioned cables are carried out with finite element method and the effect of height to diameter ratio, post-tensioning force value, liquid level in reservoir, installing position of tendons on seismic response of reservoirs are investigated.

The Impact of HIV/AIDS on Micro-enterprise Development in Kenya: A Study of Obunga Slum in Kisumu

The performances of small and medium enterprises have stagnated in the last two decades. This has mainly been due to the emergence of HIV / Aids. The disease has had a detrimental effect on the general economy of the country leading to morbidity and mortality of the Kenyan workforce in their primary age. The present study sought to establish the economic impact of HIV / Aids on the micro-enterprise development in Obunga slum – Kisumu, in terms of production loss, increasing labor related cost and to establish possible strategies to address the impact of HIV / Aids on microenterprises. The study was necessitated by the observation that most micro-enterprises in the slum are facing severe economic and social crisis due to the impact of HIV / Aids, they get depleted and close down within a short time due to death of skilled and experience workforce. The study was carried out between June 2008 and June 2009 in Obunga slum. Data was subjected to computer aided statistical analysis that included descriptive statistic, chi-squared and ANOVA techniques. Chi-squared analysis on the micro-enterprise owners opinion on the impact of HIV / Aids on depletion of microenterprise compared to other diseases indicated high levels of the negative effects of the disease at significance levels of P

Program Camouflage: A Systematic Instruction Hiding Method for Protecting Secrets

This paper proposes an easy-to-use instruction hiding method to protect software from malicious reverse engineering attacks. Given a source program (original) to be protected, the proposed method (1) takes its modified version (fake) as an input, (2) differences in assembly code instructions between original and fake are analyzed, and, (3) self-modification routines are introduced so that fake instructions become correct (i.e., original instructions) before they are executed and that they go back to fake ones after they are executed. The proposed method can add a certain amount of security to a program since the fake instructions in the resultant program confuse attackers and it requires significant effort to discover and remove all the fake instructions and self-modification routines. Also, this method is easy to use (with little effort) because all a user (who uses the proposed method) has to do is to prepare a fake source code by modifying the original source code.

Brand Equity and Factors Affecting Consumer-s Purchase Intention towards Luxury Brands in Bangkok Metropolitan Area

The purposes of this research were 1) to study consumer-based equity of luxury brands, 2) to study consumers- purchase intention for luxury brands, 3) to study direct factors affecting purchase intention towards luxury brands, and 4) to study indirect factors affecting purchase intention towards luxury brands through brand consciousness and brand equity to analyze information by descriptive statistic and hierarchical stepwise regression analysis. The findings revealed that the eight variables of the framework which were: need for uniqueness, normative susceptibility, status consumption, brand consciousness, brand awareness, perceived quality, brand association, and brand loyalty affected the purchase intention of the luxury brands (at the significance of 0.05). Brand Loyalty had the strongest direct effect while status consumption had the strongest indirect effect affecting the purchase intention towards luxury brands. Brand consciousness and brand equity had the mediators through the purchase intention of the luxury brands (at the significance of 0.05).

A Study on Fantasy Images Represented on the Films: Focused on Mise-en-Scène Element

The genre of fantasy depicts a world of imagine that triggers popular interest from a created view of world, and a fantasy is defined as a story that illustrates a world of imagine where scientific or horror elements are stand in its center. This study is not focused on the narrative of the fantasy, i.e. not on the adventurous story, but is concentrated on the image of the fantasy to work on its relationship with intended themes and differences among cultures due to meanings of materials. As for films, we have selected some films in the 2000's that are internationally recognized as expressing unique images of fantasy containing the theme of love in them. The selected films are 5 pieces including two European films, Amelie from Montmartre (2001) and The Science of Sleep (2005) and three Asian films, Citizen Dog from Thailand (2004), Memories of Matsuko from Japan (2006), and I'm a Cyborg, but That's OK from Korea (2006). These films share some common characteristics to the effect that they give tiny lessons and feelings for life with expressions of fantasy images as if they were fairy tales for adults and that they lead the audience to reflect on their days and revive forgotten dreams of childhood. We analyze the images of fantasy in each of the films on the basis of the elements of Mise-en-Scène (setting and props, costume, hair and make-up, facial expressions and body language, lighting and color, positioning of characters, and objects within a frame).

A 16Kb 10T-SRAM with 4x Read-Power Reduction

This work aims to reduce the read power consumption as well as to enhance the stability of the SRAM cell during the read operation. A new 10-transisor cell is proposed with a new read scheme to minimize the power consumption within the memory core. It has separate read and write ports, thus cell read stability is significantly improved. A 16Kb SRAM macro operating at 1V supply voltage is demonstrated in 65 nm CMOS process. Its read power consumption is reduced to 24% of the conventional design. The new cell also has lower leakage current due to its special bit-line pre-charge scheme. As a result, it is suitable for low-power mobile applications where power supply is restricted by the battery.

Development of Integrated GIS Interface for Characteristics of Regional Daily Flow

The purpose of this paper primarily intends to develop GIS interface for estimating sequences of stream-flows at ungauged stations based on known flows at gauged stations. The integrated GIS interface is composed of three major steps. The first, precipitation characteristics using statistical analysis is the procedure for making multiple linear regression equation to get the long term mean daily flow at ungauged stations. The independent variables in regression equation are mean daily flow and drainage area. Traditionally, mean flow data are generated by using Thissen polygon method. However, method for obtaining mean flow data can be selected by user such as Kriging, IDW (Inverse Distance Weighted), Spline methods as well as other traditional methods. At the second, flow duration curve (FDC) is computing at unguaged station by FDCs in gauged stations. Finally, the mean annual daily flow is computed by spatial interpolation algorithm. The third step is to obtain watershed/topographic characteristics. They are the most important factors which govern stream-flows. In summary, the simulated daily flow time series are compared with observed times series. The results using integrated GIS interface are closely similar and are well fitted each other. Also, the relationship between the topographic/watershed characteristics and stream flow time series is highly correlated.

Simulation of Organic Matter Variability on a Sugarbeet Field Using the Computer Based Geostatistical Methods

Computer based geostatistical methods can offer effective data analysis possibilities for agricultural areas by using vectorial data and their objective informations. These methods will help to detect the spatial changes on different locations of the large agricultural lands, which will lead to effective fertilization for optimal yield with reduced environmental pollution. In this study, topsoil (0-20 cm) and subsoil (20-40 cm) samples were taken from a sugar beet field by 20 x 20 m grids. Plant samples were also collected from the same plots. Some physical and chemical analyses for these samples were made by routine methods. According to derived variation coefficients, topsoil organic matter (OM) distribution was more than subsoil OM distribution. The highest C.V. value of 17.79% was found for topsoil OM. The data were analyzed comparatively according to kriging methods which are also used widely in geostatistic. Several interpolation methods (Ordinary,Simple and Universal) and semivariogram models (Spherical, Exponential and Gaussian) were tested in order to choose the suitable methods. Average standard deviations of values estimated by simple kriging interpolation method were less than average standard deviations (topsoil OM ± 0.48, N ± 0.37, subsoil OM ± 0.18) of measured values. The most suitable interpolation method was simple kriging method and exponantial semivariogram model for topsoil, whereas the best optimal interpolation method was simple kriging method and spherical semivariogram model for subsoil. The results also showed that these computer based geostatistical methods should be tested and calibrated for different experimental conditions and semivariogram models.