Prediction of a Human Facial Image by ANN using Image Data and its Content on Web Pages

Choosing the right metadata is a critical, as good information (metadata) attached to an image will facilitate its visibility from a pile of other images. The image-s value is enhanced not only by the quality of attached metadata but also by the technique of the search. This study proposes a technique that is simple but efficient to predict a single human image from a website using the basic image data and the embedded metadata of the image-s content appearing on web pages. The result is very encouraging with the prediction accuracy of 95%. This technique may become a great assist to librarians, researchers and many others for automatically and efficiently identifying a set of human images out of a greater set of images.

European Ecological Network Natura 2000 - Opportunities and Threats

The research objective of the project and article “European Ecological Network Natura 2000 – opportunities and threats” Natura 2000 sites constitute a form of environmental protection, several legal problems are likely to result. Most controversially, certain sites will be subject to two regimes of protection: as national parks and as Natura 2000 sites. This dualism of the legal regulation makes it difficult to perform certain legal obligations related to the regimes envisaged under each form of environmental protection. Which regime and which obligations resulting from the particular form of environmental protection have priority and should prevail? What should be done if these obligations are contradictory? Furthermore, an institutional problem consists in that no public administration authority has the power to resolve legal conflicts concerning the application of a particular regime on a given site. There are also no criteria to decide priority and superiority of one form of environmental protection over the other. Which regulations are more important, those that pertain to national parks or to Natura 2000 sites? In the light of the current regulations, it is impossible to give a decisive answer to these questions. The internal hierarchy of forms of environmental protection has not been determined, and all such forms should be treated equally.

Study on Ultrasonic Vibration Effects on Grinding Process of Alumina Ceramic (Al2O3)

Nowadays, engineering ceramics have significant applications in different industries such as; automotive, aerospace, electrical, electronics and even martial industries due to their attractive physical and mechanical properties like very high hardness and strength at elevated temperatures, chemical stability, low friction and high wear resistance. However, these interesting properties plus low heat conductivity make their machining processes too hard, costly and time consuming. Many attempts have been made in order to make the grinding process of engineering ceramics easier and many scientists have tried to find proper techniques to economize ceramics' machining processes. This paper proposes a new diamond plunge grinding technique using ultrasonic vibration for grinding Alumina ceramic (Al2O3). For this purpose, a set of laboratory equipments have been designed and simulated using Finite Element Method (FEM) and constructed in order to be used in various measurements. The results obtained have been compared with the conventional plunge grinding process without ultrasonic vibration and indicated that the surface roughness and fracture strength improved and the grinding forces decreased.

Data Traffic Dynamics and Saturation on a Single Link

The dynamics of User Datagram Protocol (UDP) traffic over Ethernet between two computers are analyzed using nonlinear dynamics which shows that there are two clear regimes in the data flow: free flow and saturated. The two most important variables affecting this are the packet size and packet flow rate. However, this transition is due to a transcritical bifurcation rather than phase transition in models such as in vehicle traffic or theorized large-scale computer network congestion. It is hoped this model will help lay the groundwork for further research on the dynamics of networks, especially computer networks.

Knowledge Sharing Behaviour among Academic Staff at a Public Higher Education Institution in Malaysia

This study applied Theory of Planned Behaviour (TPB) to explain the knowledge sharing behaviour among academic staff at a Public Higher Education Institution (HEI) in Malaysia. The main objectives of this study are; to identify the components that influence knowledge sharing behaviour and to determine the levels of knowledge sharing behaviour among academic staff. A total of 200 respondents were participated in answering questionnaires. The findings of this study revealed that level of perceiving and implementing knowledge sharing behaviour among academic staff at a Public HEI in Malaysia exist but not openly or strongly practiced. The findings were discussed and recommendations for the future research were also addressed.

Paradigm of Relocation of Urban Poor Habitats (Slums): Case Study of Nagpur City

Developing countries are facing a problem of slums and there appears to be no fool proof solution to eradicate them. For improving the quality of life there are three approaches of slum development and In-situ up-gradation approach is found to be the best one, while the relocation approach has proved to be failure. Factors responsible for failure of relocation projects are needed to be assessed, which is the basic aim of the paper. Factors responsible for failure of relocation projects are loss of livelihood, security of tenure and inefficiency of the Government. These factors are traced out & mapped from the examples of Western & Indian cities. National habitat, Resettlement policy emphasized relationship between shelter and work place. SRA has identified 55 slums for relocation due reservation of land uses, security of tenure and non- notified status of slums. The policy guidelines have been suggested for successful relocation projects. KeywordsLivelihood, Relocation, Slums, Urban poor.

Mobile Phone as a Tool for Data Collection in Field Research

The necessity of accurate and timely field data is shared among organizations engaged in fundamentally different activities, public services or commercial operations. Basically, there are three major components in the process of the qualitative research: data collection, interpretation and organization of data, and analytic process. Representative technological advancements in terms of innovation have been made in mobile devices (mobile phone, PDA-s, tablets, laptops, etc). Resources that can be potentially applied on the data collection activity for field researches in order to improve this process. This paper presents and discuss the main features of a mobile phone based solution for field data collection, composed of basically three modules: a survey editor, a server web application and a client mobile application. The data gathering process begins with the survey creation module, which enables the production of tailored questionnaires. The field workforce receives the questionnaire(s) on their mobile phones to collect the interviews responses and sending them back to a server for immediate analysis.

A Perceptually Optimized Foveation Based Wavelet Embedded Zero Tree Image Coding

In this paper, we propose a Perceptually Optimized Foveation based Embedded ZeroTree Image Coder (POEFIC) that introduces a perceptual weighting to wavelet coefficients prior to control SPIHT encoding algorithm in order to reach a targeted bit rate with a perceptual quality improvement with respect to a given bit rate a fixation point which determines the region of interest ROI. The paper also, introduces a new objective quality metric based on a Psychovisual model that integrates the properties of the HVS that plays an important role in our POEFIC quality assessment. Our POEFIC coder is based on a vision model that incorporates various masking effects of human visual system HVS perception. Thus, our coder weights the wavelet coefficients based on that model and attempts to increase the perceptual quality for a given bit rate and observation distance. The perceptual weights for all wavelet subbands are computed based on 1) foveation masking to remove or reduce considerable high frequencies from peripheral regions 2) luminance and Contrast masking, 3) the contrast sensitivity function CSF to achieve the perceptual decomposition weighting. The new perceptually optimized codec has the same complexity as the original SPIHT techniques. However, the experiments results show that our coder demonstrates very good performance in terms of quality measurement.

Performance and Emission Characteristics of a DI Diesel Engine Fuelled with Cashew Nut Shell Liquid (CNSL)-Diesel Blends

The increased number of automobiles in recent years has resulted in great demand for fossil fuel. This has led to the development of automobile by using alternative fuels which include gaseous fuels, biofuels and vegetables oils as fuel. Energy from biomass and more specific bio-diesel is one of the opportunities that could cover the future demand of fossil fuel shortage. Biomass in the form of cashew nut shell represents a new energy source and abundant source of energy in India. The bio-fuel is derived from cashew nut shell oil and its blend with diesel are promising alternative fuel for diesel engine. In this work the pyrolysis Cashew Nut Shell Liquid (CNSL)-Diesel Blends (CDB) was used to run the Direct Injection (DI) diesel engine. The experiments were conducted with various blends of CNSL and Diesel namely B20, B40, B60, B80 and B100. The results are compared with neat diesel operation. The brake thermal efficiency was decreased for blends of CNSL and Diesel except the lower blends of B20. The brake thermal efficiency of B20 is nearly closer to that of diesel fuel. Also the emission level of the all CNSL and Diesel blends was increased compared to neat diesel. The higher viscosity and lower volatility of CNSL leads to poor mixture formation and hence lower brake thermal efficiency and higher emission levels. The higher emission level can be reduced by adding suitable additives and oxygenates with CNSL and Diesel blends.

Improving Fault Resilience and Reconstruction of Overlay Multicast Tree Using Leaving Time of Participants

Network layer multicast, i.e. IP multicast, even after many years of research, development and standardization, is not deployed in large scale due to both technical (e.g. upgrading of routers) and political (e.g. policy making and negotiation) issues. Researchers looked for alternatives and proposed application/overlay multicast where multicast functions are handled by end hosts, not network layer routers. Member hosts wishing to receive multicast data form a multicast delivery tree. The intermediate hosts in the tree act as routers also, i.e. they forward data to the lower hosts in the tree. Unlike IP multicast, where a router cannot leave the tree until all members below it leave, in overlay multicast any member can leave the tree at any time thus disjoining the tree and disrupting the data dissemination. All the disrupted hosts have to rejoin the tree. This characteristic of the overlay multicast causes multicast tree unstable, data loss and rejoin overhead. In this paper, we propose that each node sets its leaving time from the tree and sends join request to a number of nodes in the tree. The nodes in the tree will reject the request if their leaving time is earlier than the requesting node otherwise they will accept the request. The node can join at one of the accepting nodes. This makes the tree more stable as the nodes will join the tree according to their leaving time, earliest leaving time node being at the leaf of the tree. Some intermediate nodes may not follow their leaving time and leave earlier than their leaving time thus disrupting the tree. For this, we propose a proactive recovery mechanism so that disrupted nodes can rejoin the tree at predetermined nodes immediately. We have shown by simulation that there is less overhead when joining the multicast tree and the recovery time of the disrupted nodes is much less than the previous works. Keywords

Technique for Grounding System Design in Distribution Substation

This paper presents the significant factor and give some suggestion that should know before design. The main objective of this paper is guide the first step for someone who attends to design of grounding system before study in details later. The overview of grounding system can protect damage from fault such as can save a human life and power system equipment. The unsafe conditions have three cases. Case 1) maximum touch voltage exceeds the safety criteria. In this case, the conductor compression ratio of the ground gird should be first adjusted to have optimal spacing of ground grid conductors. If it still over limit, earth resistivity should be consider afterward. Case 2) maximum step voltage exceeds the safety criteria. In this case, increasing the number of ground grid conductors around the boundary can solve this problem. Case 3) both of maximum touch and step voltage exceed the safety criteria. In this case, follow the solutions explained in case 1 and case 2. Another suggestion, vary depth of ground grid until maximum step and touch voltage do not exceed the safety criteria.

Adaptive MPC Using a Recursive Learning Technique

A model predictive controller based on recursive learning is proposed. In this SISO adaptive controller, a model is automatically updated using simple recursive equations. The identified models are then stored in the memory to be re-used in the future. The decision for model update is taken based on a new control performance index. The new controller allows the use of simple linear model predictive controllers in the control of nonlinear time varying processes.

Linear Elasticity Problems Solved by Using the Fictitious Domain Method and Total - FETI Domain Decomposition

The main goal of this paper is to show a possibility, how to solve numerically elliptic boundary value problems arising in 2D linear elasticity by using the fictitious domain method (FDM) and the Total-FETI domain decomposition method. We briefly mention the theoretical background of these methods and demonstrate their performance on a benchmark.

A Laboratory Assistance Module

We propose that Virtual Learning Environments (VLEs) should be designed by taking into account the characteristics, the special needs and the specific operating rules of the academic institutions in which they are employed. In this context, we describe a VLE module that extends the support of the organization and delivery of course material by including administration activities related to the various stages of teaching. These include the co-ordination, collaboration and monitoring of the course material development process and institution-specific course material delivery modes. Our specialized module, which enhances VLE capabilities by Helping Educators and Learners through a Laboratory Assistance System, is willing to assist the Greek tertiary technological sector, which includes Technological Educational Institutes (T.E.I.).

Losses Analysis in TEP Considering Uncertainity in Demand by DPSO

This paper presents a mathematical model and a methodology to analyze the losses in transmission expansion planning (TEP) under uncertainty in demand. The methodology is based on discrete particle swarm optimization (DPSO). DPSO is a useful and powerful stochastic evolutionary algorithm to solve the large-scale, discrete and nonlinear optimization problems like TEP. The effectiveness of the proposed idea is tested on an actual transmission network of the Azerbaijan regional electric company, Iran. The simulation results show that considering the losses even for transmission expansion planning of a network with low load growth is caused that operational costs decreases considerably and the network satisfies the requirement of delivering electric power more reliable to load centers.

Continuity Planning in Supply Chain Networks: Degrees of Freedom and Application in the Risk Management Process

Supply chain networks are frequently hit by unplanned events which lead to disruptions and cause operational and financial consequences. It is neither possible to avoid disruption risk entirely, nor are network members able to prepare for every possible disruptive event. Therefore a continuity planning should be set up which supports effective operational responses in supply chain networks in times of emergencies. In this research network related degrees of freedom which determine the options for responsive actions are derived from interview data. The findings are further embedded into a common risk management process. The paper provides support for researchers and practitioners to identify the network related options for responsive actions and to determine the need for improving the reaction capabilities.

An Algorithm Proposed for FIR Filter Coefficients Representation

Finite impulse response (FIR) filters have the advantage of linear phase, guaranteed stability, fewer finite precision errors, and efficient implementation. In contrast, they have a major disadvantage of high order need (more coefficients) than IIR counterpart with comparable performance. The high order demand imposes more hardware requirements, arithmetic operations, area usage, and power consumption when designing and fabricating the filter. Therefore, minimizing or reducing these parameters, is a major goal or target in digital filter design task. This paper presents an algorithm proposed for modifying values and the number of non-zero coefficients used to represent the FIR digital pulse shaping filter response. With this algorithm, the FIR filter frequency and phase response can be represented with a minimum number of non-zero coefficients. Therefore, reducing the arithmetic complexity needed to get the filter output. Consequently, the system characteristic i.e. power consumption, area usage, and processing time are also reduced. The proposed algorithm is more powerful when integrated with multiplierless algorithms such as distributed arithmetic (DA) in designing high order digital FIR filters. Here the DA usage eliminates the need for multipliers when implementing the multiply and accumulate unit (MAC) and the proposed algorithm will reduce the number of adders and addition operations needed through the minimization of the non-zero values coefficients to get the filter output.

Neuro-fuzzy Model and Regression Model a Comparison Study of MRR in Electrical Discharge Machining of D2 Tool Steel

In the current research, neuro-fuzzy model and regression model was developed to predict Material Removal Rate in Electrical Discharge Machining process for AISI D2 tool steel with copper electrode. Extensive experiments were conducted with various levels of discharge current, pulse duration and duty cycle. The experimental data are split into two sets, one for training and the other for validation of the model. The training data were used to develop the above models and the test data, which was not used earlier to develop these models were used for validation the models. Subsequently, the models are compared. It was found that the predicted and experimental results were in good agreement and the coefficients of correlation were found to be 0.999 and 0.974 for neuro fuzzy and regression model respectively

Empirical Analyses of Determinants of D.J.S.I.US Mean Returns

This study investigates the relationship between 10 year bond value, Yen/U.S dollar exchange rate, non-farm payrolls (all employs) and crude oil to U.S. Dow Jones Sustainability Index. A GARCH model is used to test these relationships for the period January 1st 1999 to January 31st 2008 using monthly data. Results show that an increase of the 10 year bond and non farm payrolls (all employs) lead to an increase of the D.J.S.I returns. On the contrary the volatility of the Yen/U.S dollar exchange rates as well as the increase of crude oil returns has negative effects on the U.S D.J.S.I returns. This study aims at assisting investors to understand the influences certain macroeconomic indicators have on the companies- stock returns as reported by the D.J.S.I.

Ranking Genes from DNA Microarray Data of Cervical Cancer by a local Tree Comparison

The major objective of this paper is to introduce a new method to select genes from DNA microarray data. As criterion to select genes we suggest to measure the local changes in the correlation graph of each gene and to select those genes whose local changes are largest. More precisely, we calculate the correlation networks from DNA microarray data of cervical cancer whereas each network represents a tissue of a certain tumor stage and each node in the network represents a gene. From these networks we extract one tree for each gene by a local decomposition of the correlation network. The interpretation of a tree is that it represents the n-nearest neighbor genes on the n-th level of a tree, measured by the Dijkstra distance, and, hence, gives the local embedding of a gene within the correlation network. For the obtained trees we measure the pairwise similarity between trees rooted by the same gene from normal to cancerous tissues. This evaluates the modification of the tree topology due to tumor progression. Finally, we rank the obtained similarity values from all tissue comparisons and select the top ranked genes. For these genes the local neighborhood in the correlation networks changes most between normal and cancerous tissues. As a result we find that the top ranked genes are candidates suspected to be involved in tumor growth. This indicates that our method captures essential information from the underlying DNA microarray data of cervical cancer.