Paradigm of Relocation of Urban Poor Habitats (Slums): Case Study of Nagpur City

Developing countries are facing a problem of slums and there appears to be no fool proof solution to eradicate them. For improving the quality of life there are three approaches of slum development and In-situ up-gradation approach is found to be the best one, while the relocation approach has proved to be failure. Factors responsible for failure of relocation projects are needed to be assessed, which is the basic aim of the paper. Factors responsible for failure of relocation projects are loss of livelihood, security of tenure and inefficiency of the Government. These factors are traced out & mapped from the examples of Western & Indian cities. National habitat, Resettlement policy emphasized relationship between shelter and work place. SRA has identified 55 slums for relocation due reservation of land uses, security of tenure and non- notified status of slums. The policy guidelines have been suggested for successful relocation projects. KeywordsLivelihood, Relocation, Slums, Urban poor.

Improving Fault Resilience and Reconstruction of Overlay Multicast Tree Using Leaving Time of Participants

Network layer multicast, i.e. IP multicast, even after many years of research, development and standardization, is not deployed in large scale due to both technical (e.g. upgrading of routers) and political (e.g. policy making and negotiation) issues. Researchers looked for alternatives and proposed application/overlay multicast where multicast functions are handled by end hosts, not network layer routers. Member hosts wishing to receive multicast data form a multicast delivery tree. The intermediate hosts in the tree act as routers also, i.e. they forward data to the lower hosts in the tree. Unlike IP multicast, where a router cannot leave the tree until all members below it leave, in overlay multicast any member can leave the tree at any time thus disjoining the tree and disrupting the data dissemination. All the disrupted hosts have to rejoin the tree. This characteristic of the overlay multicast causes multicast tree unstable, data loss and rejoin overhead. In this paper, we propose that each node sets its leaving time from the tree and sends join request to a number of nodes in the tree. The nodes in the tree will reject the request if their leaving time is earlier than the requesting node otherwise they will accept the request. The node can join at one of the accepting nodes. This makes the tree more stable as the nodes will join the tree according to their leaving time, earliest leaving time node being at the leaf of the tree. Some intermediate nodes may not follow their leaving time and leave earlier than their leaving time thus disrupting the tree. For this, we propose a proactive recovery mechanism so that disrupted nodes can rejoin the tree at predetermined nodes immediately. We have shown by simulation that there is less overhead when joining the multicast tree and the recovery time of the disrupted nodes is much less than the previous works. Keywords

Genetic Algorithms with Oracle for the Traveling Salesman Problem

By introducing the concept of Oracle we propose an approach for improving the performance of genetic algorithms for large-scale asymmetric Traveling Salesman Problems. The results have shown that the proposed approach allows overcoming some traditional problems for creating efficient genetic algorithms.

Linear Elasticity Problems Solved by Using the Fictitious Domain Method and Total - FETI Domain Decomposition

The main goal of this paper is to show a possibility, how to solve numerically elliptic boundary value problems arising in 2D linear elasticity by using the fictitious domain method (FDM) and the Total-FETI domain decomposition method. We briefly mention the theoretical background of these methods and demonstrate their performance on a benchmark.

Losses Analysis in TEP Considering Uncertainity in Demand by DPSO

This paper presents a mathematical model and a methodology to analyze the losses in transmission expansion planning (TEP) under uncertainty in demand. The methodology is based on discrete particle swarm optimization (DPSO). DPSO is a useful and powerful stochastic evolutionary algorithm to solve the large-scale, discrete and nonlinear optimization problems like TEP. The effectiveness of the proposed idea is tested on an actual transmission network of the Azerbaijan regional electric company, Iran. The simulation results show that considering the losses even for transmission expansion planning of a network with low load growth is caused that operational costs decreases considerably and the network satisfies the requirement of delivering electric power more reliable to load centers.

A New Approach to Face Recognition Using Dual Dimension Reduction

In this paper a new approach to face recognition is presented that achieves double dimension reduction, making the system computationally efficient with better recognition results and out perform common DCT technique of face recognition. In pattern recognition techniques, discriminative information of image increases with increase in resolution to a certain extent, consequently face recognition results change with change in face image resolution and provide optimal results when arriving at a certain resolution level. In the proposed model of face recognition, initially image decimation algorithm is applied on face image for dimension reduction to a certain resolution level which provides best recognition results. Due to increased computational speed and feature extraction potential of Discrete Cosine Transform (DCT), it is applied on face image. A subset of coefficients of DCT from low to mid frequencies that represent the face adequately and provides best recognition results is retained. A tradeoff between decimation factor, number of DCT coefficients retained and recognition rate with minimum computation is obtained. Preprocessing of the image is carried out to increase its robustness against variations in poses and illumination level. This new model has been tested on different databases which include ORL , Yale and EME color database.

Influence of Hydrocarbons on Plant Cell Ultrastructure and Main Metabolic Enzymes

Influence of octane and benzene on plant cell ultrastructure and enzymes of basic metabolism, such as nitrogen assimilation and energy generation have been studied. Different plants: perennial ryegrass (Lolium perenne) and alfalfa (Medicago sativa); crops- maize (Zea mays L.) and bean (Phaseolus vulgaris); shrubs – privet (Ligustrum sempervirens) and trifoliate orange (Poncirus trifoliate); trees - poplar (Populus deltoides) and white mulberry (Morus alba L.) were exposed to hydrocarbons of different concentrations (1, 10 and 100 mM). Destructive changes in bean and maize leaves cells ultrastructure under the influence of benzene vapour were revealed at the level of photosynthetic and energy generation subcellular organells. Different deviations at the level of subcellular organelles structure and distribution were observed in alfalfa and ryegrass root cells under the influence of benzene and octane, absorbed through roots. The level of destructive changes is concentration dependent. Benzene at low 1 and 10 mM concentration caused the increase in glutamate dehydrogenase (GDH) activity in maize roots and leaves and in poplar and mulberry shoots, though to higher extent in case of lower, 1mM concentration. The induction was more intensive in plant roots. The highest tested 100mM concentration of benzene was inhibitory to the enzyme in all plants. Octane caused induction of GDH in all grassy plants at all tested concentrations; however the rate of induction decreased parallel to increase of the hydrocarbon concentration. Octane at concentration 1 mM caused induction of GDH in privet, trifoliate and white mulberry shoots. The highest, 100mM octane was characterized by inhibitory effect to GDH activity in all plants. Octane had inductive effect on malate dehydrogenase in almost all plants and tested concentrations, indicating the intensification of Trycarboxylic Acid Cycle. The data could be suggested for elaboration of criteria for plant selection for phytoremediation of oil hydrocarbons contaminated soils.

Thermal Modeling of Dry-Transformers and Estimating Temperature Rise

Temperature rise in a transformer depends on variety of parameters such as ambient temperature, output current and type of the core. Considering these parameters, temperature rise estimation is still complicated procedure. In this paper, we present a new model based on simple electrical equivalent circuit. This method avoids the complication associated to accurate estimation and is in very good agreement with practice.

The Management in Large Emergency Situations – A Best Practise Case Study based on GIS for Management of Evacuation

In most of the cases, natural disasters lead to the necessity of evacuating people. The quality of evacuation management is dramatically improved by the use of information provided by decision support systems, which become indispensable in case of large scale evacuation operations. This paper presents a best practice case study. In November 2007, officers from the Emergency Situations Inspectorate “Crisana" of Bihor County from Romania participated to a cross-border evacuation exercise, when 700 people have been evacuated from Netherlands to Belgium. One of the main objectives of the exercise was the test of four different decision support systems. Afterwards, based on that experience, software system called TEVAC (Trans Border Evacuation) has been developed “in house" by the experts of this institution. This original software system was successfully tested in September 2008, during the deployment of the international exercise EU-HUROMEX 2008, the scenario involving real evacuation of 200 persons from Hungary to Romania. Based on the lessons learned and results, starting from April 2009, the TEVAC software is used by all Emergency Situations Inspectorates all over Romania.

A Comparison between Heterogeneous and Homogeneous Gas Flow Model in Slurry Bubble Column Reactor for Direct Synthesis of DME

In the present study, a heterogeneous and homogeneous gas flow dispersion model for simulation and optimisation of a large-scale catalytic slurry reactor for the direct synthesis of dimethyl ether (DME) from syngas and CO2, using a churn-turbulent regime was developed. In the heterogeneous gas flow model the gas phase was distributed into two bubble phases: small and large, however in the homogeneous one, the gas phase was distributed into only one large bubble phase. The results indicated that the heterogeneous gas flow model was in more agreement with experimental pilot plant data than the homogeneous one.

The Development of a Teachers- Self-Efficacy Instrument for High School Physical Education Teacher

The purpose of this study was to develop a “teachers’ self-efficacy scale for high school physical education teachers (TSES-HSPET)” in Taiwan. This scale is based on the self-efficacy theory of Bandura [1], [2]. This study used exploratory and confirmatory factor analyses to test the reliability and validity. The participants were high school physical education teachers in Taiwan. Both stratified random sampling and cluster sampling were used to sample participants for the study. 350 teachers were sampled in the first stage and 234 valid scales (male 133, female 101) returned. During the second stage, 350 teachers were sampled and 257 valid scales (male 143, female 110, 4 did not indicate gender) returned. The exploratory factor analysis was used in the first stage, and it got 60.77% of total variance for construct validity. The Cronbach’s alpha coefficient of internal consistency was 0.91 for sumscale, and subscales were 0.84 and 0.90. In the second stage, confirmatory factor analysis was used to test construct validity. The result showed that the fit index could be accepted (χ2 (75) =167.94, p

Reconstruction of the Most Energetic Modes in a Fully Developed Turbulent Channel Flow with Density Variation

Proper orthogonal decomposition (POD) is used to reconstruct spatio-temporal data of a fully developed turbulent channel flow with density variation at Reynolds number of 150, based on the friction velocity and the channel half-width, and Prandtl number of 0.71. To apply POD to the fully developed turbulent channel flow with density variation, the flow field (velocities, density, and temperature) is scaled by the corresponding root mean square values (rms) so that the flow field becomes dimensionless. A five-vector POD problem is solved numerically. The reconstructed second-order moments of velocity, temperature, and density from POD eigenfunctions compare favorably to the original Direct Numerical Simulation (DNS) data.

Effect of Sintering Temperature Curve in Wick Manufactured for Loop Heat Pipe

This investigation examines the effect of the sintering temperature curve in manufactured nickel powder capillary structure (wick) for a loop heat pipe (LHP). The sintering temperature curve is composed of a region of increasing temperature; a region of constant temperature and a region of declining temperature. The most important region is that in which the temperature increases, as an index in the stage in which the temperature increases. The wick of nickel powder is manufactured in the stage of fixed sintering temperature and the time between the stage of constant temperature and the stage of falling temperature. When the slope of the curve in the region of increasing temperature is unity (equivalent to 10 °C/min), the structure of the wick is complete and the heat transfer performance is optimal. The result of experiment test demonstrates that the heat transfer performance is optimal at 320W; the minimal total thermal resistance is approximately 0.18°C/W, and the heat flux is 17W/cm2; the internal parameters of the wick are an effective pore radius of 3.1 μm, a permeability of 3.25×10-13m2 and a porosity of 71%.

Conventional and PSO Based Approaches for Model Reduction of SISO Discrete Systems

Reduction of Single Input Single Output (SISO) discrete systems into lower order model, using a conventional and an evolutionary technique is presented in this paper. In the conventional technique, the mixed advantages of Modified Cauer Form (MCF) and differentiation are used. In this method the original discrete system is, first, converted into equivalent continuous system by applying bilinear transformation. The denominator of the equivalent continuous system and its reciprocal are differentiated successively, the reduced denominator of the desired order is obtained by combining the differentiated polynomials. The numerator is obtained by matching the quotients of MCF. The reduced continuous system is converted back into discrete system using inverse bilinear transformation. In the evolutionary technique method, Particle Swarm Optimization (PSO) is employed to reduce the higher order model. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example.

Observation of the Correlations between Pair Wise Interaction and Functional Organization of the Proteins, in the Protein Interaction Network of Saccaromyces Cerevisiae

Understanding the cell's large-scale organization is an interesting task in computational biology. Thus, protein-protein interactions can reveal important organization and function of the cell. Here, we investigated the correspondence between protein interactions and function for the yeast. We obtained the correlations among the set of proteins. Then these correlations are clustered using both the hierarchical and biclustering methods. The detailed analyses of proteins in each cluster were carried out by making use of their functional annotations. As a result, we found that some functional classes appear together in almost all biclusters. On the other hand, in hierarchical clustering, the dominancy of one functional class is observed. In brief, from interaction data to function, some correlated results are noticed about the relationship between interaction and function which might give clues about the organization of the proteins.

An Approach to Polynomial Curve Comparison in Geometric Object Database

In image processing and visualization, comparing two bitmapped images needs to be compared from their pixels by matching pixel-by-pixel. Consequently, it takes a lot of computational time while the comparison of two vector-based images is significantly faster. Sometimes these raster graphics images can be approximately converted into the vector-based images by various techniques. After conversion, the problem of comparing two raster graphics images can be reduced to the problem of comparing vector graphics images. Hence, the problem of comparing pixel-by-pixel can be reduced to the problem of polynomial comparisons. In computer aided geometric design (CAGD), the vector graphics images are the composition of curves and surfaces. Curves are defined by a sequence of control points and their polynomials. In this paper, the control points will be considerably used to compare curves. The same curves after relocated or rotated are treated to be equivalent while two curves after different scaled are considered to be similar curves. This paper proposed an algorithm for comparing the polynomial curves by using the control points for equivalence and similarity. In addition, the geometric object-oriented database used to keep the curve information has also been defined in XML format for further used in curve comparisons.

A Tool for Creation Artificial Symbiotic Associations of Wheat

This paper reports optimization of characteristics of bioballistic transformation of spring soft wheat (Triticum aestivum L. cultivar Raduga) and getting of transgenic plants, carrying pea lectin gene. This gene will let to create new associative wheat symbiosis with nodule bacteria of field pea, which has growth encouraging, fungistatic and other useful characteristics.

Effects of Li2O Thickness and Moisture Content on LiH Hydrolysis Kinetics in Slightly Humidified Argon

The hydrolysis kinetics of polycrystalline lithium hydride (LiH) in argon at various low humidities was measured by gravimetry and Raman spectroscopy with ambient water concentration ranging from 200 to 1200 ppm. The results showed that LiH hydrolysis curve revealed a paralinear shape, which was attributed to two different reaction stages that forming different products as explained by the 'Layer Diffusion Control' model. Based on the model, a novel two-stage rate equation for LiH hydrolysis reactions was developed and used to fit the experimental data for determination of Li2O steady thickness Hs and the ultimate hydrolysis rate vs. The fitted data presented a rise of Hs as ambient water concentration cw increased. However, in spite of the negative effect imposed by Hs increasing, the upward trend of vs remained, which implied that water concentration, rather than Li2O thickness, played a predominant role in LiH hydrolysis kinetics. In addition, the proportional relationship between vsHs and cw predicted by rate equation and confirmed by gravimetric data validated the model in such conditions.

Pattern Recognition as an Internalized Motor Programme

A new conceptual architecture for low-level neural pattern recognition is presented. The key ideas are that the brain implements support vector machines and that support vectors are represented as memory patterns in competitive queuing memories. A binary classifier is built from two competitive queuing memories holding positive and negative valence training examples respectively. The support vector machine classification function is calculated in synchronized evaluation cycles. The kernel is computed by bisymmetric feed-forward networks feed by sensory input and by competitive queuing memories traversing the complete sequence of support vectors. Temporary summation generates the output classification. It is speculated that perception apparatus in the brain reuses structures that have evolved for enabling fluent execution of prepared action sequences so that pattern recognition is built on internalized motor programmes.

Classroom Teacher Candidates' Definitions and Beliefs about Technology Integration

The purpose of this paper is to present teacher candidates- beliefs about technology integration in their field of study, which is classroom teaching in this case. The study was conducted among the first year students in college of education in Turkey. This study is based on both quantitative and qualitative data. For the quantitative data- Likert scale was used and for the qualitative data pattern matching was employed. The primary findings showed that students defined educational technology as technologies that improve learning with their visual, easily accessible, and productive features. They also believe these technologies could affect their future students- learning positively.