The Importance of Zakat in Struggle against Circle of Poverty and Income Redistribution

This paper examines how “Zakat” provides fair income redistribution and aids the struggle against poverty. Providing fair income redistribution and combating poverty constitutes some of the fundamental tasks performed by countries all over the world. Each country seeks a solution for these problems according to their political, economic and administrative styles through applying various economic and financial policies. The same situation can be handled via “zakat” association in Islam. Nowadays, we observe different versions of “zakat” in developed countries. Applications such as negative income tax denote merely a different form of “zakat” that is being applied almost in the same way but under changed names. However, the minimum values to donate under zakat (e.g. 85 gr. gold and 40 animals) get altered and various amounts are put into practice. It might be named as negative income tax instead of zakat, nonetheless, these applications are based on the Holy Koran and the hadith released 1400 years ago. Besides, considering the savage and slavery in the world at those times, we might easily recognize the true value of the zakat being applied for the first time then in the Islamic system. Through zakat, governments are able to transfer incomes to the poor as a means of enabling them achieve the minimum standard of living required. With regards to who benefits from the Zakat, an objective and fair criteria was used to determine who benefits from the zakat contrary to the notion that it was based on peoples’ own choices. Since the zakat is obligatory, the transfers do not get forwarded directly but via the government and get distributed, which requires vast governmental organizations. Through the application of Zakat, reduced levels of poverty can be achieved and also ensure the fair income redistribution.

Application of RS and GIS Technique for Identifying Groundwater Potential Zone in Gomukhi Nadhi Sub Basin, South India

India holds 17.5% of the world’s population but has only 2% of the total geographical area of the world where 27.35% of the area is categorized as wasteland due to lack of or less groundwater. So there is a demand for excessive groundwater for agricultural and non agricultural activities to balance its growth rate. With this in mind, an attempt is made to find the groundwater potential zone in Gomukhi Nadhi sub basin of Vellar River basin, TamilNadu, India covering an area of 1146.6 Sq.Km consists of 9 blocks from Peddanaickanpalayam to Virudhachalam in the sub basin. The thematic maps such as Geology, Geomorphology, Lineament, Landuse and Landcover and Drainage are prepared for the study area using IRS P6 data. The collateral data includes rainfall, water level, soil map are collected for analysis and inference. The digital elevation model (DEM) is generated using Shuttle Radar Topographic Mission (SRTM) and the slope of the study area is obtained. ArcGIS 10.1 acts as a powerful spatial analysis tool to find out the ground water potential zones in the study area by means of weighted overlay analysis. Each individual parameter of the thematic maps are ranked and weighted in accordance with their influence to increase the water level in the ground. The potential zones in the study area are classified viz., Very Good, Good, Moderate, Poor with its aerial extent of 15.67, 381.06, 575.38, 174.49 Sq.Km respectively.

SVPWM Based Two Level VSI for Micro Grids

With advances in solid-state power electronic devices and microprocessors, various pulse-width-modulation (PWM) techniques have been developed for industrial applications. This paper presents the comparison of two different PWM techniques, the sinusoidal PWM (SPWM) technique and the space-vector PWM (SVPWM) technique applied to two level VSI for micro grid applications. These two methods are compared by discussing their ease of implementation and by analyzing the output harmonic spectra of various output voltages (line-to-neutral voltages, and line-to-line voltages) and their total harmonic distortion (THD). The SVPWM technique in the under-modulation region can increase the fundamental output voltage by 15.5% over the SPWM technique.

Interactive Planning of Suburban Apartment Buildings

Construction in Finland is focusing increasingly on renovation instead of conventional new construction, and this trend will continue to grow in the coming years and decades. Renovation of the large number of suburban residential apartment buildings built in the 1960s and 1970s poses a particular challenge. However, renovation projects are demanding for the residents of these buildings, since they usually are uninitiated in construction issues. On the other hand, renovation projects generally apply the operating models of new construction. Nevertheless, the residents of an existing residential apartment building are some of the best experts on the site. Thus, in this research project we applied a relational model in developing and testing at case sites a planning process that employs interactive planning methods. Current residents, housing company managers, the city zoning manager, the contractor’s and prefab element supplier’s representatives, professional designers and researchers all took part in the planning. The entire interactive planning process progressed phase by phase as the participants’ and designers’ concerted discussion and ideation process, so that the end result was a renovation plan desired by the residents.

Spatial Time Series Models for Rice and Cassava Yields Based On Bayesian Linear Mixed Models

This paper proposes a linear mixed model (LMM) with spatial effects to forecast rice and cassava yields in Thailand at the same time. A multivariate conditional autoregressive (MCAR) model is assumed to present the spatial effects. A Bayesian method is used for parameter estimation via Gibbs sampling Markov Chain Monte Carlo (MCMC). The model is applied to the rice and cassava yields monthly data which have been extracted from the Office of Agricultural Economics, Ministry of Agriculture and Cooperatives of Thailand. The results show that the proposed model has better performance in most provinces in both fitting part and validation part compared to the simple exponential smoothing and conditional auto regressive models (CAR) from our previous study.

The Early Stages of the Standardization of Finnish Building Sector

Early 20th century functionalism aimed at generalising living and rationalising construction, thus laying the foundation for the standardisation of construction components and products. From the 1930s onwards, all measurement and quality instructions for building products, different types of building components, descriptions of working methods complying with advisable building practises, planning, measurement and calculation guidelines, terminology, etc. were called standards. Standardisation was regarded as a necessary prerequisite for the mass production of housing. This article examines the early stages of standardisation in Finland in the 1940s and 1950s, as reflected on the working history of an individual architect, ErkkiKoiso-Kanttila (1914-2006). In 1950 Koiso-Kanttila was appointed the Head of Design of the Finnish Association of Architects’ Building Standards Committee, a position which he held until 1958. His main responsibilities were the development of the RT Building Information File and compiling of the files.

Dynamic Analysis of Transmission Line Towers

The transmission line towers are one of the important life line structures in the distribution of power from the source to the various places for several purposes. The predominant external loads which act on these towers are wind and earthquake loads. In this present study tower is analyzed using Indian Standards IS: 875:1987(Wind Load), IS: 802:1995(Structural steel), IS:1893:2002 (Earthquake) and dynamic analysis of tower has been performed considering ground motion of 2001 Bhuj Earthquake (India). The dynamic analysis was performed considering a tower system consisting two towers spaced 800m apart and 35m height each. This analysis has been performed using numerical time stepping finite difference method which is central difference method were employed by a developed MATLAB program to get the normalized ground motion parameters includes acceleration, frequency, velocity which are important in designing the tower. The tower is analyzed using response spectrum analysis.

Voltage Stability Enhancement Using Cat Swarm Optimization Algorithm

Optimal Power Flow (OPF) problem in electrical power system is considered as a static, non-linear, multi-objective or a single objective optimization problem. This paper presents an algorithm for solving the voltage stability objective reactive power dispatch problem in a power system .The proposed approach employs cat swarm optimization algorithm for optimal settings of RPD control variables. Generator terminal voltages, reactive power generation of the capacitor banks and tap changing transformer setting are taken as the optimization variables. CSO algorithm is tested on standard IEEE 30 bus system and the results are compared with other methods to prove the effectiveness of the new algorithm. As a result, the proposed method is the best for solving optimal reactive power dispatch problem.

Isolation and Screening of Fungi for Aerobic Delignification and Reduction of AOX of Pulp and Paper Mill Effluent

Water pollution is a major concern for the pulp and paper industry due to the large quantities of effluents generated. Biodegradation of industrial Lignin and AOX by a fungal isolate identified as Aspergillus flavus, white rot fungi which was isolated from Pulp and Paper effluent was studied in batch flask system with industrial effluent and synthetic solution. The flasks were operated at temperature 32°C at 200rpm for eight days in continuous mode. The average overall pH, Temperature, DO, C.O.D, T.D.S, T.S.S, Lignin, AOX were up to 4.56, 32oC, 4.2mg/l, 104mg/l, 6000 mg/l, 4000mg/l, 575.5mg/l, 2195 mg/l respectively after treatment. The Aspergillus flavus sp was the most effective in the biodegradation of Lignin of pulp industry for 94% at 480nm, AOX for 62% at 510nm and of chemical oxygen demand levels for 45% after 8 days of incubation. The optimal conditions found were 4 pH and 32oC temperature for lignin and AOX degradation.

Optimal Choice and Location of Multi Type Facts Devices in Deregulated Electricity Market Using Evolutionary Programming Method

This paper deals with the optimal choice and allocation of multi FACTS devices in Deregulated power system using Evolutionary Programming method. The objective is to achieve the power system economic generation allocation and dispatch in deregulated electricity market. Using the proposed method, the locations of the FACTS devices, their types and ratings are optimized simultaneously. Different kinds of FACTS devices are simulated in this study such as UPFC, TCSC, TCPST, and SVC. Simulation results validate the capability of this new approach in minimizing the overall system cost function, which includes the investment costs of the FACTS devices and the bid offers of the market participants. The proposed algorithm is an effective and practical method for the choice and allocation of FACTS devices in deregulated electricity market environment. The standard data of IEEE 14 Bus systems has been taken into account and simulated with aid of MAT-lab software and results were obtained.

Computer Aided Docking Studies on Antiviral Drugs for SARS

Severe acute respiratory syndrome (SARS) is a respiratory disease in humans which is caused by the SARS coronavirus. The treatment of coronavirus-associated SARS has been evolving and so far there is no consensus on an optimal regimen. The mainstream therapeutic interventions for SARS involve broad-spectrum antibiotics and supportive care, as well as antiviral agents and immunomodulatory therapy. The Protein- Ligand interaction plays a significant role in structural based drug designing. In the present work we have taken the receptor Angiotensin converting enzyme 2 and identified the drugs that are commonly used against SARS. They are Lopinavir, Ritonavir, Ribavirin, and Oseltamivir. The receptor Angiotensin converting enzyme 2 (ACE-2) was docked with above said drugs and the energy value obtained are as follows, Lopinavir (-292.3), Ritonavir (-325.6), Oseltamivir (- 229.1), Ribavirin (-208.8). Depending on the least energy value we have chosen the best two drugs out of the four conventional drugs. We tried to improve the binding efficiency and steric compatibility of the two drugs namely Ritonavir and Lopinavir. Several modifications were made to the probable functional groups (phenylic, ketonic groups in case of Ritonavir and carboxylic groups in case of Lopinavir respectively) which were interacting with the receptor molecule. Analogs were prepared by Marvin Sketch software and were docked using HEX docking software. Lopinavir analog 8 and Ritonavir analog 11 were detected with significant energy values and are probable lead molecule. It infers that some of the modified drugs are better than the original drugs. Further work can be carried out to improve the steric compatibility of the drug based upon the work done above for a more energy efficient binding of the drugs to the receptor.

Comparison of Different Types of Sources of Traffic Using SFQ Scheduling Discipline

In this paper, SFQ (Start Time Fair Queuing) algorithm is analyzed when this is applied in computer networks to know what kind of behavior the traffic in the net has when different data sources are managed by the scheduler. Using the NS2 software the computer networks were simulated to be able to get the graphs showing the performance of the scheduler. Different traffic sources were introduced in the scripts, trying to establish the real scenario. Finally the results were that depending on the data source, the traffic can be affected in different levels, when Constant Bite Rate is applied, the scheduler ensures a constant level of data sent and received, but the truth is that in the real life it is impossible to ensure a level that resists the changes in work load.

Production of Novel Bioactive Yogurt Enriched with Olive Fruit Polyphenols

In the course of the present work, plain (nonencapsulated) and microencapsulated polyphenols were produced using olive mill wastewater (OMW) as raw material, in order to be used for enrichment of yogurt and dairy products. The OMW was first clarified by using membrane technology and subsequently the contained poly-phenols were isolated by adsorption-desorption technique using selective macro-porous resins and finally recovered in dry form after been processed by RO membrane technique followed by freeze drying. Moreover, the polyphenols were encapsulated in modified starch by freeze drying in order to mask the color and bitterness effect and improve their functionality. The two products were used successfully as additives in yogurt preparations and the produced products were acceptable by the consumers and presented with certain advantage to the plain yogurt. For the herein proposed production scheme a patent application was already submitted.

Dynamic Decompression for Text Files

Compression algorithms reduce the redundancy in data representation to decrease the storage required for that data. Lossless compression researchers have developed highly sophisticated approaches, such as Huffman encoding, arithmetic encoding, the Lempel-Ziv (LZ) family, Dynamic Markov Compression (DMC), Prediction by Partial Matching (PPM), and Burrows-Wheeler Transform (BWT) based algorithms. Decompression is also required to retrieve the original data by lossless means. A compression scheme for text files coupled with the principle of dynamic decompression, which decompresses only the section of the compressed text file required by the user instead of decompressing the entire text file. Dynamic decompressed files offer better disk space utilization due to higher compression ratios compared to most of the currently available text file formats.

Intelligent Caching in on-demand Routing Protocol for Mobile Adhoc Networks

An on-demand routing protocol for wireless ad hoc networks is one that searches for and attempts to discover a route to some destination node only when a sending node originates a data packet addressed to that node. In order to avoid the need for such a route discovery to be performed before each data packet is sent, such routing protocols must cache routes previously discovered. This paper presents an analysis of the effect of intelligent caching in a non clustered network, using on-demand routing protocols in wireless ad hoc networks. The analysis carried out is based on the Dynamic Source Routing protocol (DSR), which operates entirely on-demand. DSR uses the cache in every node to save the paths that are learnt during route discovery procedure. In this implementation, caching these paths only at intermediate nodes and using the paths from these caches when required is tried. This technique helps in storing more number of routes that are learnt without erasing the entries in the cache, to store a new route that is learnt. The simulation results on DSR have shown that this technique drastically increases the available memory for caching the routes discovered without affecting the performance of the DSR routing protocol in any way, except for a small increase in end to end delay.

Rethinking the Analysis of Means-End Chain Data in Marketing Research

This paper proposes a new procedure for analyzing means-end chain data in marketing research. Most commonly the collected data is summarized in the Hierarchical Value Map (HVM) illustrating the main attribute-consequence-value linkages. This paper argues that traditionally constructed HVM may give an erroneous impression of the results of a means-end study. To justify the arguments, an alternative procedure to (1) determine the dominant attribute-consequence-value linkages and (2) construct HVM in a precise manner is presented. The current approach makes a contribution to means-end analysis, allowing marketers to address a set of marketing problems, such as advertising strategy.

Flux Cored Arc Welding Parameter Optimization of AISI 316L (N) Austenitic Stainless Steel

Bead-on-plate welds were carried out on AISI 316L (N) austenitic stainless steel (ASS) using flux cored arc welding (FCAW) process. The bead on plates weld was conducted as per L25 orthogonal array. In this paper, the weld bead geometry such as depth of penetration (DOP), bead width (BW) and weld reinforcement (R) of AISI 316L (N) ASS are investigated. Taguchi approach is used as statistical design of experiment (DOE) technique for optimizing the selected welding input parameters. Grey relational analysis and desirability approach are applied to optimize the input parameters considering multiple output variables simultaneously. Confirmation experiment has also been conducted to validate the optimized parameters.

Molecular Docking Studies of Mycobacterium tuberculosis RNA Polymerase β Subunit (rpoB) Receptor

Tuberculosis (TB) is a bacterial infectious disease caused by the obligate human pathogen, Mycobacterium tuberculosis. Multidrug-resistant tuberculosis (MDR-TB) is a global reality that threatens tuberculosis control. Resistance to antibiotic Rifampicin, occurs in 95% of cases through nucleotide substitutions in an 81-bp core region of the rpoB i.e; beta subunit of DNA dependant RNA polymerase. In this paper, we studied the Rifampicin-rpoB receptor interactions In silico. First, homology modeling was performed to obtain the three dimensional structure of Mycobacterium rpoB. Sixty analogs of Rifampicin were prepared using Marvin sketch software. Both original Rifampicin and the analogs were docked with rpoB and energy values were obtained. Out of sixty analogs, 43 analogs had lesser energy values than conventional Rifampicin and hence are predicted to have greater binding affinity to rpoB. Thus, this study offers a route for the development of Rifampicin analogs against multi drug resistant Mycobacterium rpoB.

Proposing an Efficient Method for Frequent Pattern Mining

Data mining, which is the exploration of knowledge from the large set of data, generated as a result of the various data processing activities. Frequent Pattern Mining is a very important task in data mining. The previous approaches applied to generate frequent set generally adopt candidate generation and pruning techniques for the satisfaction of the desired objective. This paper shows how the different approaches achieve the objective of frequent mining along with the complexities required to perform the job. This paper will also look for hardware approach of cache coherence to improve efficiency of the above process. The process of data mining is helpful in generation of support systems that can help in Management, Bioinformatics, Biotechnology, Medical Science, Statistics, Mathematics, Banking, Networking and other Computer related applications. This paper proposes the use of both upward and downward closure property for the extraction of frequent item sets which reduces the total number of scans required for the generation of Candidate Sets.

The Role of Ga(Gallium)-flux and AlN(Aluminum Nitride) as the Interface Materials, between (Ga-face)GaN and (Siface)4H-SiC, through Molecular Dynamics Simulation

We report here, the results of molecular dynamics simulation of p-doped (Ga-face)GaN over n-doped (Siface)( 0001)4H-SiC hetero-epitaxial material system with one-layer each of Ga-flux and (Al-face)AlN, as the interface materials, in the form of, the total Density of States (DOS). It is found that the total DOS at the Fermi-level for the heavily p-doped (Ga-face)GaN and ndoped (Si-face)4H-SiC hetero-epitaxial system, with one layer of (Al-face)AlN as the interface material, is comparatively higher than that of the various cases studied, indicating that there could be good vertical conduction across the (Ga-face)GaN over (Si-face)(0001)4HSiC hetero-epitaxial material system.