Aqueous Extract of Flacourtia indica Prevents Carbon Tetrachloride Induced Hepatotoxicity in Rat

Carbon tetrachloride (CCl4) is a well-known hepatotoxin and exposure to this chemical is known to induce oxidative stress and causes liver injury by the formation of free radicals. Flacourtia indica commonly known as 'Baichi' has been reported as an effective remedy for the treatment of a variety of diseases. The objective of this study was to investigate the hepatoprotective activity of aqueous extract of leaves of Flacourtia indica against CCl4 induced hepatotoxicity. Animals were pretreated with the aqueous extract of Flacourtia indica (250 & 500 mg/kg body weight) for one week and then challenged with CCl4 (1.5 ml/kg bw) in olive oil (1:1, v/v) on 7th day. Serum marker enzymes (ALP, AST, ALT, Total Protein & Total Bilirubin) and TBARS level (Marker for oxidative stress) were estimated in all the study groups. Alteration in the levels of biochemical markers of hepatic damage like AST, ALT, ALP, Total Protein, Total Bilirubin and lipid peroxides (TBARS) were tested in both CCl4 treated and extract treated groups. CCl4 has enhanced the AST, ALT, ALP and the Lipid peroxides (TBARS) in liver. Treatment of aqueous extract of Flacourtia indica leaves (250 & 500 mg/kg) exhibited a significant protective effect by altering the serum levels of AST, ALT, ALP, Total Protein, Total Bilirubin and liver TBARS. These biochemical observations were supported by histopathological study of liver sections. From this preliminary study it has been concluded that the aqueous extract of the leaves of Flacourtia indica protects liver against oxidative damages and could be used as an effective protector against CCl4 induced hepatic damage. Our findings suggested that Flacourtia indica possessed good hepatoprotective activity

Using the Polynomial Approximation Algorithm in the Algorithm 2 for Manipulator's Control in an Unknown Environment

The Algorithm 2 for a n-link manipulator movement amidst arbitrary unknown static obstacles for a case when a sensor system supplies information about local neighborhoods of different points in the configuration space is presented. The Algorithm 2 guarantees the reaching of a target position in a finite number of steps. The Algorithm 2 is reduced to a finite number of calls of a subroutine for planning a trajectory in the presence of known forbidden states. The polynomial approximation algorithm which is used as the subroutine is presented. The results of the Algorithm2 implementation are given.

On The Elliptic Divisibility Sequences over Finite Fields

In this work we study elliptic divisibility sequences over finite fields. MorganWard in [11, 12] gave arithmetic theory of elliptic divisibility sequences. We study elliptic divisibility sequences, equivalence of these sequences and singular elliptic divisibility sequences over finite fields Fp, p > 3 is a prime.

On the Need to have an Additional Methodology for the Psychological Product Measurement and Evaluation

Cognitive Science appeared about 40 years ago, subsequent to the challenge of the Artificial Intelligence, as common territory for several scientific disciplines such as: IT, mathematics, psychology, neurology, philosophy, sociology, and linguistics. The new born science was justified by the complexity of the problems related to the human knowledge on one hand, and on the other by the fact that none of the above mentioned sciences could explain alone the mental phenomena. Based on the data supplied by the experimental sciences such as psychology or neurology, models of the human mind operation are built in the cognition science. These models are implemented in computer programs and/or electronic circuits (specific to the artificial intelligence) – cognitive systems – whose competences and performances are compared to the human ones, leading to the psychology and neurology data reinterpretation, respectively to the construction of new models. During these processes if psychology provides the experimental basis, philosophy and mathematics provides the abstraction level utterly necessary for the intermission of the mentioned sciences. The ongoing general problematic of the cognitive approach provides two important types of approach: the computational one, starting from the idea that the mental phenomenon can be reduced to 1 and 0 type calculus operations, and the connection one that considers the thinking products as being a result of the interaction between all the composing (included) systems. In the field of psychology measurements in the computational register use classical inquiries and psychometrical tests, generally based on calculus methods. Deeming things from both sides that are representing the cognitive science, we can notice a gap in psychological product measurement possibilities, regarded from the connectionist perspective, that requires the unitary understanding of the quality – quantity whole. In such approach measurement by calculus proves to be inefficient. Our researches, deployed for longer than 20 years, lead to the conclusion that measuring by forms properly fits to the connectionism laws and principles.

Multiscale Analysis and Change Detection Based on a Contrario Approach

Automatic methods of detecting changes through satellite imaging are the object of growing interest, especially beca²use of numerous applications linked to analysis of the Earth’s surface or the environment (monitoring vegetation, updating maps, risk management, etc...). This work implemented spatial analysis techniques by using images with different spatial and spectral resolutions on different dates. The work was based on the principle of control charts in order to set the upper and lower limits beyond which a change would be noted. Later, the a contrario approach was used. This was done by testing different thresholds for which the difference calculated between two pixels was significant. Finally, labeled images were considered, giving a particularly low difference which meant that the number of “false changes” could be estimated according to a given limit.

Clustering Mixed Data Using Non-normal Regression Tree for Process Monitoring

In the semiconductor manufacturing process, large amounts of data are collected from various sensors of multiple facilities. The collected data from sensors have several different characteristics due to variables such as types of products, former processes and recipes. In general, Statistical Quality Control (SQC) methods assume the normality of the data to detect out-of-control states of processes. Although the collected data have different characteristics, using the data as inputs of SQC will increase variations of data, require wide control limits, and decrease performance to detect outof- control. Therefore, it is necessary to separate similar data groups from mixed data for more accurate process control. In the paper, we propose a regression tree using split algorithm based on Pearson distribution to handle non-normal distribution in parametric method. The regression tree finds similar properties of data from different variables. The experiments using real semiconductor manufacturing process data show improved performance in fault detecting ability.

Bank Business Models and The Changes in CEE Countries

The aim of this article is to assess the existing business models used by the banks operating in the CEE countries in the time period from 2006 till 2011. In order to obtain research results, the authors performed qualitative analysis of the scientific literature on bank business models, which have been grouped into clusters that consist of such components as: 1) capital and reserves; 2) assets; 3) deposits, and 4) loans. In their turn, bank business models have been developed based on the types of core activities of the banks, and have been divided into four groups: Wholesale, Investment, Retail and Universal Banks. Descriptive statistics have been used to analyse the models, determining mean, minimal and maximal values of constituent cluster components, as well as standard deviation. The analysis of the data is based on such bank variable indices as Return on Assets (ROA) and Return on Equity (ROE).

A Comparison Study of Inspector's Performance between Regular and Complex Tasks

This research was to study a comparison of inspector-s performance between regular and complex visual inspection task. Visual task was simulated on DVD read control circuit. Inspection task was performed by using computer. Subjects were 10 undergraduate randomly selected and test for 20/20. Then, subjects were divided into two groups, five for regular inspection (control group) and five for complex inspection (treatment group) tasks. Result was showed that performance on regular and complex inspectors was significantly difference at the level of 0.05. Inspector performance on regular inspection was showed high percentage on defects detected by using equal time to complex inspection. This would be indicated that inspector performance was affected by visual inspection task.

Simulation and 40 Years of Object-Oriented Programming

2007 is a jubilee year: in 1967, programming language SIMULA 67 was presented, which contained all aspects of what was later called object-oriented programming. The present paper contains a description of the development unto the objectoriented programming, the role of simulation in this development and other tools that appeared in SIMULA 67 and that are nowadays called super-object-oriented programming.

Study of Barriers to Women's Entrepreneurship Development among Iranian Women (Case Entrepreneur Women)

In this research, effort was made to identify and evaluate barriers to the development of entrepreneurship among Iranian entrepreneur women who were graduated from universities. In this study, perspectives of thirty-seven available entrepreneur women were examined. In order to prepare questionnaires and receive knowledge about barriers among these women, seven cases of entrepreneur women took part in in-depth interviews. Then, to evaluate the importance of barriers, the researchers made a questionnaire with closed questions in which the barriers were classified into the following categories: personal-familial barriers; socio-cultural barriers; economic-financial-commercial barriers; and structural barriers. Entrepreneur women were requested to rate the importance of each item. The results indicated that there were different obstacles among entrepreneur women. The order of the important barriers was as fallow: economic-financial-commercial, structural, socio-cultural, and personal-familial.

A New Self-Adaptive EP Approach for ANN Weights Training

Evolutionary Programming (EP) represents a methodology of Evolutionary Algorithms (EA) in which mutation is considered as a main reproduction operator. This paper presents a novel EP approach for Artificial Neural Networks (ANN) learning. The proposed strategy consists of two components: the self-adaptive, which contains phenotype information and the dynamic, which is described by genotype. Self-adaptation is achieved by the addition of a value, called the network weight, which depends on a total number of hidden layers and an average number of neurons in hidden layers. The dynamic component changes its value depending on the fitness of a chromosome, exposed to mutation. Thus, the mutation step size is controlled by two components, encapsulated in the algorithm, which adjust it according to the characteristics of a predefined ANN architecture and the fitness of a particular chromosome. The comparative analysis of the proposed approach and the classical EP (Gaussian mutation) showed, that that the significant acceleration of the evolution process is achieved by using both phenotype and genotype information in the mutation strategy.

Modeling and Control of Two Manipulators Handling a Flexible Beam

This paper seeks to develop simple yet practical and efficient control scheme that enables cooperating arms to handle a flexible beam. Specifically the problem studied herein is that of two arms rigidly grasping a flexible beam and such capable of generating forces/moments in such away as to move a flexible beam along a predefined trajectory. The paper develops a sliding mode control law that provides robustness against model imperfection and uncertainty. It also provides an implicit stability proof. Simulation results for two three joint arms moving a flexible beam, are presented to validate the theoretical results.

Potential of Exopolysaccharides in Yoghurt Production

Consumer demand for products with low fat or sugar content and low levels of food additives, as well as cost factors, make exopolysaccharides (EPS) a viable alternative. EPS remain an interesting tool to modulate the sensory properties of yoghurt. This study was designed to evaluate EPS production potential of commercial yoghurt starter cultures (Yo-Flex starters: Harmony 1.0, TWIST 1.0 and YF-L902, Chr.Hansen, Denmark) and their influence on an apparent viscosity of yoghurt samples. The production of intracellularly synthesized EPS by different commercial yoghurt starters varies roughly from 144,08 to 440,81 mg/l. Analysing starters’ producing EPS, they showed large variations in concentration and supposedly composition. TWIST 1.0 had produced greater amounts of EPS in MRS medium and in yoghurt samples but there wasn’t determined significant contribution to development of texture as well as an apparent viscosity of the final product. YF-L902 and Harmony 1.0 starters differed considerably in EPS yields, but not in apparent viscosities (p>0.05) of the final yoghurts. Correlation between EPS concentration and viscosity of yoghurt samples was not established in the study.

A Novel Genetic Algorithm Designed for Hardware Implementation

A new genetic algorithm, termed the 'optimum individual monogenetic genetic algorithm' (OIMGA), is presented whose properties have been deliberately designed to be well suited to hardware implementation. Specific design criteria were to ensure fast access to the individuals in the population, to keep the required silicon area for hardware implementation to a minimum and to incorporate flexibility in the structure for the targeting of a range of applications. The first two criteria are met by retaining only the current optimum individual, thereby guaranteeing a small memory requirement that can easily be stored in fast on-chip memory. Also, OIMGA can be easily reconfigured to allow the investigation of problems that normally warrant either large GA populations or individuals many genes in length. Local convergence is achieved in OIMGA by retaining elite individuals, while population diversity is ensured by continually searching for the best individuals in fresh regions of the search space. The results given in this paper demonstrate that both the performance of OIMGA and its convergence time are superior to those of a range of existing hardware GA implementations.

Denoising by Spatial Domain Averaging for Wireless Local Area Network Terminal Localization

Terminal localization for indoor Wireless Local Area Networks (WLANs) is critical for the deployment of location-aware computing inside of buildings. A major challenge is obtaining high localization accuracy in presence of fluctuations of the received signal strength (RSS) measurements caused by multipath fading. This paper focuses on reducing the effect of the distance-varying noise by spatial filtering of the measured RSS. Two different survey point geometries are tested with the noise reduction technique: survey points arranged in sets of clusters and survey points uniformly distributed over the network area. The results show that the location accuracy improves by 16% when the filter is used and by 18% when the filter is applied to a clustered survey set as opposed to a straight-line survey set. The estimated locations are within 2 m of the true location, which indicates that clustering the survey points provides better localization accuracy due to superior noise removal.

Prototype for Enhancing Information Security Awareness in Industry

Human-related information security breaches within organizations are primarily caused by employees who have not been made aware of the importance of protecting the information they work with. Information security awareness is accordingly attracting more attention from industry, because stakeholders are held accountable for the information with which they work. The authors developed an Information Security Retrieval and Awareness model – entitled “ISRA" – that is tailored specifically towards enhancing information security awareness in industry amongst all users of information, to address shortcomings in existing information security awareness models. This paper is principally aimed at expounding a prototype for the ISRA model to highlight the advantages of utilizing the model. The prototype will focus on the non-technical, humanrelated information security issues in industry. The prototype will ensure that all stakeholders in an organization are part of an information security awareness process, and that these stakeholders are able to retrieve specific information related to information security issues relevant to their job category, preventing them from being overburdened with redundant information.

Low Power Bus Binding Based on Dynamic Bit Reordering

In this paper, the problem of reducing switching activity in on-chip buses at the stage of high-level synthesis is considered, and a high-level low power bus binding based on dynamic bit reordering is proposed. Whereas conventional methods use a fixed bit ordering between variables within a bus, the proposed method switches a bit ordering dynamically to obtain a switching activity reduction. As a result, the proposed method finds a binding solution with a smaller value of total switching activity (TSA). Experimental result shows that the proposed method obtains a binding solution having 12.0-34.9% smaller TSA compared with the conventional methods.

Characterisation of Hydrocarbons in Atmospheric Aerosols from Different European Sites

The concentrations of aliphatic and polycyclic aromatic hydrocarbons (PAH) were determined in atmospheric aerosol samples collected at a rural site in Hungary (K-puszta, summer 2008), a boreal forest (Hyytiälä,  April 2007) and a polluted rural area in Italy (San Pietro Capofiume, Po Valley, April 2008). A clear distinction between “clean" and “polluted" periods was observed. Concentrations obtained for Hyytiälä are significantly lower than those for the other two sites. Source reconciliation was performed using diagnostic parameters, such as the carbon preference index and ratios between PAH. The presence of an unresolved complex mixture of hydrocarbons, especially for the Finnish and Italian samples, is indicative of petrogenic inputs. In K-puszta, the aliphatic hydrocarbons are dominated by leaf wax n-alkanes. The long range transport of anthropogenic pollution contributed to the Finnish aerosol. Industrial activities and vehicular emissions represent major sources in San Pietro Capofiume. PAH in K-puszta consist of both pyrogenic and petrogenic compounds.

A Novel FFT-Based Frequency Offset Estimator for OFDM Systems

This paper proposes a novel frequency offset (FO) estimator for orthogonal frequency division multiplexing. Simplicity is most significant feature of this algorithm and can be repeated to achieve acceptable accuracy. Also fractional and integer part of FO is estimated jointly with use of the same algorithm. To do so, instead of using conventional algorithms that usually use correlation function, we use DFT of received signal. Therefore, complexity will be reduced and we can do synchronization procedure by the same hardware that is used to demodulate OFDM symbol. Finally, computer simulation shows that the accuracy of this method is better than other conventional methods.

Optimization Approaches for a Complex Dairy Farm Simulation Model

This paper describes the optimization of a complex dairy farm simulation model using two quite different methods of optimization, the Genetic algorithm (GA) and the Lipschitz Branch-and-Bound (LBB) algorithm. These techniques have been used to improve an agricultural system model developed by Dexcel Limited, New Zealand, which describes a detailed representation of pastoral dairying scenarios and contains an 8-dimensional parameter space. The model incorporates the sub-models of pasture growth and animal metabolism, which are themselves complex in many cases. Each evaluation of the objective function, a composite 'Farm Performance Index (FPI)', requires simulation of at least a one-year period of farm operation with a daily time-step, and is therefore computationally expensive. The problem of visualization of the objective function (response surface) in high-dimensional spaces is also considered in the context of the farm optimization problem. Adaptations of the sammon mapping and parallel coordinates visualization are described which help visualize some important properties of the model-s output topography. From this study, it is found that GA requires fewer function evaluations in optimization than the LBB algorithm.