Regional Development Programs: A Reason for Them Failing

This paper contributes to the analysis of the design of regional development programs. This is a case study the birth, life, death and afterlife of a stately development program in Norway, supporting diffusion of innovations by promoting e-business in SMEs (small and medium sized enterprises). The study shows that joint projects like regional development programs have to be designed such that the present value of the future benefits always exceeds the present value of the future effort for all stakeholders vital for the survival of the project. The study also indicate that a development program not always have one common goal which all the stakeholders agree upon. There are several stakeholders who may have different goals by playing a part in the realization of the program. Even if some parties evaluate the results of a development program as a failure, other may have attained their goals. The lessons learned from this study may advise the designers of development programs involving many independent stakeholders. There is a lack of research examining failing development programs, investigating the reasons for it to be considered a failure. This paper shows why a development program was terminated and gives hint to how joint programs could be designed in order for the program to deliver the wanted results to all the key stakeholders.

Adaptive Score Normalization: A Novel Approach for Multimodal Biometric Systems

Multimodal biometric systems integrate the data presented by multiple biometric sources, hence offering a better performance than the systems based on a single biometric modality. Although the coupling of biometric systems can be done at different levels, the fusion at the scores level is the most common since it has been proven effective than the rest of the fusion levels. However, the scores from different modalities are generally heterogeneous. A step of normalizing the scores is needed to transform these scores into a common domain before combining them. In this paper, we study the performance of several normalization techniques with various fusion methods in a context relating to the merger of three unimodal systems based on the face, the palmprint and the fingerprint. We also propose a new adaptive normalization method that takes into account the distribution of client scores and impostor scores. Experiments conducted on a database of 100 people show that the performances of a multimodal system depend on the choice of the normalization method and the fusion technique. The proposed normalization method has given the best results.

Effect of Zr Addition on Mechanical Properties of Cr-Mo Plastic Mold Steels

We investigated the effects of the additions of Zr and other alloying elements on the mechanical properties and microstructure in Cr-Mo plastic mold steels. The addition of alloying elements changed the microstructure of the normalized samples from the upper bainite to lower bainite due to the increased hardenability. The tempering temperature influenced the strength and hardness values, especially the phenomenon of 350oC embrittlement was observed. The alloy additions of Cr, Mo, and V improved the resistance to the temper embrittlement. The addition of Zr improved the tensile strength and yield strength, but the impact energy was sharply decreased. It may be caused by the formation of Zr-MnS inclusion and rectangular-shaped Zr inclusion due to the Zr addition.

Early-Warning Lights Classification Management System for Industrial Parks in Taiwan

This paper presents the early-warning lights classification management system for industrial parks promoted by the Taiwan Environmental Protection Administration (EPA) since 2011, including the definition of each early-warning light, objectives, action program and accomplishments. All of the 151 industrial parks in Taiwan were classified into four early-warning lights, including red, orange, yellow and green, for carrying out respective pollution management according to the monitoring data of soil and groundwater quality, regulatory compliance, and regulatory listing of control site or remediation site. The Taiwan EPA set up a priority list for high potential polluted industrial parks and investigated their soil and groundwater qualities based on the results of the light classification and pollution potential assessment. In 2011-2013, there were 44 industrial parks selected and carried out different investigation, such as the early warning groundwater well networks establishment and pollution investigation/verification for the red and orange-light industrial parks and the environmental background survey for the yellow-light industrial parks. Among them, 22 industrial parks were newly or continuously confirmed that the concentrations of pollutants exceeded those in soil or groundwater pollution control standards. Thus, the further investigation, groundwater use restriction, listing of pollution control site or remediation site, and pollutant isolation measures were implemented by the local environmental protection and industry competent authorities; the early warning lights of those industrial parks were proposed to adjust up to orange or red-light. Up to the present, the preliminary positive effect of the soil and groundwater quality management system for industrial parks has been noticed in several aspects, such as environmental background information collection, early warning of pollution risk, pollution investigation and control, information integration and application, and inter-agency collaboration. Finally, the work and goal of self-initiated quality management of industrial parks will be carried out on the basis of the inter-agency collaboration by the classified lights system of early warning and management as well as the regular announcement of the status of each industrial park.

Isolation and Classification of Red Blood Cells in Anemic Microscopic Images

Red blood cells (RBCs) are among the most commonly and intensively studied type of blood cells in cell biology. Anemia is a lack of RBCs is characterized by its level compared to the normal hemoglobin level. In this study, a system based image processing methodology was developed to localize and extract RBCs from microscopic images. Also, the machine learning approach is adopted to classify the localized anemic RBCs images. Several textural and geometrical features are calculated for each extracted RBCs. The training set of features was analyzed using principal component analysis (PCA). With the proposed method, RBCs were isolated in 4.3secondsfrom an image containing 18 to 27 cells. The reasons behind using PCA are its low computation complexity and suitability to find the most discriminating features which can lead to accurate classification decisions. Our classifier algorithm yielded accuracy rates of 100%, 99.99%, and 96.50% for K-nearest neighbor (K-NN) algorithm, support vector machine (SVM), and neural network RBFNN, respectively. Classification was evaluated in highly sensitivity, specificity, and kappa statistical parameters. In conclusion, the classification results were obtained within short time period, and the results became better when PCA was used.

Using Multi-Linguistic Techniques for Thailand Herb and Traditional Medicine Registration Systems

Thailand has evolved many unique culture and knowledge, and the leading is the Thai traditional medicine (TTM). Recently, a number of researchers have tried to save this indigenous knowledge. However, the system to do so has still been scant. To preserve this ancient knowledge, we therefore invented and integrated multi-linguistic techniques to create the system of the collected all of recipes. This application extracted the medical recipes from antique scriptures then normalized antiquarian words, primitive grammar and antiquated measurement of them to the modern ones. Then, we applied ingredient-duplication-calculation, proportion-similarity-calculation and score-ranking to examine duplicate recipes. We collected the questionnaires from registrants and people to investigate the users’ satisfaction. The satisfactory results were found. This application assists not only registrants to validating the copyright violation in TTM registration process but also people to cure their illness that aids both Thai people and all mankind to fight for intractable diseases.

Partial Purification of Cytotoxic Peptides against Gastric Cancer Cells from Protein Hydrolysate of Euphorbia hirta Linn.

Protein hydrolysates prepared from a number of medicinal plants are promising sources of various bioactive peptides. In this work, proteins from dried whole plant of Euphorbia hirta Linn. were extracted and digested with pepsin for 12h. The hydrolysates of lesser than 3 KDa were fractionated by a cut-off membrane. The peptide hydrolysate was then purified by an anion-exchange chromatography on DEAE-Sephacel™ column and reverse-phase chromatography on Sep-pak C18 column, respectively. The cytotoxic effect of each peptide fraction against a gastric carcinoma cell line (KATO-III, ATCC No. HTB103) was investigated using colorimetric MTT viability assay. A human liver cell line (Chang Liver, CLS No. 300139) was used as a control normal cell line. Two purified peptide peaks, peak l and peak ll at 100µg peptides mL-1 affected cell viability of the gastric cancer cell lines to 63.85±4.94 and 66.92±6.46%, respectively. Our result showed for the first time that the peptide fractions derived from protein hydrolysate of Euphorbia hirta Linn. have anti-gastric cancer activity, which offers a potential novel and natural anti-gastric cancer remedy.

Using Strategic CSR to Achieve the Hybrid Middle Ground in Social Entrepreneurship: The Case of Telenor Hungary

To be considered a socially entrepreneurial organization today requires achieving what can be termed a “hybrid middle ground” equilibrium, comprising of economic as well as social sustainability. This middle ground requires some blend of both business and social commitments. In this paper, we use the case of Hungary's second ranked mobile operator, Telenor Hungary to illustrate an example of a company that is moving to the hybrid middle ground by transitioning from a for-profit company to a socially responsible business using the concept of strategic CSR. In this line of thinking, the organization explicitly supports programs and initiatives that have a direct link to the core business and bring operational and/or financial advantages for the company, while creating a positive social and/or environmental impact. The important lessons learned from the company transition are also discussed. 

Designing Software Quality Measurement System for Telecommunication Industry Using Object-Oriented Technique

Numbers of software quality measurement system have been implemented over the past few years, but none of them focuses on telecommunication industry. Software quality measurement system for telecommunication industry was a system that could calculate the quality value of the measured software that totally focused in telecommunication industry. Before designing a system, quality factors, quality attributes and quality metrics were identified based on literature review and survey. Then, using the identified quality factors, quality attributes and quality metrics, quality model for telecommunication industry was constructed. Each identified quality metrics had its own formula. Quality value for the system was measured based on the quality metrics and aggregated by referring to the quality model. It would classify the quality level of the software based on Net Satisfaction Index (NSI). The system was designed using object-oriented approach in web-based environment. Thus, existing of software quality measurement system was important to both developers and users in order to produce high quality software product for telecommunication industry.

Future Logistics - Challenges, Requirements and Solutions for Logistics Networks

The importance of logistics has changed enormously in the last few decades. While logistics was formerly one of the core functions of most companies, logistics or at least parts of their functions are nowadays outsourced to external logistic service providers in terms of contracts. As a result of this shift new business models like the fourth party logistics provider emerged, which designs, plans and monitors the resulting logistics networks. This new business model and topics such as Synchromodality or Big Data impose new requirements on the underlying IT, which cannot be met with conventional concepts and approaches. In this paper, the challenges of logistics network monitoring are outlined by using a scenario. The most common layers in a logical multilayered architecture for an information system are used to point out the arising challenges for IT. In addition, first appropriate solution approaches are introduced.  

Grid–SVC: An Improvement in SVC Algorithm, Based On Grid Based Clustering

Support vector clustering (SVC) is an important kernelbased clustering algorithm in multi applications. It has got two main bottle necks, the high computation price and labeling piece. In this paper, we presented a modified SVC method, named Grid–SVC, to improve the original algorithm computationally. First we normalized and then we parted the interval, where the SVC is processing, using a novel Grid–based clustering algorithm. The algorithm parts the intervals, based on the density function of the data set and then applying the cartesian multiply makes multi-dimensional grids. Eliminating many outliers and noise in the preprocess, we apply an improved SVC method to each parted grid in a parallel way. The experimental results show both improvement in time complexity order and the accuracy.

Remote Sensing, GIS, and AHP for Assessing Physical Vulnerability to Tsunami Hazard

Remote sensing image processing, spatial data analysis through GIS approach, and analytical hierarchy process were introduced in this study for assessing the vulnerability area and inundation area due to tsunami hazard in the area of Rikuzentakata, Iwate Prefecture, Japan. Appropriate input parameters were derived from GSI DEM data, ALOS AVNIR-2, and field data. We used the parameters of elevation, slope, shoreline distance, and vegetation density. Five classes of vulnerability were defined and weighted via pairwise comparison matrix. The assessment results described that 14.35km2 of the study area was under tsunami vulnerability zone. Inundation areas are those of high and slightly high vulnerability. The farthest area reached by a tsunami was about 7.50km from the shoreline and shows that rivers act as flooding strips that transport tsunami waves into the hinterland. This study can be used for determining a priority for land-use planning in the scope of tsunami hazard risk management.

Development of Perez-Du Mortier Calibration Algorithm for Ground-Based Aerosol Optical Depth Measurement with Validation using SMARTS Model

Aerosols are small particles suspended in air that have wide varying spatial and temporal distributions. The concentration of aerosol in total columnar atmosphere is normally measured using aerosol optical depth (AOD). In long-term monitoring stations, accurate AOD retrieval is often difficult due to the lack of frequent calibration. To overcome this problem, a near-sea-level Langley calibration algorithm is developed using the combination of clear-sky detection model and statistical filter. It attempts to produce a dataset that consists of only homogenous and stable atmospheric condition for the Langley calibration purposes. In this paper, a radiance-based validation method is performed to further investigate the feasibility and consistency of the proposed algorithm at different location, day, and time. The algorithm is validated using SMARTS model based n DNI value. The overall results confirmed that the proposed calibration algorithm feasible and consistent for measurements taken at different sites and weather conditions.

Enhanced Gram-Schmidt Process for Improving the Stability in Signal and Image Processing

The Gram-Schmidt Process (GSP) is used to convert a non-orthogonal basis (a set of linearly independent vectors) into an orthonormal basis (a set of orthogonal, unit-length vectors). The process consists of taking each vector and then subtracting the elements in common with the previous vectors. This paper introduces an Enhanced version of the Gram-Schmidt Process (EGSP) with inverse, which is useful for signal and image processing applications.

How Stock Market Reacts to Guidance Revisions and Actual Earnings Surprises

According to the existing literature, companies manage analysts’ expectations of their future earnings by issuing pessimistic earnings guidance to meet the expectations. Consequently, one could expect that markets price this pessimistic bias in advance and penalize companies more for lowering the guidance than reward for beating the guidance. In this paper we confirm this empirically. In addition we show that although guidance revisions have a statistically significant relation to stock returns, that is not the case with the actual earnings surprise. Reason for this could be that, after the annual earnings report also information on future earnings power is given at the same time.

Fixture Layout Optimization Using Element Strain Energy and Genetic Algorithm

The stiffness of the workpiece is very important to reduce the errors in manufacturing process. The high stiffness of the workpiece can be achieved by optimal positioning of fixture elements in the fixture. The minimization of the sum of the nodal deflection normal to the surface is used as objective function in previous research. The deflection in other direction has been neglected. The 3-2-1 fixturing principle is not valid for metal sheets due to its flexible nature. We propose a new fixture layout optimization method N-3-2-1 for metal sheets that uses the strain energy of the finite elements. This method combines the genetic algorithm and finite element analysis. The objective function in this method is to minimize the sum of all the element strain energy. By using the concept of element strain energy, the deformations in all the directions have been considered. Strain energy and stiffness are inversely proportional to each other. So, lower the value of strain energy, higher will be the stiffness. Two different kinds of case studies are presented. The case studies are solved for both objective functions; element strain energy and nodal deflection. The result are compared to verify the propose method.

An Analysis of New Service Interchange Designs

An efficient freeway system will be essential to the development of Africa, and interchanges are a key to that efficiency. Around the world, many interchanges between freeways and surface streets, called service interchanges, are of the diamond configuration, and interchanges using roundabouts or loop ramps are also popular. However, many diamond interchanges have serious operational problems, interchanges with roundabouts fail at high demand levels, and loops use lots of expensive land. Newer service interchange designs provide other options. The most popular new interchange design in the US at the moment is the double crossover diamond (DCD), also known as the diverging diamond. The DCD has enormous potential, but also has several significant limitations. The objectives of this paper are to review new service interchange options and to highlight some of the main features of those alternatives. The paper tests four conventional and seven unconventional designs using seven measures related to efficiency, cost, and safety. The results show that there is no superior design in all measures investigated. The DCD is better than most designs tested on most measures examined. However, the DCD was only superior to all other designs for bridge width. The DCD performed relatively poorly for capacity and for serving pedestrians. Based on the results, African freeway designers are encouraged to investigate the full range of alternatives that could work at the spot of interest. Diamonds and DCDs have their niches, but some of the other designs investigated could be optimum at some spots.

Hypolipidemic and Antioxidant Effects of Black Tea Extract and Quercetin in Atherosclerotic Rats

Background: Atherosclerosis is the main cause of cardiovascular disease (CVD) with complex and multifactorial process including atherogenic lipoprotein, oxidized low density lipoprotein (LDL), endothelial dysfunction, plaque stability, vascular inflammation, thrombotic and fibrinolytic disorder, exercises and genetic factor Epidemiological studies have shown tea consumption inversely associated with the development and progression of atherosclerosis. The research objectives: to elucidate hypolipidemic, antioxidant effects, as well as ability to improve coronary artery’s histopathologyof black tea extract (BTE) and quercetin in atherosclerotic rats. Methods: The antioxidant activity was determined by using Superoxide Dismutase activity (SOD) of serum and lipid peroxidation product (Malondialdehyde) of plasma and lipid profile including cholesterol total, LDL, triglyceride (TG), High Density Lipoprotein (HDL) of atherosclerotic rats. Inducing atherosclerotic, rats were given cholesterol and cholic acid in feed during ten weeks until rats indicated atherosclerotic symptom with narrowed artery and foamy cells in the artery’s wall. After rats suffered atherosclerotic, the high cholesterol feed and cholic acid were stopped and rats were given BTE 450; 300; 150 mg/kg body weight (BW) daily, quercetin 15; 10; 5 mg/kg BW daily, compared to rats were given vitamin E 60 mg/kg/BW; simvastatin 2.7 mg/kg BW, probucol 30 mg/kg BW daily for 21 days (first treatment) and 42 days (second treatment), negative control (normal feed), positive control (atherosclerotic rats). Results: BTE and quercetin could lower cholesterol total, triglyceride, LDL MDA and increase HDL, SOD were comparable with simvastatin, probucol both for 21 days and 42 days treatment, as well to improve coronary arteries histopathology. Conclusions: BTE andquercetin have hypolipidemic and antioxidant effects, as well as improve coronary arteries histopathology in atherosclerotic rats.

Determination and Comparison of Fabric Pills Distribution Using Image Processing and Spatial Data Analysis Tools

This work deals with the determination and comparison of pill patterns in 2 sets of fabric samples which differ in way of pill creation. The first set contains fabric samples with the pills created by simulation on a Martindale abrasion machine, while pills in the second set originated during normal wearing and maintenance. The goal of the study is to determine whether the pattern of the fabric pills created by simulation is the same as the pattern of naturally occurring pills. The system of determination and comparison of the pills is based on image processing and spatial data analysis tools. Firstly, 3D reconstruction of the fabric surfaces with the pills is realized with using a gradient fields method. The gradient fields method creates a 3D fabric surface from a set of 4 images. Thereafter, the pills are detected in 3D fabric surfaces using image-processing tools in the MATLAB software. Determination and comparison of the pills patterns of two sets of fabric samples is based on spatial data analysis using tools in R software.

A 3 Dimensional Simulation of the Repeated Load Triaxial Test

A typical flexible pavement structure consists of the surface, base, sub-base and subgrade soil. The loading traffic is transferred from the top layer with higher stiffness to the layer below with less stiffness. Under normal traffic loading, the behaviour of flexible pavement is very complex and can be predicted by using the repeated load triaxial test equipment in the laboratory. However, the nature of the repeated load triaxial testing procedure is considered time-consuming, complicated and expensive, and it is a challenge to carry out as a routine test in the laboratory. Therefore, the current paper proposes a numerical approach to simulate the repeated load triaxial test by employing the discrete element method. A sample with particle size ranging from 2.36mm to 19.0mm was constructed. Material properties, which included normal stiffness, shear stiffness, coefficient of friction, maximum dry density and particle density, were used as the input for the simulation. The sample was then subjected to a combination of deviator and confining stress and it was found that the discrete element method is able to simulate the repeated load triaxial test in the laboratory.