Application of l1-Norm Minimization Technique to Image Retrieval

Image retrieval is a topic where scientific interest is currently high. The important steps associated with image retrieval system are the extraction of discriminative features and a feasible similarity metric for retrieving the database images that are similar in content with the search image. Gabor filtering is a widely adopted technique for feature extraction from the texture images. The recently proposed sparsity promoting l1-norm minimization technique finds the sparsest solution of an under-determined system of linear equations. In the present paper, the l1-norm minimization technique as a similarity metric is used in image retrieval. It is demonstrated through simulation results that the l1-norm minimization technique provides a promising alternative to existing similarity metrics. In particular, the cases where the l1-norm minimization technique works better than the Euclidean distance metric are singled out.

The Determination of Rating Points of Objects with Qualitative Characteristics and their Usagein Decision Making Problems

The paper presents the method developed to assess rating points of objects with qualitative indexes. The novelty of the method lies in the fact that the authors use linguistic scales that allow to formalize the values of the indexes with the help of fuzzy sets. As a result it is possible to operate correctly with dissimilar indexes on the unified basis and to get stable final results. The obtained rating points are used in decision making based on fuzzy expert opinions.

Three Computational Mathematics Techniques: Comparative Determination of Area under Curve

The objective of this manuscript is to find area under the plasma concentration- time curve (AUC) for multiple doses of salbutamol sulphate sustained release tablets (Ventolin® oral tablets SR 8 mg, GSK, Pakistan) in the group of 18 healthy adults by using computational mathematics techniques. Following the administration of 4 doses of Ventolin® tablets 12 hourly to 24 healthy human subjects and bioanalysis of obtained plasma samples, plasma drug concentration-time profile was constructed. AUC, an important pharmacokinetic parameter, was measured using integrated equation of multiple oral dose regimens. The approximated AUC was also calculated by using computational mathematics techniques such as repeated rectangular, repeated trapezium and repeated Simpson's rule and compared with exact value of AUC calculated by using integrated equation of multiple oral dose regimens to find best computational mathematics method that gives AUC values closest to exact. The exact values of AUC for four consecutive doses of Ventolin® oral tablets were 150.5819473, 157.8131756, 164.4178231 and 162.78 ng.h/ml while the closest values approximated AUC values were 149.245962, 157.336171, 164.2585768 and 162.289224 ng.h/ml, respectively as found by repeated rectangular rule. The errors in the approximated values of AUC were negligible. It is concluded that all computational tools approximated values of AUC accurately but the repeated rectangular rule gives slightly better approximated values of AUC as compared to repeated trapezium and repeated Simpson's rules.

Noise Factors of RFID-Aided Positioning

In recent years, Radio Frequency Identification (RFID) is followed with interest by many researches, especially for the purpose of indoor positioning as the innate properties of RFID are profitable for achieving it. A lot of algorithms or schemes are proposed to be used in the RFID-based positioning system, but most of them are lack of environmental consideration and it induces inaccuracy of application. In this research, a lot of algorithms and schemes of RFID indoor positioning are discussed to see whether effective or not on application, and some rules are summarized for achieving accurate positioning. On the other hand, a new term “Noise Factor" is involved to describe the signal loss between the target and the obstacle. As a result, experimental data can be obtained but not only simulation; and the performance of the positioning system can be expressed substantially.

Design, Modeling and Fabrication of a Tactile Sensor and Display System for Application in Laparoscopic Surgery

One of the major disadvantages of the minimally invasive surgery (MIS) is the lack of tactile feedback to the surgeon. In order to identify and avoid any damage to the grasped complex tissue by endoscopic graspers, it is important to measure the local softness of tissue during MIS. One way to display the measured softness to the surgeon is a graphical method. In this paper, a new tactile sensor has been reported. The tactile sensor consists of an array of four softness sensors, which are integrated into the jaws of a modified commercial endoscopic grasper. Each individual softness sensor consists of two piezoelectric polymer Polyvinylidene Fluoride (PVDF) films, which are positioned below a rigid and a compliant cylinder. The compliant cylinder is fabricated using a micro molding technique. The combination of output voltages from PVDF films is used to determine the softness of the grasped object. The theoretical analysis of the sensor is also presented. A method has been developed with the aim of reproducing the tactile softness to the surgeon by using a graphical method. In this approach, the proposed system, including the interfacing and the data acquisition card, receives signals from the array of softness sensors. After the signals are processed, the tactile information is displayed by means of a color coding method. It is shown that the degrees of softness of the grasped objects/tissues can be visually differentiated and displayed on a monitor.

A New Heuristic Approach to Solving U-shape Assembly Line Balancing Problems Type-1

Assembly line balancing is a very important issue in mass production systems due to production cost. Although many studies have been done on this topic, but because assembly line balancing problems are so complex they are categorized as NP-hard problems and researchers strongly recommend using heuristic methods. This paper presents a new heuristic approach called the critical task method (CTM) for solving U-shape assembly line balancing problems. The performance of the proposed heuristic method is tested by solving a number of test problems and comparing them with 12 other heuristics available in the literature to confirm the superior performance of the proposed heuristic. Furthermore, to prove the efficiency of the proposed CTM, the objectives are increased to minimize the number of workstation (or equivalently maximize line efficiency), and minimizing the smoothness index. Finally, it is proven that the proposed heuristic is more efficient than the others to solve the U-shape assembly line balancing problem.

Adaptive Digital Watermarking Integrating Fuzzy Inference HVS Perceptual Model

An adaptive Fuzzy Inference Perceptual model has been proposed for watermarking of digital images. The model depends on the human visual characteristics of image sub-regions in the frequency multi-resolution wavelet domain. In the proposed model, a multi-variable fuzzy based architecture has been designed to produce a perceptual membership degree for both candidate embedding sub-regions and strength watermark embedding factor. Different sizes of benchmark images with different sizes of watermarks have been applied on the model. Several experimental attacks have been applied such as JPEG compression, noises and rotation, to ensure the robustness of the scheme. In addition, the model has been compared with different watermarking schemes. The proposed model showed its robustness to attacks and at the same time achieved a high level of imperceptibility.

LAYMOD; A Layered and Modular Platform for CAx Collaboration Management and Supporting Product data Integration based on STEP Standard

Nowadays companies strive to survive in a competitive global environment. To speed up product development/modifications, it is suggested to adopt a collaborative product development approach. However, despite the advantages of new IT improvements still many CAx systems work separately and locally. Collaborative design and manufacture requires a product information model that supports related CAx product data models. To solve this problem many solutions are proposed, which the most successful one is adopting the STEP standard as a product data model to develop a collaborative CAx platform. However, the improvement of the STEP-s Application Protocols (APs) over the time, huge number of STEP AP-s and cc-s, the high costs of implementation, costly process for conversion of older CAx software files to the STEP neutral file format; and lack of STEP knowledge, that usually slows down the implementation of the STEP standard in collaborative data exchange, management and integration should be considered. In this paper the requirements for a successful collaborative CAx system is discussed. The STEP standard capability for product data integration and its shortcomings as well as the dominant platforms for supporting CAx collaboration management and product data integration are reviewed. Finally a platform named LAYMOD to fulfil the requirements of CAx collaborative environment and integrating the product data is proposed. The platform is a layered platform to enable global collaboration among different CAx software packages/developers. It also adopts the STEP modular architecture and the XML data structures to enable collaboration between CAx software packages as well as overcoming the STEP standard limitations. The architecture and procedures of LAYMOD platform to manage collaboration and avoid contradicts in product data integration are introduced.

Langmuir–Blodgett Films of Polyaniline for Efficient Detection of Uric Acid

Langmuir–Blodgett (LB) films of polyaniline (PANI) grown onto ITO coated glass substrates were utilized for the fabrication of Uric acid biosensor for efficient detection of uric acid by immobilizing Uricase via EDC–NHS coupling. The modified electrodes were characterized by atomic force microscopy (AFM). The response characteristics after immobilization of uricase were studied using cyclic voltammetry and electrochemical impedance spectroscopy techniques. The uricase/PANI/ITO/glass bioelectrode studied by CV and EIS techniques revealed detection of uric acid in a wide range of 0.05 mM to 1.0 mM, covering the physiological range in blood. A low Michaelis–Menten constant (Km) of 0.21 mM indicates the higher affinity of immobilized Uricase towards its analyte (uric acid). The fabricated uric acid biosensor based on PANI LB films exhibits excellent sensitivity of 0.21 mA/mM with a response time of 4 s, good reproducibility, long shelf life (8 weeks) and high selectivity.

Sensory Evaluation of Meatballs with Jerusalem Artichoke (Helianthus tuberosus L.)

Meat and meat products for human consumption are one of main sources of protein, amino acids, fatty acids, vitamins, and minerals. Popular variety of meat product is meatballs, which can be enriched with valuable product – Jerusalem artichoke powder, made from dried and grinded Jerusalem artichoke tubers, it is raw material with low-calorie, low fat, rich in dietary fibres, minerals, and vitamins. The results of this study indicate that that people could accept the new product - meatballs with Jerusalem artichoke powder and Jerusalem artichoke powder is suitable for meatballs preparation, in result them is possible to improve meatballs sensory and physical properties.

Recovering the Clipped OFDM Figurebased on the Conic Function

In Orthogonal Frequency Division Multiplexing (OFDM) systems, the peak to average power ratio (PAR) is much high. The clipping signal scheme is a useful method to reduce PAR. Clipping the OFDM signal, however, increases the overall noise level by introducing clipping noise. It is necessary to recover the figure of the original signal at receiver in order to reduce the clipping noise. Considering the continuity of the signal and the figure of the peak, we obtain a certain conic function curve to replace the clipped signal module within the clipping time. The results of simulation show that the proposed scheme can reduce the systems? BER (bit-error rate) 10 times when signal-to-interference-and noise-ratio (SINR) equals to 12dB. And the BER performance of the proposed scheme is superior to that of kim's scheme, too.

OCR for Script Identification of Hindi (Devnagari) Numerals using Error Diffusion Halftoning Algorithm with Neural Classifier

The applications on numbers are across-the-board that there is much scope for study. The chic of writing numbers is diverse and comes in a variety of form, size and fonts. Identification of Indian languages scripts is challenging problems. In Optical Character Recognition [OCR], machine printed or handwritten characters/numerals are recognized. There are plentiful approaches that deal with problem of detection of numerals/character depending on the sort of feature extracted and different way of extracting them. This paper proposes a recognition scheme for handwritten Hindi (devnagiri) numerals; most admired one in Indian subcontinent our work focused on a technique in feature extraction i.e. Local-based approach, a method using 16-segment display concept, which is extracted from halftoned images & Binary images of isolated numerals. These feature vectors are fed to neural classifier model that has been trained to recognize a Hindi numeral. The archetype of system has been tested on varieties of image of numerals. Experimentation result shows that recognition rate of halftoned images is 98 % compared to binary images (95%).

On Symmetry Analysis and Exact Wave Solutions of New Modified Novikov Equation

In this paper, we study a new modified Novikov equation for its classical and nonclassical symmetries and use the symmetries to reduce it to a nonlinear ordinary differential equation (ODE). With the aid of solutions of the nonlinear ODE by using the modified (G/G)-expansion method proposed recently, multiple exact traveling wave solutions are obtained and the traveling wave solutions are expressed by the hyperbolic functions, trigonometric functions and rational functions.

Communication and Quality in Distributed Agile Development: An Empirical Case Study

Through inward perceptions, we intuitively expect distributed software development to increase the risks associated with achieving cost, schedule, and quality goals. To compound this problem, agile software development (ASD) insists one of the main ingredients of its success is cohesive communication attributed to collocation of the development team. The following study identified the degree of communication richness needed to achieve comparable software quality (reduce pre-release defects) between distributed and collocated teams. This paper explores the relevancy of communication richness in various development phases and its impact on quality. Through examination of a large distributed agile development project, this investigation seeks to understand the levels of communication required within each ASD phase to produce comparable quality results achieved by collocated teams. Obviously, a multitude of factors affects the outcome of software projects. However, within distributed agile software development teams, the mode of communication is one of the critical components required to achieve team cohesiveness and effectiveness. As such, this study constructs a distributed agile communication model (DAC-M) for potential application to similar distributed agile development efforts using the measurement of the suitable level of communication. The results of the study show that less rich communication methods, in the appropriate phase, might be satisfactory to achieve equivalent quality in distributed ASD efforts.

Face Recognition using a Kernelization of Graph Embedding

Linearization of graph embedding has been emerged as an effective dimensionality reduction technique in pattern recognition. However, it may not be optimal for nonlinearly distributed real world data, such as face, due to its linear nature. So, a kernelization of graph embedding is proposed as a dimensionality reduction technique in face recognition. In order to further boost the recognition capability of the proposed technique, the Fisher-s criterion is opted in the objective function for better data discrimination. The proposed technique is able to characterize the underlying intra-class structure as well as the inter-class separability. Experimental results on FRGC database validate the effectiveness of the proposed technique as a feature descriptor.

Automatic Building an Extensive Arabic FA Terms Dictionary

Field Association (FA) terms are a limited set of discriminating terms that give us the knowledge to identify document fields which are effective in document classification, similar file retrieval and passage retrieval. But the problem lies in the lack of an effective method to extract automatically relevant Arabic FA Terms to build a comprehensive dictionary. Moreover, all previous studies are based on FA terms in English and Japanese, and the extension of FA terms to other language such Arabic could be definitely strengthen further researches. This paper presents a new method to extract, Arabic FA Terms from domain-specific corpora using part-of-speech (POS) pattern rules and corpora comparison. Experimental evaluation is carried out for 14 different fields using 251 MB of domain-specific corpora obtained from Arabic Wikipedia dumps and Alhyah news selected average of 2,825 FA Terms (single and compound) per field. From the experimental results, recall and precision are 84% and 79% respectively. Therefore, this method selects higher number of relevant Arabic FA Terms at high precision and recall.

A Universal Approach for the Intuitive Control of Mobile Robots using an AR/VR-based Interface

Mobile robots are used in a large field of scenarios, like exploring contaminated areas, repairing oil rigs under water, finding survivors in collapsed buildings, etc. Currently, there is no unified intuitive user interface (UI) to control such complex mobile robots. As a consequence, some scenarios are done without the exploitation of experience and intuition of human teleoperators. A novel framework has been developed to embed a flexible and modular UI into a complete 3-D virtual reality simulation system. This new approach wants to access maximum benefits of human operators. Sensor information received from the robot is prepared for an intuitive visualization. Virtual reality metaphors support the operator in his decisions. These metaphors are integrated into a real time stereo video stream. This approach is not restricted to any specific type of mobile robot and allows for the operation of different robot types with a consistent concept and user interface.

Flame Stability and Structure of Liquefied Petroleum Gas-Fired Inverse Diffusion Flame with Hydrogen Enrichment

The present project was conducted with the circumferential-fuel-jets inverse diffusion flame (CIDF) burner burning liquefied petroleum gas (LPG) enriched with 50% of hydrogen fuel (H2). The range of stable operation of the CIDF burner in terms of Reynolds number (from laminar to turbulent flow regions), equivalence ratio and fuel jet velocity of LPG of the 50% H2-LPG mixed fuel was identified. Experiments were also carried out to investigate the flame structures of the LPG flame and LPG enriched H2 flame. Experimental results obtained from these two flames were compared to fully explore the influence of hydrogen addition on flame stability. Flame heights obtained by burning these two kinds of fuels at various equivalence ratios were compared and correlated with the Global Momentum Ratio (GMR).

Response Quality Evaluation in Heterogeneous Question Answering System: A Black-box Approach

The evaluation of the question answering system is a major research area that needs much attention. Before the rise of domain-oriented question answering systems based on natural language understanding and reasoning, evaluation is never a problem as information retrieval-based metrics are readily available for use. However, when question answering systems began to be more domains specific, evaluation becomes a real issue. This is especially true when understanding and reasoning is required to cater for a wider variety of questions and at the same time achieve higher quality responses The research in this paper discusses the inappropriateness of the existing measure for response quality evaluation and in a later part, the call for new standard measures and the related considerations are brought forward. As a short-term solution for evaluating response quality of heterogeneous systems, and to demonstrate the challenges in evaluating systems of different nature, this research presents a black-box approach using observation, classification scheme and a scoring mechanism to assess and rank three example systems (i.e. AnswerBus, START and NaLURI).

The Effect of Interlamellar Distance in Pearlite on CGI Machining

Swedish truck industry is investigating the possibility for implementing the use of Compacted Graphite Iron (CGI) in their heavy duty diesel engines. Compared to the alloyed gray iron used today, CGI has superior mechanical properties but not as good machinability. Another issue that needs to be addressed when implementing CGI is the inhomogeneous microstructure when the cast component has different section thicknesses, as in cylinder blocks. Thinner sections results in finer pearlite, in the material, with higher strength. Therefore an investigation on its influence on machinability was needed. This paper focuses on the effect that interlamellar distance in pearlite has on CGI machinability and material physical properties. The effect of pearlite content and nodularity is also examined. The results showed that interlamellar distance in pearlite did not have as large effect on the material physical properties or machinability as pearlite content. The paper also shows the difficulties of obtaining a homogeneous microstructure in inhomogeneous workpieces.