Speckle Characterization in Laser Projector Display

Speckle phenomena results from when coherent radiation is reflected from a rough surface. Characterizing the speckle strongly depends on the measurement condition and experimental setup. In this paper we report the experimental results produced with different parameters in the setup. We investigated the factors which affects the speckle contrast, such as, F-number, gamma value and exposure time of the camera, rather than geometric factors like the distance between the projector lens to the screen, the viewing distance, etc. The measurement results show that the speckle contrast decreases by decreasing F-number, by increasing gamma value, and slightly affects by exposure time of the camera and the gain value of the camera.

Performance Modeling and Availability Analysis of Yarn Dyeing System of a Textile Industry

This paper discusses the performance modeling and availability analysis of Yarn Dyeing System of a Textile Industry. The Textile Industry is a complex and repairable engineering system. Yarn Dyeing System of Textile Industry consists of five subsystems arranged in series configuration. For performance modeling and analysis of availability, a performance evaluating model has been developed with the help of mathematical formulation based on Markov-Birth-Death Process. The differential equations have been developed on the basis of Probabilistic Approach using a Transition Diagram. These equations have further been solved using normalizing condition in order to develop the steady state availability, a performance measure of the system concerned. The system performance has been further analyzed with the help of decision matrices. These matrices provide various availability levels for different combinations of failure and repair rates for various subsystems. The findings of this paper are therefore, considered to be useful for the analysis of availability and determination of the best possible maintenance strategies which can be implemented in future to enhance the system performance.

Replicating Data Objects in Large-scale Distributed Computing Systems using Extended Vickrey Auction

This paper proposes a novel game theoretical technique to address the problem of data object replication in largescale distributed computing systems. The proposed technique draws inspiration from computational economic theory and employs the extended Vickrey auction. Specifically, players in a non-cooperative environment compete for server-side scarce memory space to replicate data objects so as to minimize the total network object transfer cost, while maintaining object concurrency. Optimization of such a cost in turn leads to load balancing, fault-tolerance and reduced user access time. The method is experimentally evaluated against four well-known techniques from the literature: branch and bound, greedy, bin-packing and genetic algorithms. The experimental results reveal that the proposed approach outperforms the four techniques in both the execution time and solution quality.

Enhanced Spectral Envelope Coding Based On NLMS for G.729.1

In this paper, a new encoding algorithm of spectral envelope based on NLMS in G.729.1 for VoIP is proposed. In the TDAC part of G.729.1, the spectral envelope and MDCT coefficients extracted in the weighted CELP coding error (lower-band) and the higher-band input signal are encoded. In order to reduce allocation bits for spectral envelope coding, a new quantization algorithm based on NLMS is proposed. Also, reduced bits are used to enhance sound quality. The performance of the proposed algorithm is evaluated by sound quality and bit reduction rates in clean and frame loss conditions.

Dielectric Studies on Nano Zirconium Dioxide Synthesized through Co-Precipitation Process

Nano sized zirconium dioxide in monoclinic phase (m-ZrO2) has been synthesized in pure form through co-precipitation processing at different calcination temperatures and has been characterized by several techniques such as XRD, FT-IR, UV-Vis Spectroscopy and SEM. The dielectric and capacitance values of the pelletized samples have been examined at room temperature as the functions of frequency. The higher dielectric constant value of the sample having larger grain size proves the strong influence of grain size on the dielectric constant.

TFRank: An Evaluation of Users Importance with Fractal Views in Social Networks

One of research issues in social network analysis is to evaluate the position/importance of users in social networks. As the information diffusion in social network is evolving, it seems difficult to evaluate the importance of users using traditional approaches. In this paper, we propose an evaluation approach for user importance with fractal view in social networks. In this approach, the global importance (Fractal Importance) and the local importance (Topological Importance) of nodes are considered. The basic idea is that the bigger the product of fractal importance and topological importance of a node is, the more important of the node is. We devise the algorithm called TFRank corresponding to the proposed approach. Finally, we evaluate TFRank by experiments. Experimental results demonstrate our TFRank has the high correlations with PageRank algorithm and potential ranking algorithm, and it shows the effectiveness and advantages of our approach.

Investigation of the GFR2400 Reactivity Control System

The presented paper is related to the design methods and neutronic characterization of the reactivity control system in the large power unit of Generation IV Gas cooled Fast Reactor – GFR2400. The reactor core is based on carbide pin fuel type with the application of refractory metallic liners used to enhance the fission product retention of the SiCcladding. The heterogeneous design optimization of control rod is presented and the results of rods worth and their interferences in a core are evaluated. In addition, the idea of reflector removal as an additive reactivity management option is investigated and briefly described.

Reduced Inventories, High Reliability and Short Throughput Times by Using CONWIP Production Planning System

CONWIP (constant work-in-process) as a pull production system have been widely studied by researchers to date. The CONWIP pull production system is an alternative to pure push and pure pull production systems. It lowers and controls inventory levels which make the throughput better, reduces production lead time, delivery reliability and utilization of work. In this article a CONWIP pull production system was simulated. It was simulated push and pull planning system. To compare these systems via a production planning system (PPS) game were adjusted parameters of each production planning system. The main target was to reduce the total WIP and achieve throughput and delivery reliability to minimum values. Data was recorded and evaluated. A future state was made for real production of plastic components and the setup of the two indicators with CONWIP pull production system which can greatly help the company to be more competitive on the market.

An Efficient Method for Solving Multipoint Equation Boundary Value Problems

In this work, we solve multipoint boundary value problems where the boundary value conditions are equations using the Newton-Broyden Shooting method (NBSM).The proposed method is tested upon several problems from the literature and the results are compared with the available exact solution. The experiments are given to illustrate the efficiency and implementation of the method.

A Multiple-Objective Environmental Rationalization and Optimization for Material Substitution in the Production of Stone-Washed Jeans- Garments

As the Textile Industry is the second largest industry in Egypt and as small and medium-sized enterprises (SMEs) make up a great portion of this industry therein it is essential to apply the concept of Cleaner Production for the purpose of reducing pollution. In order to achieve this goal, a case study concerned with ecofriendly stone-washing of jeans-garments was investigated. A raw material-substitution option was adopted whereby the toxic potassium permanganate and sodium sulfide were replaced by the environmentally compatible hydrogen peroxide and glucose respectively where the concentrations of both replaced chemicals together with the operating time were optimized. In addition, a process-rationalization option involving four additional processes was investigated. By means of criteria such as product quality, effluent analysis, mass and heat balance; and cost analysis with the aid of a statistical model, a process optimization treatment revealed that the superior process optima were 50%, 0.15% and 50min for H2O2 concentration, glucose concentration and time, respectively. With these values the superior process ought to reduce the annual cost by about EGP 105 relative to the currently used conventional method.

A Study on the Quality of Hexapod Machine Tool's Workspace

One of the main concerns about parallel mechanisms is the presence of singular points within their workspaces. In singular positions the mechanism gains or loses one or several degrees of freedom. It is impossible to control the mechanism in singular positions. Therefore, these positions have to be avoided. This is a vital need especially in computer controlled machine tools designed and manufactured on the basis of parallel mechanisms. This need has to be taken into consideration when selecting design parameters. A prerequisite to this is a thorough knowledge about the effect of design parameters and constraints on singularity. In this paper, quality condition index was introduced as a criterion for evaluating singularities of different configurations of a hexapod mechanism obtainable by different design parameters. It was illustrated that this method can effectively be employed to obtain the optimum configuration of hexapod mechanism with the aim of avoiding singularity within the workspace. This method was then employed to design the hexapod table of a CNC milling machine.

Investigation of Corona wind Effect on Heat and Mass Transfer Enhancement

Applying corona wind as a novel technique can lead to a great level of heat and mass transfer augmentation by using very small amount of energy. Enhancement of forced flow evaporation rate by applying electric field (corona wind) has been experimentally evaluated in this study. Corona wind produced by a fine wire electrode which is charged with positive high DC voltage impinges to water surface and leads to evaporation enhancement by disturbing the saturated air layer over water surface. The study was focused on the effect of corona wind velocity, electrode spacing and air flow velocity on the level of evaporation enhancement. Two sets of experiments, i.e. with and without electric field, have been conducted. Data obtained from the first experiment were used as reference for evaluation of evaporation enhancement at the presence of electric field. Applied voltages ranged from corona threshold voltage to spark over voltage at 1 kV increments. The results showed that corona wind has great enhancement effect on water evaporation rate, but its effectiveness gradually diminishes by increasing air flow velocity. Maximum enhancements were 7.3 and 3.6 for air velocities of 0.125 and 1.75 m/s, respectively.

Fuzzy Clustering Analysis in Real Estate Companies in China

This paper applies fuzzy clustering algorithm in classifying real estate companies in China according to some general financial indexes, such as income per share, share accumulation fund, net profit margins, weighted net assets yield and shareholders' equity. By constructing and normalizing initial partition matrix, getting fuzzy similar matrix with Minkowski metric and gaining the transitive closure, the dynamic fuzzy clustering analysis for real estate companies is shown clearly that different clustered result change gradually with the threshold reducing, and then, it-s shown there is the similar relationship with the prices of those companies in stock market. In this way, it-s great valuable in contrasting the real estate companies- financial condition in order to grasp some good chances of investment, and so on.

Svision: Visual Identification of Scanning and Denial of Service Attacks

We propose a novel graphical technique (SVision) for intrusion detection, which pictures the network as a community of hosts independently roaming in a 3D space defined by the set of services that they use. The aim of SVision is to graphically cluster the hosts into normal and abnormal ones, highlighting only the ones that are considered as a threat to the network. Our experimental results using DARPA 1999 and 2000 intrusion detection and evaluation datasets show the proposed technique as a good candidate for the detection of various threats of the network such as vertical and horizontal scanning, Denial of Service (DoS), and Distributed DoS (DDoS) attacks.

Preliminary Study on Determining Stem Diameter Variations of Sympodial Orchid

Changes in stem diameter of orchid plants were investigated in a control growing climate. Previous studies have focused on stem diameter in relation to plant water on terrestrial plants in order to schedule the irrigation. The objective of this work was to evaluate the ability of the strain gauges to capture changes in the epiphytes plant stem. Experiments were carried out by using the sympodial orchid, Dendrobium Sonia in a stressed condition. From the findings, the sensor can detect changes in the plant stem and the result can easily be used as a reference for further studies for the development of a proper watering system.

An Edge Detection and Filtering Mechanism of Two Dimensional Digital Objects Based on Fuzzy Inference

The general idea behind the filter is to average a pixel using other pixel values from its neighborhood, but simultaneously to take care of important image structures such as edges. The main concern of the proposed filter is to distinguish between any variations of the captured digital image due to noise and due to image structure. The edges give the image the appearance depth and sharpness. A loss of edges makes the image appear blurred or unfocused. However, noise smoothing and edge enhancement are traditionally conflicting tasks. Since most noise filtering behaves like a low pass filter, the blurring of edges and loss of detail seems a natural consequence. Techniques to remedy this inherent conflict often encompass generation of new noise due to enhancement. In this work a new fuzzy filter is presented for the noise reduction of images corrupted with additive noise. The filter consists of three stages. (1) Define fuzzy sets in the input space to computes a fuzzy derivative for eight different directions (2) construct a set of IFTHEN rules by to perform fuzzy smoothing according to contributions of neighboring pixel values and (3) define fuzzy sets in the output space to get the filtered and edged image. Experimental results are obtained to show the feasibility of the proposed approach with two dimensional objects.

A Similarity Function for Global Quality Assessment of Retinal Vessel Segmentations

Retinal vascularity assessment plays an important role in diagnosis of ophthalmic pathologies. The employment of digital images for this purpose makes possible a computerized approach and has motivated development of many methods for automated vascular tree segmentation. Metrics based on contingency tables for binary classification have been widely used for evaluating performance of these algorithms and, concretely, the accuracy has been mostly used as measure of global performance in this topic. However, this metric shows very poor matching with human perception as well as other notable deficiencies. Here, a new similarity function for measuring quality of retinal vessel segmentations is proposed. This similarity function is based on characterizing the vascular tree as a connected structure with a measurable area and length. Tests made indicate that this new approach shows better behaviour than the current one does. Generalizing, this concept of measuring descriptive properties may be used for designing functions for measuring more successfully segmentation quality of other complex structures.

The Suitability of GPS Receivers Update Rates for Navigation Applications

Navigation is the processes of monitoring and controlling the movement of an object from one place to another. Currently, Global Positioning System (GPS) is the main navigation system used all over the world for navigation applications. GPS receiver receives signals from at least three satellites to locate and display itself. Displayed positioning information is updated continuously. Update rate is the number of times per second that a display is illuminated. The speed of update is governed by receiver update rate. A higher update rate decreases display lag time and improves distance measurements and tracking especially when moving on a curvy route. The majority of GPS receivers used nowadays are updated every second continuously. This period is considered reasonable for some applications while it is long relatively for high speed applications. In this paper, the suitability and feasibility of GPS receiver with different update rates will be evaluated for various applications according to the level of speed and update rate needed for particular applications.

Study on Characterization of Tuncbilek Fly Ash

Fly ash is one of the residues generated in combustion, and comprises the fine particles that rise with the flue gases. Ash which does not rise is termed bottom ash [1]. In our country, it is expected that will be occurred 50 million tons of waste ash per year until 2020. Released waste from the thermal power plants is caused very significant problems as known. The fly ashes can be evaluated by using as adsorbent material. The purpose of this study is to investigate the possibility of use of Tuncbilek fly ash like low-cost adsorbents for heavy metal adsorption. First of all, Tuncbilek fly ash was characterized. For this purpose; analysis such as sieve analysis, XRD, XRF, SEM and FT-IR were performed.

Integrating Security Indifference Curve to Formal Decision Evaluation

Decisions are regularly made during a project or daily life. Some decisions are critical and have a direct impact on project or human success. Formal evaluation is thus required, especially for crucial decisions, to arrive at the optimal solution among alternatives to address issues. According to microeconomic theory, all people-s decisions can be modeled as indifference curves. The proposed approach supports formal analysis and decision by constructing indifference curve model from the previous experts- decision criteria. These knowledge embedded in the system can be reused or help naïve users select alternative solution of the similar problem. Moreover, the method is flexible to cope with unlimited number of factors influencing the decision-making. The preliminary experimental results of the alternative selection are accurately matched with the expert-s decisions.