Wavelet-Based Despeckling of Synthetic Aperture Radar Images Using Adaptive and Mean Filters

In this paper we introduced new wavelet based algorithm for speckle reduction of synthetic aperture radar images, which uses combination of undecimated wavelet transformation, wiener filter (which is an adaptive filter) and mean filter. Further more instead of using existing thresholding techniques such as sure shrinkage, Bayesian shrinkage, universal thresholding, normal thresholding, visu thresholding, soft and hard thresholding, we use brute force thresholding, which iteratively run the whole algorithm for each possible candidate value of threshold and saves each result in array and finally selects the value for threshold that gives best possible results. That is why it is slow as compared to existing thresholding techniques but gives best results under the given algorithm for speckle reduction.

Asset Management for Educational Buildings in Egypt

In Egypt, the concept of Asset Management (AM) is new; however, the need for applying it has become crucial because deteriorating or losing an asset is unaffordable in a developing country like Egypt. Therefore the current study focuses on educational buildings as one of the most important assets regarding planning, building, operating and maintenance expenditures. The main objective of this study is to develop a SAMF for educational buildings in Egypt. The General Authority for Educational Buildings (GAEB) was chosen as a case study of the current research as it represents the biggest governmental organization responsible for planning, operating and maintaining schools in Egypt. To achieve the research objective, structured interviews were conducted with senior managers of GAEB using a pre designed questionnaire to explore the current practice of AM. Gab analysis technique was applied against best practices compounded from a vast literature review to identify gaps between current practices and the desired one. The previous steps mainly revealed; limited knowledge about strategic asset management, no clear goals, no training, no real risk plan and lack of data, technical and financial resources. Based on the findings, a SAMF for GAEB was introduced and Framework implementation steps and assessment techniques were explained in detail.

Classification and Resolving Urban Problems by Means of Fuzzy Approach

Urban problems are problems of organized complexity. Thus, many models and scientific methods to resolve urban problems are failed. This study is concerned with proposing of a fuzzy system driven approach for classification and solving urban problems. The proposed study investigated mainly the selection of the inputs and outputs of urban systems for classification of urban problems. In this research, five categories of urban problems, respect to fuzzy system approach had been recognized: control, polytely, optimizing, open and decision making problems. Grounded Theory techniques were then applied to analyze the data and develop new solving method for each category. The findings indicate that the fuzzy system methods are powerful processes and analytic tools for helping planners to resolve urban complex problems. These tools can be successful where as others have failed because both incorporate or address uncertainty and risk; complexity and systems interacting with other systems.

Enhance Performance of Secure Image Using Wavelet Compression

The increase popularity of multimedia application especially in image processing places a great demand on efficient data storage and transmission techniques. Network communication such as wireless network can easily be intercepted and cause of confidential information leaked. Unfortunately, conventional compression and encryption methods are too slow; it is impossible to carry out real time secure image processing. In this research, Embedded Zerotree Wavelet (EZW) encoder which specially designs for wavelet compression is examined. With this algorithm, three methods are proposed to reduce the processing time, space and security protection that will be secured enough to protect the data.

Automation of Fishhooks Objective Measures

Fishing has always been an essential component of the Polynesians- life. Fishhooks, mostly in pearl shell, found during archaeological excavations are the artifacts related to this activity the most numerous. Thanks to them, we try to reconstruct the ancient techniques of resources exploitation, inside the lagoons and offshore. They can also be used as chronological and cultural indicators. The shapes and dimensions of these artifacts allow comparisons and classifications used in both functional approach and chrono-cultural perspective. Hence it is very important for the ethno-archaeologists to dispose of reliable methods and standardized measurement of these artifacts. Such a reliable objective and standardized method have been previously proposed. But this method cannot be envisaged manually because of the very important time required to measure each fishhook manually and the quantity of fishhooks to measure (many hundreds). We propose in this paper a detailed acquisition protocol of fishhooks and an automation of every step of this method. We also provide some experimental results obtained on the fishhooks coming from three archaeological excavations sites.

Observations about the Principal Components Analysis and Data Clustering Techniques in the Study of Medical Data

The medical data statistical analysis often requires the using of some special techniques, because of the particularities of these data. The principal components analysis and the data clustering are two statistical methods for data mining very useful in the medical field, the first one as a method to decrease the number of studied parameters, and the second one as a method to analyze the connections between diagnosis and the data about the patient-s condition. In this paper we investigate the implications obtained from a specific data analysis technique: the data clustering preceded by a selection of the most relevant parameters, made using the principal components analysis. Our assumption was that, using the principal components analysis before data clustering - in order to select and to classify only the most relevant parameters – the accuracy of clustering is improved, but the practical results showed the opposite fact: the clustering accuracy decreases, with a percentage approximately equal with the percentage of information loss reported by the principal components analysis.

Particle Swarm Optimization for Design of Water Distribution Systems

Particle swarm optimization (PSO) technique is applied to design the water distribution pipeline network. A simulation-optimization model is formulated with the objective of minimizing cost and is applied to a benchmark water distribution system optimization problem. The benchmark problem taken for the application of PSO technique to optimize the pipe size of the water distribution network is New York City water supply system problem. The results from the analysis infer that PSO is a potential alternative optimization technique when compared to other heuristic techniques for optimal sizing of water distribution systems.

Layout Based Spam Filtering

Due to the constant increase in the volume of information available to applications in fields varying from medical diagnosis to web search engines, accurate support of similarity becomes an important task. This is also the case of spam filtering techniques where the similarities between the known and incoming messages are the fundaments of making the spam/not spam decision. We present a novel approach to filtering based solely on layout, whose goal is not only to correctly identify spam, but also warn about major emerging threats. We propose a mathematical formulation of the email message layout and based on it we elaborate an algorithm to separate different types of emails and find the new, numerically relevant spam types.

A Performance Comparison of Golay and Reed-Muller Coded OFDM Signal for Peak-to-Average Power Ratio Reduction

Multicarrier transmission system such as Orthogonal Frequency Division Multiplexing (OFDM) is a promising technique for high bit rate transmission in wireless communication systems. OFDM is a spectrally efficient modulation technique that can achieve high speed data transmission over multipath fading channels without the need for powerful equalization techniques. A major drawback of OFDM is the high Peak-to-Average Power Ratio (PAPR) of the transmit signal which can significantly impact the performance of the power amplifier. In this paper we have compared the PAPR reduction performance of Golay and Reed-Muller coded OFDM signal. From our simulation it has been found that the PAPR reduction performance of Golay coded OFDM is better than the Reed-Muller coded OFDM signal. Moreover, for the optimum PAPR reduction performance, code configuration for Golay and Reed-Muller codes has been identified.

On The Comparison of Fuzzy Logic and State Space Averaging based Sliding Control Methods Applied onan Arc Welding Machine

In this study, the performance of a high-frequency arc welding machine including a two-switch inverter is analyzed. The control of the system is achieved using two different control techniques i- fuzzy logic control (FLC) ii- state space averaging based sliding control. Fuzzy logic control does not need accurate mathematical model of a plant and can be used in nonlinear applications. The second method needs the mathematical model of the system. In this method the state space equations of the system are derived for two different “on" and “off" states of the switches. The derived state equations are combined with the sliding control rule considering the duty-cycle of the converter. The performance of the system is analyzed by simulating the system using SIMULINK tool box of MATLAB. The simulation results show that fuzzy logic controller is more robust and less sensitive to parameter variations.

An Experimental Study of a Self-Supervised Classifier Ensemble

Learning using labeled and unlabelled data has received considerable amount of attention in the machine learning community due its potential in reducing the need for expensive labeled data. In this work we present a new method for combining labeled and unlabeled data based on classifier ensembles. The model we propose assumes each classifier in the ensemble observes the input using different set of features. Classifiers are initially trained using some labeled samples. The trained classifiers learn further through labeling the unknown patterns using a teaching signals that is generated using the decision of the classifier ensemble, i.e. the classifiers self-supervise each other. Experiments on a set of object images are presented. Our experiments investigate different classifier models, different fusing techniques, different training sizes and different input features. Experimental results reveal that the proposed self-supervised ensemble learning approach reduces classification error over the single classifier and the traditional ensemble classifier approachs.

Dynamics in Tangible Chemical Reactions

Spatial understanding and the understanding of dynamic change in the spatial structure of molecules during a reaction is essential for designing new molecules. Knowing the physical processes in the reactions helps to speed up the designing process. To support the designer with the correct representation of the designed molecule as well as showing the dynamic behavior of the whole reacting system is the goal of our application. Our system shows the spatial deformation of the molecules at every time interval by minimizing the energy level of the molecules. The position and orientation of the molecules can be intuitively controlled by manipulating objects of the real world using Augmented Reality techniques. Our approach has the potential to speed up the design of new molecules and help students to understand the chemical processes better.

Automated Optic Disc Detection in Retinal Images of Patients with Diabetic Retinopathy and Risk of Macular Edema

In this paper, a new automated methodology to detect the optic disc (OD) automatically in retinal images from patients with risk of being affected by Diabetic Retinopathy (DR) and Macular Edema (ME) is presented. The detection procedure comprises two independent methodologies. On one hand, a location methodology obtains a pixel that belongs to the OD using image contrast analysis and structure filtering techniques and, on the other hand, a boundary segmentation methodology estimates a circular approximation of the OD boundary by applying mathematical morphology, edge detection techniques and the Circular Hough Transform. The methodologies were tested on a set of 1200 images composed of 229 retinographies from patients affected by DR with risk of ME, 431 with DR and no risk of ME and 540 images of healthy retinas. The location methodology obtained 98.83% success rate, whereas the OD boundary segmentation methodology obtained good circular OD boundary approximation in 94.58% of cases. The average computational time measured over the total set was 1.67 seconds for OD location and 5.78 seconds for OD boundary segmentation.

Dynamics and Control of a Chaotic Electromagnetic System

In this paper, different nonlinear dynamics analysis techniques are employed to unveil the rich nonlinear phenomena of the electromagnetic system. In particular, bifurcation diagrams, time responses, phase portraits, Poincare maps, power spectrum analysis, and the construction of basins of attraction are all powerful and effective tools for nonlinear dynamics problems. We also employ the method of Lyapunov exponents to show the occurrence of chaotic motion and to verify those numerical simulation results. Finally, two cases of a chaotic electromagnetic system being effectively controlled by a reference signal or being synchronized to another nonlinear electromagnetic system are presented.

Simulating and Forecasting Qualitative Marcoeconomic Models Using Rule-Based Fuzzy Cognitive Maps

Economic models are complex dynamic systems with a lot of uncertainties and fuzzy data. Conventional modeling approaches using well known methods and techniques cannot provide realistic and satisfactory answers to today-s challenging economic problems. Qualitative modeling using fuzzy logic and intelligent system theories can be used to model macroeconomic models. Fuzzy Cognitive maps (FCM) is a new method been used to model the dynamic behavior of complex systems. For the first time FCMs and the Mamdani Model of Intelligent control is used to model macroeconomic models. This new model is referred as the Mamdani Rule-Based Fuzzy Cognitive Map (MBFCM) and provides the academic and research community with a new promising integrated advanced computational model. A new economic model is developed for a qualitative approach to Macroeconomic modeling. Fuzzy Controllers for such models are designed. Simulation results for an economic scenario are provided and extensively discussed

Design Analysis of a Slotted Microstrip Antenna for Wireless Communication

In this paper, a new design technique for enhancing bandwidth that improves the performance of a conventional microstrip patch antenna is proposed. This paper presents a novel wideband probe fed inverted slotted microstrip patch antenna. The design adopts contemporary techniques; coaxial probe feeding, inverted patch structure and slotted patch. The composite effect of integrating these techniques and by introducing the proposed patch, offer a low profile, broadband, high gain, and low cross-polarization level. The results for the VSWR, gain and co-and cross-polarization patterns are presented. The antenna operating the band of 1.80-2.36 GHz shows an impedance bandwidth (2:1 VSWR) of 27% and a gain of 10.18 dBi with a gain variation of 1.12 dBi. Good radiation characteristics, including a cross-polarization level in xz-plane less than -42 dB, have been obtained.

The Effect of Waste Magnesium to Boric Acid Ratio in Hydrothermal Magnesium Borate Synthesis at 70oC

Magnesium wastes are produced by many industrial activities. This waste problem is becoming a future problem for the world. Magnesium borates have many advantages such as; high corrosion resistance, heat resistance, high coefficient of elasticity and can also be used in the production of material against radiation. Addition, magnesium borates have great potential in sectors including ceramic and detergents industry and superconducting materials. In this study, using the starting materials of waste magnesium and H3BO3 the hydrothermal method was applied at a moderate temperature of 70oC. Several mole ratios of waste magnesium to H3BO3 are selected as; 1:2, 1:4, 1:6, 1:8, 1:10. Reaction time was determined as 1 hour. After the synthesis, X-Ray Diffraction (XRD) and Fourier Transform Infrared Spectroscopy (FT-IR) techniques are applied to products. As a result the forms of mcallisterite “Mg2(B6O7(OH)6)2.9(H2O)”, admontite “MgO(B2O3)3.7(H2O)” and magnesium boron hydrate (MgO(B2O3)3.6(H2O)” are obtained. 

Processor Scheduling on Parallel Computers

Many problems in computer vision and image processing present potential for parallel implementations through one of the three major paradigms of geometric parallelism, algorithmic parallelism and processor farming. Static process scheduling techniques are used successfully to exploit geometric and algorithmic parallelism, while dynamic process scheduling is better suited to dealing with the independent processes inherent in the process farming paradigm. This paper considers the application of parallel or multi-computers to a class of problems exhibiting spatial data characteristic of the geometric paradigm. However, by using processor farming paradigm, a dynamic scheduling technique is developed to suit the MIMD structure of the multi-computers. A hybrid scheme of scheduling is also developed and compared with the other schemes. The specific problem chosen for the investigation is the Hough transform for line detection.

New Nonlinear Filtering Strategies for Eliminating Short and Long Tailed Noise in Images with Edge Preservation Properties

Midpoint filter is quite effective in recovering the images confounded by the short-tailed (uniform) noise. It, however, performs poorly in the presence of additive long-tailed (impulse) noise and it does not preserve the edge structures of the image signals. Median smoother discards outliers (impulses) effectively, but it fails to provide adequate smoothing for images corrupted with nonimpulse noise. In this paper, two nonlinear techniques for image filtering, namely, New Filter I and New Filter II are proposed based on a nonlinear high-pass filter algorithm. New Filter I is constructed using a midpoint filter, a highpass filter and a combiner. It suppresses uniform noise quite well. New Filter II is configured using an alpha trimmed midpoint filter, a median smoother of window size 3x3, the high pass filter and the combiner. It is robust against impulse noise and attenuates uniform noise satisfactorily. Both the filters are shown to exhibit good response at the image boundaries (edges). The proposed filters are evaluated for their performance on a test image and the results obtained are included.

M-band Wavelet and Cosine Transform Based Watermark Algorithm Using Randomization and Principal Component Analysis

Computational techniques derived from digital image processing are playing a significant role in the security and digital copyrights of multimedia and visual arts. This technology has the effect within the domain of computers. This research presents discrete M-band wavelet transform (MWT) and cosine transform (DCT) based watermarking algorithm by incorporating the principal component analysis (PCA). The proposed algorithm is expected to achieve higher perceptual transparency. Specifically, the developed watermarking scheme can successfully resist common signal processing, such as geometric distortions, and Gaussian noise. In addition, the proposed algorithm can be parameterized, thus resulting in more security. To meet these requirements, the image is transformed by a combination of MWT & DCT. In order to improve the security further, we randomize the watermark image to create three code books. During the watermark embedding, PCA is applied to the coefficients in approximation sub-band. Finally, first few component bands represent an excellent domain for inserting the watermark.