Enhanced QoS Mechanisms for IEEE 802.11e Wireless Networks

The quality-of-service (QoS) support for wireless LANs has been a hot research topic during the past few years. In this paper, two QoS provisioning mechanisms are proposed for the employment in 802.11e EDCA MAC scheme. First, the proposed call admission control mechanism can not only guarantee the QoS for the higher priority existing connections but also provide the minimum reserved bandwidth for traffic flows with lower priority. In addition, the adaptive contention window adjustment mechanism can adjust the maximum and minimum contention window size dynamically according to the existing connection number of each AC. The collision probability as well as the packet delay will thus be reduced effectively. Performance results via simulations have revealed the enhanced QoS property achieved by employing these two mechanisms.

Normalized Cumulative Spectral Distribution in Music

As the remedy used music becomes active and meditation effect through the music is verified, people take a growing interest about psychological balance or remedy given by music. From traditional studies, it is verified that the music of which spectral envelop varies approximately as 1/f (f is frequency) down to a frequency of low frequency bandwidth gives psychological balance. In this paper, we researched signal properties of music which gives psychological balance. In order to find this, we derived the property from voice. Music composed by voice shows large value in NCSD. We confirmed the degree of deference between music by curvature of normalized cumulative spectral distribution. In the music that gives psychological balance, the curvature shows high value, otherwise, the curvature shows low value.

A New Heuristic Approach for the Large-Scale Generalized Assignment Problem

This paper presents a heuristic approach to solve the Generalized Assignment Problem (GAP) which is NP-hard. It is worth mentioning that many researches used to develop algorithms for identifying the redundant constraints and variables in linear programming model. Some of the algorithms are presented using intercept matrix of the constraints to identify redundant constraints and variables prior to the start of the solution process. Here a new heuristic approach based on the dominance property of the intercept matrix to find optimal or near optimal solution of the GAP is proposed. In this heuristic, redundant variables of the GAP are identified by applying the dominance property of the intercept matrix repeatedly. This heuristic approach is tested for 90 benchmark problems of sizes upto 4000, taken from OR-library and the results are compared with optimum solutions. Computational complexity is proved to be O(mn2) of solving GAP using this approach. The performance of our heuristic is compared with the best state-ofthe- art heuristic algorithms with respect to both the quality of the solutions. The encouraging results especially for relatively large size test problems indicate that this heuristic approach can successfully be used for finding good solutions for highly constrained NP-hard problems.

Skolem Sequences and Erdosian Labellings of m Paths with 2 and 3 Vertices

Assume that we have m identical graphs where the graphs consists of paths with k vertices where k is a positive integer. In this paper, we discuss certain labelling of the m graphs called c-Erdösian for some positive integers c. We regard labellings of the vertices of the graphs by positive integers, which induce the edge labels for the paths as the sum of the two incident vertex labels. They have the property that each vertex label and edge label appears only once in the set of positive integers {c, . . . , c+6m- 1}. Here, we show how to construct certain c-Erdösian of m paths with 2 and 3 vertices by using Skolem sequences.

Property of Polyurethane: from Soy-derived Phosphate Ester

Polyurethane foams (PUF) were formed by a chemical reaction of polyol and isocyanate. The polyol was manufactured by ring-opening hydrolysis of epoxidized soybean oil in the presence of phosphoric acid under varying experimental conditions. Other factors in the foam formulation such as water content and surfactant were kept constant. The effect of the amount of solvents, phosphoric acid, and their derivates in the foam formulation on the properties of polyurethane foams were studied. The properties of the material were measured via a number of parameters, which are water content of prepared polyol, polymer density and cellular structures.

Fingerprint Compression Using Contourlet Transform and Multistage Vector Quantization

This paper presents a new fingerprint coding technique based on contourlet transform and multistage vector quantization. Wavelets have shown their ability in representing natural images that contain smooth areas separated with edges. However, wavelets cannot efficiently take advantage of the fact that the edges usually found in fingerprints are smooth curves. This issue is addressed by directional transforms, known as contourlets, which have the property of preserving edges. The contourlet transform is a new extension to the wavelet transform in two dimensions using nonseparable and directional filter banks. The computation and storage requirements are the major difficulty in implementing a vector quantizer. In the full-search algorithm, the computation and storage complexity is an exponential function of the number of bits used in quantizing each frame of spectral information. The storage requirement in multistage vector quantization is less when compared to full search vector quantization. The coefficients of contourlet transform are quantized by multistage vector quantization. The quantized coefficients are encoded by Huffman coding. The results obtained are tabulated and compared with the existing wavelet based ones.

Experimental Determination of the Critical Locus of the Acetone + Chloroform Binary System

In this paper, vapour-liquid critical locus for the binary system acetone + chloroform was determined experimentally over the whole range of composition. The critical property measurements were carried out using a dynamic-synthetic apparatus, employed in the dynamic mode. The critical points are visually determined by observing the critical opalescence and the simultaneous disappearance and reappearance of the meniscus in the middle of a high-pressure view cell which withstands operations up to 673K and 20MPa. The experimental critical points measured in this work were compared to those available in literature.

3D Locomotion and Fractal Analysis of Goldfish for Acute Toxicity Bioassay

Biological reactions of individuals of a testing animal to toxic substance are unique and can be used as an indication of the existing of toxic substance. However, to distinguish such phenomenon need a very complicate system and even more complicate to analyze data in 3 dimensional. In this paper, a system to evaluate in vitro biological activities to acute toxicity of stochastic self-affine non-stationary signal of 3D goldfish swimming by using fractal analysis is introduced. Regular digital camcorders are utilized by proposed algorithm 3DCCPC to effectively capture and construct 3D movements of the fish. A Critical Exponent Method (CEM) has been adopted as a fractal estimator. The hypothesis was that the swimming of goldfish to acute toxic would show the fractal property which related to the toxic concentration. The experimental results supported the hypothesis by showing that the swimming of goldfish under the different toxic concentration has fractal properties. It also shows that the fractal dimension of the swimming related to the pH value of FD Ôëê 0.26pH + 0.05. With the proposed system, the fish is allowed to swim freely in all direction to react to the toxic. In addition, the trajectories are precisely evaluated by fractal analysis with critical exponent method and hence the results exhibit with much higher degree of confidence.

Development of 3D Coordinates and Damaged Point Detection System for Ducts using IMU

Recently, as the scale of construction projects has increases, more ground excavation for foundations is carried out than ever before. Consequently, damage to underground ducts (gas, water/sewage or oil pipelines, communication cables or power cable ducts) or superannuated pipelines frequently cause serious accidents resulting in damage to life and property. (In Korea, the total length of city water pipelines was approximately 2,000 km as of the end of 2009.) In addition, large amounts of damage caused by fractures, water and gas leakage caused by superannuation or damage to underground ducts in construction has been reported. Therefore, a system is required to precisely detect defects and deterioration in underground pipelines and the locations of such defects, for timely and accurate maintenance or replacement of the ducts. In this study, a system was developed which can locate underground structures (gas and water pipelines, power cable ducts, etc.) in 3D-coordinates and monitor the degree and position of defects using an Inertial Measurement Unit (IMU) sensing technique. The system can prevent damage to underground ducts and superannuated pipelines during construction, and provide reliable data for maintenance. The utility of the IMU sensing technique used in aircraft and ships in civil applications was verified.

A 1.2-ns16×16-Bit Binary Multiplier Using High Speed Compressors

For higher order multiplications, a huge number of adders or compressors are to be used to perform the partial product addition. We have reduced the number of adders by introducing special kind of adders that are capable to add five/six/seven bits per decade. These adders are called compressors. Binary counter property has been merged with the compressor property to develop high order compressors. Uses of these compressors permit the reduction of the vertical critical paths. A 16×16 bit multiplier has been developed using these compressors. These compressors make the multipliers faster as compared to the conventional design that have been used 4-2 compressors and 3-2 compressors.

Zero-knowledge-like Proof of Cryptanalysis of Bluetooth Encryption

This paper presents a protocol aiming at proving that an encryption system contains structural weaknesses without disclosing any information on those weaknesses. A verifier can check in a polynomial time that a given property of the cipher system output has been effectively realized. This property has been chosen by the prover in such a way that it cannot been achieved by known attacks or exhaustive search but only if the prover indeed knows some undisclosed weaknesses that may effectively endanger the cryptosystem security. This protocol has been denoted zero-knowledge-like proof of cryptanalysis. In this paper, we apply this protocol to the Bluetooth core encryption algorithm E0, used in many mobile environments and thus we suggest that its security can seriously be put into question.

5-Aminolevulinic Acid-Loaded Gel, Sponge Collagen to Enhance the Delivery Ability to Skin

Topical photodynamic therapy (PDT) with 5-aminolevulinic acid (ALA) is an alternative therapy for treating superficial cancer, especially for skin or oral cancer. ALA, a precursor of the photosensitizer protoporphyrin IX (PpIX), is present as zwitterions and hydrophilic property which make the low permeability through the cell membrane. Collagen is a traditional carrier; its molecular composed various amino acids which bear positive charge and negative charge. In order to utilize the ion-pairs with ALA and collagen, the study employed various pH values adjusting the net charge. The aim of this study was to compare a series collagen form, including solution, gel and sponge to investigate the topical delivery behavior of ALA. The in vivo confocal laser scanning microscopy (CLSM) study demonstrated that PpIX generation ability was different pattern after apply for 6 h. Gel type could generate high PpIX, and archived more deep of skin depth.

Stability Analysis for a Multicriteria Problem with Linear Criteria and Parameterized Principle of Optimality “from Lexicographic to Slater“

A multicriteria linear programming problem with integer variables and parameterized optimality principle "from lexicographic to Slater" is considered. A situation in which initial coefficients of penalty cost functions are not fixed but may be potentially a subject to variations is studied. For any efficient solution, appropriate measures of the quality are introduced which incorporate information about variations of penalty cost function coefficients. These measures correspond to the so-called stability and accuracy functions defined earlier for efficient solutions of a generic multicriteria combinatorial optimization problem with Pareto and lexicographic optimality principles. Various properties of such functions are studied and maximum norms of perturbations for which an efficient solution preserves the property of being efficient are calculated.

Regular Data Broadcasting Plan with Grouping in Wireless Mobile Environment

The broadcast problem including the plan design is considered. The data are inserted and numbered at predefined order into customized size relations. The server ability to create a full, regular Broadcast Plan (RBP) with single and multiple channels after some data transformations is examined. The Regular Geometric Algorithm (RGA) prepares a RBP and enables the users to catch their items avoiding energy waste of their devices. Moreover, the Grouping Dimensioning Algorithm (GDA) based on integrated relations can guarantee the discrimination of services with a minimum number of channels. This last property among the selfmonitoring, self-organizing, can be offered by servers today providing also channel availability and less energy consumption by using smaller number of channels. Simulation results are provided.

Integrating Hedgerow into Town Planning: A Framework for Sustainable Residential Development

The vast rural landscape in the southern United States is conspicuously characterized by the hedgerow trees or groves. The patchwork landscape of fields surrounded by high hedgerows is a traditional and familiar feature of the American countryside. Hedgerows are in effect linear strips of trees, groves, or woodlands, which are often critical habitats for wildlife and important for the visual quality of the landscape. As landscape interfaces, hedgerows define the spaces in the landscape, give the landscape life and meaning, and enrich ecologies and cultural heritages of the American countryside. Although hedgerows were originally intended as fences and to mark property and townland boundaries, they are not merely the natural or man-made additions to the landscape--they have gradually become “naturalized" into the landscape, deeply rooted in the rural culture, and now formed an important component of the southern American rural environment. However, due to the ever expanding real estate industry and high demand for new residential development, substantial areas of authentic hedgerow landscape in the southern United States are being urbanized. Using Hudson Farm as an example, this study illustrated guidelines of how hedgerows can be integrated into town planning as green infrastructure and landscape interface to innovate and direct sustainable land use, and suggest ways in which such vernacular landscapes can be preserved and integrated into new development without losing their contextual inspiration.

Techniques with Statistics for Web Page Watermarking

Information hiding, especially watermarking is a promising technique for the protection of intellectual property rights. This technology is mainly advanced for multimedia but the same has not been done for text. Web pages, like other documents, need a protection against piracy. In this paper, some techniques are proposed to show how to hide information in web pages using some features of the markup language used to describe these pages. Most of the techniques proposed here use the white space to hide information or some varieties of the language in representing elements. Experiments on a very small page and analysis of five thousands web pages show that these techniques have a wide bandwidth available for information hiding, and they might form a solid base to develop a robust algorithm for web page watermarking.

Dynamic Slope Scaling Procedure for Stochastic Integer Programming Problem

Mathematical programming has been applied to various problems. For many actual problems, the assumption that the parameters involved are deterministic known data is often unjustified. In such cases, these data contain uncertainty and are thus represented as random variables, since they represent information about the future. Decision-making under uncertainty involves potential risk. Stochastic programming is a commonly used method for optimization under uncertainty. A stochastic programming problem with recourse is referred to as a two-stage stochastic problem. In this study, we consider a stochastic programming problem with simple integer recourse in which the value of the recourse variable is restricted to a multiple of a nonnegative integer. The algorithm of a dynamic slope scaling procedure for solving this problem is developed by using a property of the expected recourse function. Numerical experiments demonstrate that the proposed algorithm is quite efficient. The stochastic programming model defined in this paper is quite useful for a variety of design and operational problems.

Landslide, Earthquake and Flood Hazard Risks of Izmir Metropolitan City, A Case: Altindag Landslide Areas

Urban disaster risks and vulnerabilities are great problems for Turkey. The annual loss of life and property through disaster in the world-s major metropolitan areas is increasing. Urban concentrations of the poor and less-informed in environmentally fragile locations suffer the impact of disaster disproportionately. Gecekondu (squatter) developments will compound the inherent risks associated with high-density environments, in appropriate technologies, and inadequate infrastructure. On the other hand, there are many geological disadvantages such as sitting on top of active tectonic plate boundaries, and why having avalanche, flood, and landslide and drought prone areas in Turkey. However, this natural formation is inevitable; the only way to survive in such a harsh geography is to be aware of importance of these natural events and to take political and physical measures. The main aim of this research is to bring up the magnitude of natural hazard risks in Izmir built-up zone, not being taken into consideration adequately. Because the dimensions of the peril are not taken seriously enough, the natural hazard risks, which are commonly well known, are not considered important or they are being forgotten after some time passes. Within this research, the magnitude of natural hazard risks for Izmir is being presented in the scope of concrete and local researches over Izmir risky areas.

Experimental Evaluation of Drilling Damage on the Strength of Cores Extracted from RC Buildings

Concrete strength evaluated from compression tests on cores is affected by several factors causing differences from the in-situ strength at the location from which the core specimen was extracted. Among the factors, there is the damage possibly occurring during the drilling phase that generally leads to underestimate the actual in-situ strength. In order to quantify this effect, in this study two wide datasets have been examined, including: (i) about 500 core specimens extracted from Reinforced Concrete existing structures, and (ii) about 600 cube specimens taken during the construction of new structures in the framework of routine acceptance control. The two experimental datasets have been compared in terms of compression strength and specific weight values, accounting for the main factors affecting a concrete property, that is type and amount of cement, aggregates' grading, type and maximum size of aggregates, water/cement ratio, placing and curing modality, concrete age. The results show that the magnitude of the strength reduction due to drilling damage is strongly affected by the actual properties of concrete, being inversely proportional to its strength. Therefore, the application of a single value of the correction coefficient, as generally suggested in the technical literature and in structural codes, appears inappropriate. A set of values of the drilling damage coefficient is suggested as a function of the strength obtained from compressive tests on cores.

Model Checking Consistency of UML Diagrams Using Alloy

In this paper, we proposed a method for detecting consistency violation between UML state machine diagrams and communication diagrams using Alloy. Using input language of Alloy, the proposed method expresses system behaviors described by state machine diagrams, message sequences described by communication diagrams, and a consistency property. As a result of application for an example system, we confirmed that consistency violation could be detected using Alloy correctly.