Some (v + 1, b + r + λ + 1, r + λ + 1, k, λ + 1) Balanced Incomplete Block Designs (BIBDs) from Lotto Designs (LDs)

The paper considered the construction of BIBDs using potential Lotto Designs (LDs) earlier derived from qualifying parent BIBDs. The study utilized Li’s condition  pr t−1  ( t−1 2 ) + pr− pr t−1 (t−1) 2  < ( p 2 ) λ, to determine the qualification of a parent BIBD (v, b, r, k, λ) as LD (n, k, p, t) constrained on v ≥ k, v ≥ p, t ≤ min{k, p} and then considered the case k = t since t is the smallest number of tickets that can guarantee a win in a lottery. The (15, 140, 28, 3, 4) and (7, 7, 3, 3, 1) BIBDs were selected as parent BIBDs to illustrate the procedure. These BIBDs yielded three potential LDs each. Each of the LDs was completely generated and their properties studied. The three LDs from the (15, 140, 28, 3, 4) produced (9, 84, 28, 3, 7), (10, 120, 36, 3, 8) and (11, 165, 45, 3, 9) BIBDs while those from the (7, 7, 3, 3, 1) produced the (5, 10, 6, 3, 3), (6, 20, 10, 3, 4) and (7, 35, 15, 3, 5) BIBDs. The produced BIBDs follow the generalization (v + 1, b + r + λ + 1, r +λ+1, k, λ+1) where (v, b, r, k, λ) are the parameters of the (9, 84, 28, 3, 7) and (5, 10, 6, 3, 3) BIBDs. All the BIBDs produced are unreduced designs.

Using Case-Based Reasoning to New Service Development from User Innovation Community in Mobile Application Services

The emergence of mobile application services and App Store has led to the explosive growth of user innovation, which users voluntarily contribute to. User innovation communities where end users freely reveal innovative ideas and needs with other community members are becoming increasingly influential in this area. However, user-s ideas in user innovation community are not enough to be new service opportunity, because some of them can already developed as existing services in App Store. Moreover, the existing services similar to new service opportunity can be significant references to apply analogy to develop service concept. In response, this research proposes Case-Based Reasoning approach to matching the user needs and existing services, identifying unmet opportunistic user needs, and retrieving similar services with opportunity. Due to its intuitive and transparent algorithm, users related to App Store innovation communities can easily employ Case-Based Reasoning based approach to their innovation.

A Study on Algorithm Fusion for Recognition and Tracking of Moving Robot

This paper presents an algorithm for the recognition and tracking of moving objects, 1/10 scale model car is used to verify performance of the algorithm. Presented algorithm for the recognition and tracking of moving objects in the paper is as follows. SURF algorithm is merged with Lucas-Kanade algorithm. SURF algorithm has strong performance on contrast, size, rotation changes and it recognizes objects but it is slow due to many computational complexities. Processing speed of Lucas-Kanade algorithm is fast but the recognition of objects is impossible. Its optical flow compares the previous and current frames so that can track the movement of a pixel. The fusion algorithm is created in order to solve problems which occurred using the Kalman Filter to estimate the position and the accumulated error compensation algorithm was implemented. Kalman filter is used to create presented algorithm to complement problems that is occurred when fusion two algorithms. Kalman filter is used to estimate next location, compensate for the accumulated error. The resolution of the camera (Vision Sensor) is fixed to be 640x480. To verify the performance of the fusion algorithm, test is compared to SURF algorithm under three situations, driving straight, curve, and recognizing cars behind the obstacles. Situation similar to the actual is possible using a model vehicle. Proposed fusion algorithm showed superior performance and accuracy than the existing object recognition and tracking algorithms. We will improve the performance of the algorithm, so that you can experiment with the images of the actual road environment.

On the Quantizer Design for Base Station Cooperation Systems with SC-FDE Techniques

By employing BS (Base Station) cooperation we can increase substantially the spectral efficiency and capacity of cellular systems. The signals received at each BS are sent to a central unit that performs the separation of the different MT (Mobile Terminal) using the same physical channel. However, we need accurate sampling and quantization of those signals so as to reduce the backhaul communication requirements. In this paper we consider the optimization of the quantizers for BS cooperation systems. Four different quantizer types are analyzed and optimized to allow better SQNR (Signal-to-Quantization Noise Ratio) and BER (Bit Error Rate) performance.

Intellectual Capital Report for Universities

Intellectual capital reporting becomes critical at universities, mainly due to the fact that knowledge is the main output as well as input in these institutions. In addition, universities have continuous external demands for greater information and transparency about the use of public funds, and are increasingly provided with greater autonomy regarding their organization, management, and budget allocation. This situation requires new management and reporting systems. The purpose of the present study is to provide a model for intellectual capital report in Spanish universities. To this end, a questionnaire was sent to every member of the Social Councils of Spanish public universities in order to identify which intangible elements university stakeholders demand most. Our proposal for an intellectual capital report aims to act as a guide to help the Spanish universities on the road to the presentation of information on intellectual capital which can assist stakeholders to make the right decisions.

Quantitative Evaluation of Frameworks for Web Applications

An empirical study of web applications that use software frameworks is presented here. The analysis is based on two approaches. In the first, developers using such frameworks are required, based on their experience, to assign weights to parameters such as database connection. In the second approach, a performance testing tool, OpenSTA, is used to compute start time and other such measures. From such an analysis, it is concluded that open source software is superior to proprietary software. The motivation behind this research is to examine ways in which a quantitative assessment can be made of software in general and frameworks in particular. Concepts such as metrics and architectural styles are discussed along with previously published research.

A Neural-Network-Based Fault Diagnosis Approach for Analog Circuits by Using Wavelet Transformation and Fractal Dimension as a Preprocessor

This paper presents a new method of analog fault diagnosis based on back-propagation neural networks (BPNNs) using wavelet decomposition and fractal dimension as preprocessors. The proposed method has the capability to detect and identify faulty components in an analog electronic circuit with tolerance by analyzing its impulse response. Using wavelet decomposition to preprocess the impulse response drastically de-noises the inputs to the neural network. The second preprocessing by fractal dimension can extract unique features, which are the fed to a neural network as inputs for further classification. A comparison of our work with [1] and [6], which also employs back-propagation (BP) neural networks, reveals that our system requires a much smaller network and performs significantly better in fault diagnosis of analog circuits due to our proposed preprocessing techniques.

Matrix-Interleaved Serially Concatenated Block Codes for Speech Transmission in Fixed Wireless Communication Systems

In this paper, we study a class of serially concatenated block codes (SCBC) based on matrix interleavers, to be employed in fixed wireless communication systems. The performances of SCBC¬coded systems are investigated under various interleaver dimensions. Numerical results reveal that the matrix interleaver could be a competitive candidate over conventional block interleaver for frame lengths of 200 bits; hence, the SCBC coding based on matrix interleaver is a promising technique to be employed for speech transmission applications in many international standards such as pan-European Global System for Mobile communications (GSM), Digital Cellular Systems (DCS) 1800, and Joint Detection Code Division Multiple Access (JD-CDMA) mobile radio systems, where the speech frame contains around 200 bits.

Reciprocating Compressor Optimum Design and Manufacturing with Respect to Performance, Reliability and Cost

Reciprocating compressors are flexible to handle wide capacity and condition swings, offer a very efficient method of compressing almost any gas mixture in wide range of pressure, can generate high head independent of density, and have numerous applications and wide power ratings. These make them vital component in various units of industrial plants. In this paper optimum reciprocating compressor configuration regarding interstage pressures, low suction pressure, non-lubricated cylinder, speed of machine, capacity control system, compressor valve, lubrication system, piston rod coating, cylinder liner material, barring device, pressure drops, rod load, pin reversal, discharge temperature, cylinder coolant system, performance, flow, coupling, special tools, condition monitoring (including vibration, thermal and rod drop monitoring), commercial points, delivery and acoustic conditions are presented.

The Students' Learning Effects on Dance Domain of Arts Education

The purpose of this study was to explore the learning effects on dance domain in Arts Curriculum at junior and senior high levels. A total of 1,366 students from 9th to 11th grade of different areas from Taiwan were administered a self-designed dance achievement test. Data were analyzed through descriptive analysis, independent sample t test, one-way ANOVA and Post hoc comparison analysis using Scheffé Test. The results showed (1) female students

Weed Classification using Histogram Maxima with Threshold for Selective Herbicide Applications

Information on weed distribution within the field is necessary to implement spatially variable herbicide application. Since hand labor is costly, an automated weed control system could be feasible. This paper deals with the development of an algorithm for real time specific weed recognition system based on Histogram Maxima with threshold of an image that is used for the weed classification. This algorithm is specifically developed to classify images into broad and narrow class for real-time selective herbicide application. The developed system has been tested on weeds in the lab, which have shown that the system to be very effectiveness in weed identification. Further the results show a very reliable performance on images of weeds taken under varying field conditions. The analysis of the results shows over 95 percent classification accuracy over 140 sample images (broad and narrow) with 70 samples from each category of weeds.

Cryptography Over Elliptic Curve Of The Ring Fq[e], e4 = 0

Groups where the discrete logarithm problem (DLP) is believed to be intractable have proved to be inestimable building blocks for cryptographic applications. They are at the heart of numerous protocols such as key agreements, public-key cryptosystems, digital signatures, identification schemes, publicly verifiable secret sharings, hash functions and bit commitments. The search for new groups with intractable DLP is therefore of great importance.The goal of this article is to study elliptic curves over the ring Fq[], with Fq a finite field of order q and with the relation n = 0, n ≥ 3. The motivation for this work came from the observation that several practical discrete logarithm-based cryptosystems, such as ElGamal, the Elliptic Curve Cryptosystems . In a first time, we describe these curves defined over a ring. Then, we study the algorithmic properties by proposing effective implementations for representing the elements and the group law. In anther article we study their cryptographic properties, an attack of the elliptic discrete logarithm problem, a new cryptosystem over these curves.

Hiding Data in Images Using PCP

In recent years, everything is trending toward digitalization and with the rapid development of the Internet technologies, digital media needs to be transmitted conveniently over the network. Attacks, misuse or unauthorized access of information is of great concern today which makes the protection of documents through digital media a priority problem. This urges us to devise new data hiding techniques to protect and secure the data of vital significance. In this respect, steganography often comes to the fore as a tool for hiding information. Steganography is a process that involves hiding a message in an appropriate carrier like image or audio. It is of Greek origin and means "covered or hidden writing". The goal of steganography is covert communication. Here the carrier can be sent to a receiver without any one except the authenticated receiver only knows existence of the information. Considerable amount of work has been carried out by different researchers on steganography. In this work the authors propose a novel Steganographic method for hiding information within the spatial domain of the gray scale image. The proposed approach works by selecting the embedding pixels using some mathematical function and then finds the 8 neighborhood of the each selected pixel and map each bit of the secret message in each of the neighbor pixel coordinate position in a specified manner. Before embedding a checking has been done to find out whether the selected pixel or its neighbor lies at the boundary of the image or not. This solution is independent of the nature of the data to be hidden and produces a stego image with minimum degradation.

Transmission Performance of Millimeter Wave Multiband OFDM UWB Wireless Signal over Fiber System

Performance of millimeter-wave (mm-wave) multiband orthogonal frequency division multiplexing (MB-OFDM) ultrawideband (UWB) signal generation using frequency quadrupling technique and transmission over fiber is experimentally investigated. The frequency quadrupling is achived by using only one Mach- Zehnder modulator (MZM) that is biased at maximum transmission (MATB) point. At the output, a frequency quadrupling signal is obtained then sent to a second MZM. This MZM is used for MBOFDM UWB signal modulation. In this work, we demonstrate 30- GHz mm-wave wireless that carries three-bands OFDM UWB signals, and error vector magnitude (EVM) is used to analyze the transmission quality. It is found that our proposed technique leads to an improvement of 3.5 dB in EVM at 40% of local oscillator (LO) modulation with comparison to the technique using two cascaded MZMs biased at minimum transmission (MITB) point.

A Novel Convergence Accelerator for the LMS Adaptive Algorithm

The least mean square (LMS) algorithmis one of the most well-known algorithms for mobile communication systems due to its implementation simplicity. However, the main limitation is its relatively slow convergence rate. In this paper, a booster using the concept of Markov chains is proposed to speed up the convergence rate of LMS algorithms. The nature of Markov chains makes it possible to exploit the past information in the updating process. Moreover, since the transition matrix has a smaller variance than that of the weight itself by the central limit theorem, the weight transition matrix converges faster than the weight itself. Accordingly, the proposed Markov-chain based booster thus has the ability to track variations in signal characteristics, and meanwhile, it can accelerate the rate of convergence for LMS algorithms. Simulation results show that the LMS algorithm can effectively increase the convergence rate and meantime further approach the Wiener solution, if the Markov-chain based booster is applied. The mean square error is also remarkably reduced, while the convergence rate is improved.

Buckling Analysis of Rectangular Plates under the Combined Action of Shear and Uniaxial Stresses

In the classical buckling analysis of rectangular plates subjected to the concurrent action of shear and uniaxial forces, the Euler shear buckling stress is generally evaluated separately, so that no influence on the shear buckling coefficient, due to the in-plane tensile or compressive forces, is taken into account. In this paper the buckling problem of simply supported rectangular plates, under the combined action of shear and uniaxial forces, is discussed from the beginning, in order to obtain new project formulas for the shear buckling coefficient that take into account the presence of uniaxial forces. Furthermore, as the classical expression of the shear buckling coefficient for simply supported rectangular plates is considered only a “rough" approximation, as the exact one is defined by a system of intersecting curves, the convergence and the goodness of the classical solution are analyzed, too. Finally, as the problem of the Euler shear buckling stress evaluation is a very important topic for a variety of structures, (e.g. ship ones), two numerical applications are carried out, in order to highlight the role of the uniaxial stresses on the plating scantling procedures and the goodness of the proposed formulas.

Evaluation Framework for Agent-Oriented Methodologies

Many agent-oriented software engineering methodologies have been proposed for software developing; however their application is still limited due to their lack of maturity. Evaluating the strengths and weaknesses of these methodologies plays an important role in improving them and in developing new stronger methodologies. This paper presents an evaluation framework for agent-oriented methodologies, which addresses six major areas: concepts, notation, process, pragmatics, support for software engineering and marketability. The framework is then used to evaluate the Gaia methodology to identify its strengths and weaknesses, and to prove the ability of the framework for promoting the agent-oriented methodologies by detecting their weaknesses in detail.

Comparative Study of Complexity in Streetscape Composition

This research is a comparative study of complexity, as a multidimensional concept, in the context of streetscape composition in Algeria and Japan. 80 streetscapes visual arrays have been collected and then presented to 20 participants, with different cultural backgrounds, in order to be categorized and classified according to their degrees of complexity. Three analysis methods have been used in this research: cluster analysis, ranking method and Hayashi Quantification method (Method III). The results showed that complexity, disorder, irregularity and disorganization are often conflicting concepts in the urban context. Algerian daytime streetscapes seem to be balanced, ordered and regular, and Japanese daytime streetscapes seem to be unbalanced, regular and vivid. Variety, richness and irregularity with some aspects of order and organization seem to characterize Algerian night streetscapes. Japanese night streetscapes seem to be more related to balance, regularity, order and organization with some aspects of confusion and ambiguity. Complexity characterized mainly Algerian avenues with green infrastructure. Therefore, for Japanese participants, Japanese traditional night streetscapes were complex. And for foreigners, Algerian and Japanese avenues nightscapes were the most complex visual arrays.

Use of Novel Algorithms MAJE4 and MACJER-320 for Achieving Confidentiality and Message Authentication in SSL and TLS

Extensive use of the Internet coupled with the marvelous growth in e-commerce and m-commerce has created a huge demand for information security. The Secure Socket Layer (SSL) protocol is the most widely used security protocol in the Internet which meets this demand. It provides protection against eaves droppings, tampering and forgery. The cryptographic algorithms RC4 and HMAC have been in use for achieving security services like confidentiality and authentication in the SSL. But recent attacks against RC4 and HMAC have raised questions in the confidence on these algorithms. Hence two novel cryptographic algorithms MAJE4 and MACJER-320 have been proposed as substitutes for them. The focus of this work is to demonstrate the performance of these new algorithms and suggest them as dependable alternatives to satisfy the need of security services in SSL. The performance evaluation has been done by using practical implementation method.

DEA Method for Evaluation of EU Performance

The paper deals with an application of quantitative analysis – the Data Envelopment Analysis (DEA) method to performance evaluation of the European Union Member States, in the reference years 2000 and 2011. The main aim of the paper is to measure efficiency changes over the reference years and to analyze a level of productivity in individual countries based on DEA method and to classify the EU Member States to homogeneous units (clusters) according to efficiency results. The theoretical part is devoted to the fundamental basis of performance theory and the methodology of DEA. The empirical part is aimed at measuring degree of productivity and level of efficiency changes of evaluated countries by basic DEA model – CCR CRS model, and specialized DEA approach – the Malmquist Index measuring the change of technical efficiency and the movement of production possibility frontier. Here, DEA method becomes a suitable tool for setting a competitive/uncompetitive position of each country because there is not only one factor evaluated, but a set of different factors that determine the degree of economic development.