Analysis of MAC Protocols with Correlation Receiver for OCDMA Networks - Part II

In this paper optical code-division multiple-access (OCDMA) packet network is considered, which offers inherent security in the access networks. Two types of random access protocols are proposed for packet transmission. In protocol 1, all distinct codes and in protocol 2, distinct codes as well as shifted versions of all these codes are used. O-CDMA network performance using optical orthogonal codes (OOCs) 1-D and two-dimensional (2-D) wavelength/time single-pulse-per-row (W/T SPR) codes are analyzed. The main advantage of using 2-D codes instead of onedimensional (1-D) codes is to reduce the errors due to multiple access interference among different users. In this paper, correlation receiver is considered in the analysis. Using analytical model, we compute and compare packet-success probability for 1-D and 2-D codes in an O-CDMA network and the analysis shows improved performance with 2-D codes as compared to 1-D codes.

On Quantum BCH Codes and Its Duals

Classical Bose-Chaudhuri-Hocquenghem (BCH) codes C that contain their dual codes can be used to construct quantum stabilizer codes this chapter studies the properties of such codes. It had been shown that a BCH code of length n which contains its dual code satisfies the bound on weight of any non-zero codeword in C and converse is also true. One impressive difficulty in quantum communication and computation is to protect informationcarrying quantum states against undesired interactions with the environment. To address this difficulty, many good quantum errorcorrecting codes have been derived as binary stabilizer codes. We were able to shed more light on the structure of dual containing BCH codes. These results make it possible to determine the parameters of quantum BCH codes in terms of weight of non-zero dual codeword.

Numerical Simulation of the Flow Field around a 30° Inclined Flat Plate

This paper presents a CFD analysis of the flow around a 30° inclined flat plate of infinite span. Numerical predictions have been compared to experimental measurements, in order to assess the potential of the finite volume code of determining the aerodynamic forces acting on a flat plate invested by a fluid stream of infinite extent. Several turbulence models and spatial node distributions have been tested and flow field characteristics in the neighborhood of the flat plate have been numerically investigated, allowing the development of a preliminary procedure to be used as guidance in selecting the appropriate grid configuration and the corresponding turbulence model for the prediction of the flow field over a twodimensional inclined plate.

An Exact MCNP Modeling of Pebble Bed Reactors

Double heterogeneity of randomly located pebbles in the core and Coated Fuel Particles (CFPs) in the pebbles are specific features in pebble bed reactors and usually, because of difficulty to model with MCNP code capabilities, are neglected. In this study, characteristics of HTR-10, Tsinghua University research reactor, are used and not only double heterogeneous but also truncated CFPs and Pebbles are considered.Firstly, 8335 CFPs are distributed randomly in a pebble and then the core of reactor is filled with those pebbles and graphite pebbles as moderator such that 57:43 ratio of fuel and moderator pebbles is established.Finally, four different core configurations are modeled. They are Simple Cubic (SC) structure with truncated pebbles,SC structure without truncated pebble, and Simple Hexagonal(SH) structure without truncated pebbles and SH structure with truncated pebbles. Results like effective multiplication factor (Keff), critical height,etc. are compared with available data.

Enhanced Performance of Fading Dispersive Channel Using Dynamic Frequency Hopping(DFH)

techniques are examined to overcome the performance degradation caused by the channel dispersion using slow frequency hopping (SFH) with dynamic frequency hopping (DFH) pattern adaptation. In DFH systems, the frequency slots are selected by continuous quality monitoring of all frequencies available in a system and modification of hopping patterns for each individual link based on replacing slots which its signal to interference ratio (SIR) measurement is below a required threshold. Simulation results will show the improvements in BER obtained by DFH in comparison with matched frequency hopping (MFH), random frequency hopping (RFH) and multi-carrier code division multiple access (MC-CDMA) in multipath slowly fading dispersive channels using a generalized bandpass two-path transfer function model, and will show the improvement obtained according to the threshold selection.

Natural Flickering of Methane Diffusion Flames

Present study focuses on studying the oscillatory behavior of jet diffusion flames. At a particular jet exit velocity, the flames are seen to exhibit natural flickering. Initially the flickering process is not continuous. In this transition region as well as in the continuous flickering regime, the flickering displays multiple frequency oscillations. The response of the flame to the exit velocity profile of the burner is also studied using three types of burners. The entire range of natural flickering is investigated by capturing high speed digital images and processing them using a MATLAB code.

Tele-Operated Anthropomorphic Arm and Hand Design

In this project, a tele-operated anthropomorphic robotic arm and hand is designed and built as a versatile robotic arm system. The robot has the ability to manipulate objects such as pick and place operations. It is also able to function by itself, in standalone mode. Firstly, the robotic arm is built in order to interface with a personal computer via a serial servo controller circuit board. The circuit board enables user to completely control the robotic arm and moreover, enables feedbacks from user. The control circuit board uses a powerful integrated microcontroller, a PIC (Programmable Interface Controller). The PIC is firstly programmed using BASIC (Beginner-s All-purpose Symbolic Instruction Code) and it is used as the 'brain' of the robot. In addition a user friendly Graphical User Interface (GUI) is developed as the serial servo interface software using Microsoft-s Visual Basic 6. The second part of the project is to use speech recognition control on the robotic arm. A speech recognition circuit board is constructed with onboard components such as PIC and other integrated circuits. It replaces the computers- Graphical User Interface. The robotic arm is able to receive instructions as spoken commands through a microphone and perform operations with respect to the commands such as picking and placing operations.

Evaluation of Ultrasonic C-Scan Images by Fractal Dimension

In this paper, quantitative evaluation of ultrasonic Cscan images through estimation of their Fractal Dimension (FD) is discussed. Necessary algorithm for evaluation of FD of any 2-D digitized image is implemented by developing a computer code. For the evaluation purpose several C-scan images of the Kevlar composite impacted by high speed bullet and glass fibre composite having flaw in the form of inclusion is used. This analysis automatically differentiates a C-scan image showing distinct damage zone, from an image that contains no such damage.

A Semi-Fragile Watermarking Scheme for Color Image Authentication

In this paper, a semi-fragile watermarking scheme is proposed for color image authentication. In this particular scheme, the color image is first transformed from RGB to YST color space, suitable for watermarking the color media. Each channel is divided into 4×4 non-overlapping blocks and its each 2×2 sub-block is selected. The embedding space is created by setting the two LSBs of selected sub-block to zero, which will hold the authentication and recovery information. For verification of work authentication and parity bits denoted by 'a' & 'p' are computed for each 2×2 subblock. For recovery, intensity mean of each 2×2 sub-block is computed and encoded upto six to eight bits depending upon the channel selection. The size of sub-block is important for correct localization and fast computation. For watermark distribution 2DTorus Automorphism is implemented using a private key to have a secure mapping of blocks. The perceptibility of watermarked image is quite reasonable both subjectively and objectively. Our scheme is oblivious, correctly localizes the tampering and able to recovery the original work with probability of near one.

Probabilistic Center Voting Method for Subsequent Object Tracking and Segmentation

In this paper, we introduce a novel algorithm for object tracking in video sequence. In order to represent the object to be tracked, we propose a spatial color histogram model which encodes both the color distribution and spatial information. The object tracking from frame to frame is accomplished via center voting and back projection method. The center voting method has every pixel in the new frame to cast a vote on whereabouts the object center is. The back projection method segments the object from the background. The segmented foreground provides information on object size and orientation, omitting the need to estimate them separately. We do not put any assumption on camera motion; the proposed algorithm works equally well for object tracking in both static and moving camera videos.

A Comparison of Adaline and MLP Neural Network based Predictors in SIR Estimation in Mobile DS/CDMA Systems

In this paper we compare the response of linear and nonlinear neural network-based prediction schemes in prediction of received Signal-to-Interference Power Ratio (SIR) in Direct Sequence Code Division Multiple Access (DS/CDMA) systems. The nonlinear predictor is Multilayer Perceptron MLP and the linear predictor is an Adaptive Linear (Adaline) predictor. We solve the problem of complexity by using the Minimum Mean Squared Error (MMSE) principle to select the optimal predictors. The optimized Adaline predictor is compared to optimized MLP by employing noisy Rayleigh fading signals with 1.8 GHZ carrier frequency in an urban environment. The results show that the Adaline predictor can estimates SIR with the same error as MLP when the user has the velocity of 5 km/h and 60 km/h but by increasing the velocity up-to 120 km/h the mean squared error of MLP is two times more than Adaline predictor. This makes the Adaline predictor (with lower complexity) more suitable than MLP for closed-loop power control where efficient and accurate identification of the time-varying inverse dynamics of the multi path fading channel is required.

Validation Testing for Temporal Neural Networks for RBF Recognition

A neuron can emit spikes in an irregular time basis and by averaging over a certain time window one would ignore a lot of information. It is known that in the context of fast information processing there is no sufficient time to sample an average firing rate of the spiking neurons. The present work shows that the spiking neurons are capable of computing the radial basis functions by storing the relevant information in the neurons' delays. One of the fundamental findings of the this research also is that when using overlapping receptive fields to encode the data patterns it increases the network-s clustering capacity. The clustering algorithm that is discussed here is interesting from computer science and neuroscience point of view as well as from a perspective.

An Adaptive ARQ – HARQ Method with Two RS Codes

In this paper we proposed multistage adaptive ARQ/HARQ/HARQ scheme. This method combines pure ARQ (Automatic Repeat reQuest) mode in low channel bit error rate and hybrid ARQ method using two different Reed-Solomon codes in middle and high error rate conditions. It follows, that our scheme has three stages. The main goal is to increase number of states in adaptive HARQ methods and be able to achieve maximum throughput for every channel bit error rate. We will prove the proposal by calculation and then with simulations in land mobile satellite channel environment. Optimization of scheme system parameters is described in order to maximize the throughput in the whole defined Signal-to- Noise Ratio (SNR) range in selected channel environment.

Development of Thermal Model by Performance Verification of Heat Pipe Subsystem for Electronic Cooling under Space Environment

Heat pipes are used to control the thermal problem for electronic cooling. It is especially difficult to dissipate heat to a heat sink in an environment in space compared to earth. For solving this problem, in this study, the Poiseuille (Po) number, which is the main measure of the performance of a heat pipe, is studied by CFD; then, the heat pipe performance is verified with experimental results. A heat pipe is then fabricated for a spatial environment, and an in-house code is developed. Further, a heat pipe subsystem, which consists of a heat pipe, MLI (Multi Layer Insulator), SSM (Second Surface Mirror), and radiator, is tested and correlated with the TMM (Thermal Mathematical Model) through a commercial code. The correlation results satisfy the 3K requirement, and the generated thermal model is verified for application to a spatial environment.

High-Resolution 12-Bit Segmented Capacitor DAC in Successive Approximation ADC

This paper study the segmented split capacitor Digital-to-Analog Converter (DAC) implemented in a differentialtype 12-bit Successive Approximation Analog-to-Digital Converter (SA-ADC). The series capacitance split array method employed as it reduced the total area of the capacitors required for high resolution DACs. A 12-bit regular binary array structure requires 2049 unit capacitors (Cs) while the split array needs 127 unit Cs. These results in the reduction of the total capacitance and power consumption of the series split array architectures as to regular binary-weighted structures. The paper will show the 12-bit DAC series split capacitor with 4-bit thermometer coded DAC architectures as well as the simulation and measured results.

Statistics of Exon Lengths in Animals, Plants, Fungi, and Protists

Eukaryotic protein-coding genes are interrupted by spliceosomal introns, which are removed from the RNA transcripts before translation into a protein. The exon-intron structures of different eukaryotic species are quite different from each other, and the evolution of such structures raises many questions. We try to address some of these questions using statistical analysis of whole genomes. We go through all the protein-coding genes in a genome and study correlations between the net length of all the exons in a gene, the number of the exons, and the average length of an exon. We also take average values of these features for each chromosome and study correlations between those averages on the chromosomal level. Our data show universal features of exon-intron structures common to animals, plants, and protists (specifically, Arabidopsis thaliana, Caenorhabditis elegans, Drosophila melanogaster, Cryptococcus neoformans, Homo sapiens, Mus musculus, Oryza sativa, and Plasmodium falciparum). We have verified linear correlation between the number of exons in a gene and the length of a protein coded by the gene, while the protein length increases in proportion to the number of exons. On the other hand, the average length of an exon always decreases with the number of exons. Finally, chromosome clustering based on average chromosome properties and parameters of linear regression between the number of exons in a gene and the net length of those exons demonstrates that these average chromosome properties are genome-specific features.

64 bit Computer Architectures for Space Applications – A study

The more recent satellite projects/programs makes extensive usage of real – time embedded systems. 16 bit processors which meet the Mil-Std-1750 standard architecture have been used in on-board systems. Most of the Space Applications have been written in ADA. From a futuristic point of view, 32 bit/ 64 bit processors are needed in the area of spacecraft computing and therefore an effort is desirable in the study and survey of 64 bit architectures for space applications. This will also result in significant technology development in terms of VLSI and software tools for ADA (as the legacy code is in ADA). There are several basic requirements for a special processor for this purpose. They include Radiation Hardened (RadHard) devices, very low power dissipation, compatibility with existing operational systems, scalable architectures for higher computational needs, reliability, higher memory and I/O bandwidth, predictability, realtime operating system and manufacturability of such processors. Further on, these may include selection of FPGA devices, selection of EDA tool chains, design flow, partitioning of the design, pin count, performance evaluation, timing analysis etc. This project deals with a brief study of 32 and 64 bit processors readily available in the market and designing/ fabricating a 64 bit RISC processor named RISC MicroProcessor with added functionalities of an extended double precision floating point unit and a 32 bit signal processing unit acting as co-processors. In this paper, we emphasize the ease and importance of using Open Core (OpenSparc T1 Verilog RTL) and Open “Source" EDA tools such as Icarus to develop FPGA based prototypes quickly. Commercial tools such as Xilinx ISE for Synthesis are also used when appropriate.

EGCL: An Extended G-Code Language with Flow Control, Functions and Mnemonic Variables

In the context of computer numerical control (CNC) and computer aided manufacturing (CAM), the capabilities of programming languages such as symbolic and intuitive programming, program portability and geometrical portfolio have special importance. They allow to save time and to avoid errors during part programming and permit code re-usage. Our updated literature review indicates that the current state of art presents voids in parametric programming, program portability and programming flexibility. In response to this situation, this article presents a compiler implementation for EGCL (Extended G-code Language), a new, enriched CNC programming language which allows the use of descriptive variable names, geometrical functions and flow-control statements (if-then-else, while). Our compiler produces low-level generic, elementary ISO-compliant Gcode, thus allowing for flexibility in the choice of the executing CNC machine and in portability. Our results show that readable variable names and flow control statements allow a simplified and intuitive part programming and permit re-usage of the programs. Future work includes allowing the programmer to define own functions in terms of EGCL, in contrast to the current status of having them as library built-in functions.

Primary School Principles- Views about In-service Training Activities

This study was carried out to determine the feedback from the primary school headmasters on in-service training activities. 384 primary school headmasters were asked to give feedback on the in-service training facilities organized by the Ministry of Education through an open –ended question. The study group was formed by 359 primary school headmasters who attended “Management Skills" seminars which were organized in March, April and May in the extent of 2006 In-service Training Facilities Plan by the Ministry of Education and were also volunteers to give feedback on the inservice training activities. The qualitative research method was used in the analysis of the data since the primary school headmasters gave written feedback on the in-service training activities. Having analyzed the feedback, certain categories were formed by coding it. Certain frequencies and percentages were determined according to the codes. It was made obvious that the primary school headmasters benefitted from those facilities by sharing their experiences, problems and their possible solutions mutually

Codebook Generation for Vector Quantization on Orthogonal Polynomials based Transform Coding

In this paper, a new algorithm for generating codebook is proposed for vector quantization (VQ) in image coding. The significant features of the training image vectors are extracted by using the proposed Orthogonal Polynomials based transformation. We propose to generate the codebook by partitioning these feature vectors into a binary tree. Each feature vector at a non-terminal node of the binary tree is directed to one of the two descendants by comparing a single feature associated with that node to a threshold. The binary tree codebook is used for encoding and decoding the feature vectors. In the decoding process the feature vectors are subjected to inverse transformation with the help of basis functions of the proposed Orthogonal Polynomials based transformation to get back the approximated input image training vectors. The results of the proposed coding are compared with the VQ using Discrete Cosine Transform (DCT) and Pairwise Nearest Neighbor (PNN) algorithm. The new algorithm results in a considerable reduction in computation time and provides better reconstructed picture quality.