Evaluation of Linear and Geometrically Nonlinear Static and Dynamic Analysis of Thin Shells by Flat Shell Finite Elements

The choice of finite element to use in order to predict nonlinear static or dynamic response of complex structures becomes an important factor. Then, the main goal of this research work is to focus a study on the effect of the in-plane rotational degrees of freedom in linear and geometrically non linear static and dynamic analysis of thin shell structures by flat shell finite elements. In this purpose: First, simple triangular and quadrilateral flat shell finite elements are implemented in an incremental formulation based on the updated lagrangian corotational description for geometrically nonlinear analysis. The triangular element is a combination of DKT and CST elements, while the quadrilateral is a combination of DKQ and the bilinear quadrilateral membrane element. In both elements, the sixth degree of freedom is handled via introducing fictitious stiffness. Secondly, in the same code, the sixth degrees of freedom in these elements is handled differently where the in-plane rotational d.o.f is considered as an effective d.o.f in the in-plane filed interpolation. Our goal is to compare resulting shell elements. Third, the analysis is enlarged to dynamic linear analysis by direct integration using Newmark-s implicit method. Finally, the linear dynamic analysis is extended to geometrically nonlinear dynamic analysis where Newmark-s method is used to integrate equations of motion and the Newton-Raphson method is employed for iterating within each time step increment until equilibrium is achieved. The obtained results demonstrate the effectiveness and robustness of the interpolation of the in-plane rotational d.o.f. and present deficiencies of using fictitious stiffness in dynamic linear and nonlinear analysis.

Evaluation of the Displacement-Based and the Force-Based Adaptive Pushover Methods in Seismic Response Estimation of Irregular Buildings Considering Torsional Effects

Recent years, adaptive pushover methods have been developed for seismic analysis of structures. Herein, the accuracy of the displacement-based adaptive pushover (DAP) method, which is introduced by Antoniou and Pinho [2004], is evaluated for Irregular buildings. The results are compared to the force-based procedure. Both concrete and steel frame structures, asymmetric in plan and elevation are analyzed and also torsional effects are taking into the account. These analyses are performed using both near fault and far fault records. In order to verify the results, the Incremental Dynamic Analysis (IDA) is performed.

Scatterer Density in Nonlinear Diffusion for Speckle Reduction in Ultrasound Imaging: The Isotropic Case

This paper proposes a method for speckle reduction in medical ultrasound imaging while preserving the edges with the added advantages of adaptive noise filtering and speed. A nonlinear image diffusion method that incorporates local image parameter, namely, scatterer density in addition to gradient, to weight the nonlinear diffusion process, is proposed. The method was tested for the isotropic case with a contrast detail phantom and varieties of clinical ultrasound images, and then compared to linear and some other diffusion enhancement methods. Different diffusion parameters were tested and tuned to best reduce speckle noise and preserve edges. The method showed superior performance measured both quantitatively and qualitatively when incorporating scatterer density into the diffusivity function. The proposed filter can be used as a preprocessing step for ultrasound image enhancement before applying automatic segmentation, automatic volumetric calculations, or 3D ultrasound volume rendering.

Comparison of Parametric and Nonparametric Techniques for Non-peak Traffic Forecasting

Accurately predicting non-peak traffic is crucial to daily traffic for all forecasting models. In the paper, least squares support vector machines (LS-SVMs) are investigated to solve such a practical problem. It is the first time to apply the approach and analyze the forecast performance in the domain. For comparison purpose, two parametric and two non-parametric techniques are selected because of their effectiveness proved in past research. Having good generalization ability and guaranteeing global minima, LS-SVMs perform better than the others. Providing sufficient improvement in stability and robustness reveals that the approach is practically promising.

On the Need to have an Additional Methodology for the Psychological Product Measurement and Evaluation

Cognitive Science appeared about 40 years ago, subsequent to the challenge of the Artificial Intelligence, as common territory for several scientific disciplines such as: IT, mathematics, psychology, neurology, philosophy, sociology, and linguistics. The new born science was justified by the complexity of the problems related to the human knowledge on one hand, and on the other by the fact that none of the above mentioned sciences could explain alone the mental phenomena. Based on the data supplied by the experimental sciences such as psychology or neurology, models of the human mind operation are built in the cognition science. These models are implemented in computer programs and/or electronic circuits (specific to the artificial intelligence) – cognitive systems – whose competences and performances are compared to the human ones, leading to the psychology and neurology data reinterpretation, respectively to the construction of new models. During these processes if psychology provides the experimental basis, philosophy and mathematics provides the abstraction level utterly necessary for the intermission of the mentioned sciences. The ongoing general problematic of the cognitive approach provides two important types of approach: the computational one, starting from the idea that the mental phenomenon can be reduced to 1 and 0 type calculus operations, and the connection one that considers the thinking products as being a result of the interaction between all the composing (included) systems. In the field of psychology measurements in the computational register use classical inquiries and psychometrical tests, generally based on calculus methods. Deeming things from both sides that are representing the cognitive science, we can notice a gap in psychological product measurement possibilities, regarded from the connectionist perspective, that requires the unitary understanding of the quality – quantity whole. In such approach measurement by calculus proves to be inefficient. Our researches, deployed for longer than 20 years, lead to the conclusion that measuring by forms properly fits to the connectionism laws and principles.

AJcFgraph - AspectJ Control Flow Graph Builder for Aspect-Oriented Software

The ever-growing usage of aspect-oriented development methodology in the field of software engineering requires tool support for both research environments and industry. So far, tool support for many activities in aspect-oriented software development has been proposed, to automate and facilitate their development. For instance, the AJaTS provides a transformation system to support aspect-oriented development and refactoring. In particular, it is well established that the abstract interpretation of programs, in any paradigm, pursued in static analysis is best served by a high-level programs representation, such as Control Flow Graph (CFG). This is why such analysis can more easily locate common programmatic idioms for which helpful transformation are already known as well as, association between the input program and intermediate representation can be more closely maintained. However, although the current researches define the good concepts and foundations, to some extent, for control flow analysis of aspectoriented programs but they do not provide a concrete tool that can solely construct the CFG of these programs. Furthermore, most of these works focus on addressing the other issues regarding Aspect- Oriented Software Development (AOSD) such as testing or data flow analysis rather than CFG itself. Therefore, this study is dedicated to build an aspect-oriented control flow graph construction tool called AJcFgraph Builder. The given tool can be applied in many software engineering tasks in the context of AOSD such as, software testing, software metrics, and so forth.

Numerical Investigation of the Thermal Separation in a Vortex Tube

This work has been carried out in order to provide an understanding of the physical behaviors of the flow variation of pressure and temperature in a vortex tube. A computational fluid dynamics model is used to predict the flow fields and the associated temperature separation within a Ranque–Hilsch vortex tube. The CFD model is a steady axisymmetric model (with swirl) that utilizes the standard k-ε turbulence model. The second–order numerical schemes, was used to carry out all the computations. Vortex tube with a circumferential inlet stream and an axial (cold) outlet stream and a circumferential (hot) outlet stream was considered. Performance curves (temperature separation versus cold outlet mass fraction) were obtained for a specific vortex tube with a given inlet mass flow rate. Simulations have been carried out for varying amounts of cold outlet mass flow rates. The model results have a good agreement with experimental data.

Geometric Modeling of Illumination on the TFT-LCD Panel using Bezier Surface

In this paper, we propose a geometric modeling of illumination on the patterned image containing etching transistor. This image is captured by a commercial camera during the inspection of a TFT-LCD panel. Inspection of defect is an important process in the production of LCD panel, but the regional difference in brightness, which has a negative effect on the inspection, is due to the uneven illumination environment. In order to solve this problem, we present a geometric modeling of illumination consisting of an interpolation using the least squares method and 3D modeling using bezier surface. Our computational time, by using the sampling method, is shorter than the previous methods. Moreover, it can be further used to correct brightness in every patterned image.

Clustering Mixed Data Using Non-normal Regression Tree for Process Monitoring

In the semiconductor manufacturing process, large amounts of data are collected from various sensors of multiple facilities. The collected data from sensors have several different characteristics due to variables such as types of products, former processes and recipes. In general, Statistical Quality Control (SQC) methods assume the normality of the data to detect out-of-control states of processes. Although the collected data have different characteristics, using the data as inputs of SQC will increase variations of data, require wide control limits, and decrease performance to detect outof- control. Therefore, it is necessary to separate similar data groups from mixed data for more accurate process control. In the paper, we propose a regression tree using split algorithm based on Pearson distribution to handle non-normal distribution in parametric method. The regression tree finds similar properties of data from different variables. The experiments using real semiconductor manufacturing process data show improved performance in fault detecting ability.

An Evaluation of Algorithms for Single-Echo Biosonar Target Classification

A recent neurospiking coding scheme for feature extraction from biosonar echoes of various plants is examined with avariety of stochastic classifiers. Feature vectors derived are employedin well-known stochastic classifiers, including nearest-neighborhood,single Gaussian and a Gaussian mixture with EM optimization.Classifiers' performances are evaluated by using cross-validation and bootstrapping techniques. It is shown that the various classifers perform equivalently and that the modified preprocessing configuration yields considerably improved results.

High Quality Speech Coding using Combined Parametric and Perceptual Modules

A novel approach to speech coding using the hybrid architecture is presented. Advantages of parametric and perceptual coding methods are utilized together in order to create a speech coding algorithm assuring better signal quality than in traditional CELP parametric codec. Two approaches are discussed. One is based on selection of voiced signal components that are encoded using parametric algorithm, unvoiced components that are encoded perceptually and transients that remain unencoded. The second approach uses perceptual encoding of the residual signal in CELP codec. The algorithm applied for precise transient selection is described. Signal quality achieved using the proposed hybrid codec is compared to quality of some standard speech codecs.

Mathematical Model of the Respiratory System – Comparison of the Total Lung Impedance in the Adult and Neonatal Lung

A mathematical model of the respiratory system is introduced in this study. Geometrical dimensions of the respiratory system were used to compute the acoustic properties of the respiratory system using the electro-acoustic analogy. The effect of the geometrical proportions of the respiratory system is observed in the paper.

Extracting Road Signs using the Color Information

In this paper, we propose a method to extract the road signs. Firstly, the grabbed image is converted into the HSV color space to detect the road signs. Secondly, the morphological operations are used to reduce noise. Finally, extract the road sign using the geometric property. The feature extraction of road sign is done by using the color information. The proposed method has been tested for the real situations. From the experimental results, it is seen that the proposed method can extract the road sign features effectively.

Feasibility Investigation of Near Infrared Spectrometry for Particle Size Estimation of Nano Structures

Determination of nano particle size is substantial since the nano particle size exerts a significant effect on various properties of nano materials. Accordingly, proposing non-destructive, accurate and rapid techniques for this aim is of high interest. There are some conventional techniques to investigate the morphology and grain size of nano particles such as scanning electron microscopy (SEM), atomic force microscopy (AFM) and X-ray diffractometry (XRD). Vibrational spectroscopy is utilized to characterize different compounds and applied for evaluation of the average particle size based on relationship between particle size and near infrared spectra [1,4] , but it has never been applied in quantitative morphological analysis of nano materials. So far, the potential application of nearinfrared (NIR) spectroscopy with its ability in rapid analysis of powdered materials with minimal sample preparation, has been suggested for particle size determination of powdered pharmaceuticals. The relationship between particle size and diffuse reflectance (DR) spectra in near infrared region has been applied to introduce a method for estimation of particle size. Back propagation artificial neural network (BP-ANN) as a nonlinear model was applied to estimate average particle size based on near infrared diffuse reflectance spectra. Thirty five different nano TiO2 samples with different particle size were analyzed by DR-FTNIR spectrometry and the obtained data were processed by BP- ANN.

Memory Estimation of Internet Server Using Queuing Theory: Comparative Study between M/G/1, G/M/1 and G/G/1 Queuing Model

How to effectively allocate system resource to process the Client request by Gateway servers is a challenging problem. In this paper, we propose an improved scheme for autonomous performance of Gateway servers under highly dynamic traffic loads. We devise a methodology to calculate Queue Length and Waiting Time utilizing Gateway Server information to reduce response time variance in presence of bursty traffic. The most widespread contemplation is performance, because Gateway Servers must offer cost-effective and high-availability services in the elongated period, thus they have to be scaled to meet the expected load. Performance measurements can be the base for performance modeling and prediction. With the help of performance models, the performance metrics (like buffer estimation, waiting time) can be determined at the development process. This paper describes the possible queue models those can be applied in the estimation of queue length to estimate the final value of the memory size. Both simulation and experimental studies using synthesized workloads and analysis of real-world Gateway Servers demonstrate the effectiveness of the proposed system.

Kinetic Spectrophotometric Determination of Ramipril in Commercial Dosage Forms

This paper presents a simple and sensitive kinetic spectrophotometric method for the determination of ramipril in commercial dosage forms. The method is based on the reaction of the drug with 1-chloro-2,4-dinitrobenzene (CDNB) in dimethylsulfoxide (DMSO) at 100 ± 1ºC. The reaction is followed spectrophotometrically by measuring the rate of change of the absorbance at 420 nm. Fixed-time (ΔA) and equilibrium methods are adopted for constructing the calibration curves. Both the calibration curves were found to be linear over the concentration ranges 20 - 220 μg/ml. The regression analysis of calibration data yielded the linear equations: Δ A = 6.30 × 10-4 + 1.54 × 10-3 C and A = 3.62 × 10-4 + 6.35 × 10-3 C for fixed time (Δ A) and equilibrium methods, respectively. The limits of detection (LOD) for fixed time and equilibrium methods are 1.47 and 1.05 μg/ml, respectively. The method has been successfully applied to the determination of ramipril in commercial dosage forms. Statistical comparison of the results shows that there is no significant difference between the proposed methods and Abdellatef-s spectrophotometric method.

The Comparative Investigation and Calculation of Thermo-Neutronic Parameters on Two Gens II and III Nuclear Reactors with Same Powers

Whereas in the third generation nuclear reactors, dimensions of core and also the kind of coolant and enrichment percent of fuel have significantly changed than the second generation, therefore in this article the aim is based on a comparative investigation between two same power reactors of second and third generations, that the neutronic parameters of both reactors such as: K∞, Keff and its details and thermal hydraulic parameters such as: power density, specific power, volumetric heat rate, released power per fuel volume unit, volume and mass of clad and fuel (consisting fissile and fertile fuels), be calculated and compared together. By this comparing the efficiency and modification of third generation nuclear reactors than second generation which have same power can be distinguished. In order to calculate the cited parameters, some information such as: core dimensions, the pitch of lattice, the fuel matter, the percent of enrichment and the kind of coolant are used. For calculating the neutronic parameters, a neutronic program entitled: SIXFAC and also related formulas have been used. Meantime for calculating the thermal hydraulic and other parameters, analytical method and related formulas have been applied.

An Artificial Intelligent Technique for Robust Digital Watermarking in Multiwavelet Domain

In this paper, an artificial intelligent technique for robust digital image watermarking in multiwavelet domain is proposed. The embedding technique is based on the quantization index modulation technique and the watermark extraction process does not require the original image. We have developed an optimization technique using the genetic algorithms to search for optimal quantization steps to improve the quality of watermarked image and robustness of the watermark. In addition, we construct a prediction model based on image moments and back propagation neural network to correct an attacked image geometrically before the watermark extraction process begins. The experimental results show that the proposed watermarking algorithm yields watermarked image with good imperceptibility and very robust watermark against various image processing attacks.

Deformation of Water Waves by Geometric Transitions with Power Law Function Distribution

In this work, we analyze the deformation of surface waves in shallow flows conditions, propagating in a channel of slowly varying cross-section. Based on a singular perturbation technique, the main purpose is to predict the motion of waves by using a dimensionless formulation of the governing equations, considering that the longitudinal variation of the transversal section obey a power-law distribution. We show that the spatial distribution of the waves in the varying cross-section is a function of a kinematic parameter,κ , and two geometrical parameters εh and w ε . The above spatial behavior of the surface elevation is modeled by an ordinary differential equation. The use of single formulas to model the varying cross sections or transitions considered in this work can be a useful approximation to natural or artificial geometrical configurations.

Stress Intensity Factors for Plates with Collinear and Non-Aligned Straight Cracks

Multi-site damage (MSD) has been a challenge to aircraft, civil and power plant structures. In real life components are subjected to cracking at many vulnerable locations such as the bolt holes. However, we do not consider for the presence of multiple cracks. Unlike components with a single crack, these components are difficult to predict. When two cracks approach one another, their stress fields influence each other and produce enhancing or shielding effect depending on the position of the cracks. In the present study, numerical studies on fracture analysis have been conducted by using the developed code based on the modified virtual crack closure integral (MVCCI) technique and finite element analysis (FEA) software ABAQUS for computing SIF of plates with multiple cracks. Various parametric studies have been carried out and the results have been compared with literature where ever available and also with the solution, obtained by using ABAQUS. By conducting extensive numerical studies expressions for SIF have been obtained for collinear cracks and non-aligned cracks.