Abstract: Implementing quality assurance in higher education establishments is the main focus of the reform process currently undertaken by the Ministry of Higher Education and Scientific Research in the Kurdistan Region of Iraq. The reform agenda has involved attempts to improve academic quality and management processes in universities, technical institutions and colleges. The central challenge for the reform process is to produce change in higher education in a region where administration is described as centralized and bureaucratic. To make these changes, there should be a well-designed plans and follow up processes in order to monitor progress and develop responses to obstacles. Lack of skills, resources, political dilemmas, poor motivation, and readiness to face the consequences of change are factors which will determine the success of the reform process.
Abstract: The k-nearest neighbors (knn) is a simple but effective method of classification. In this paper we present an extended version of this technique for chemical compounds used in High Throughput Screening, where the distances of the nearest neighbors can be taken into account. Our algorithm uses kernel weight functions as guidance for the process of defining activity in screening data. Proposed kernel weight function aims to combine properties of graphical structure and molecule descriptors of screening compounds. We apply the modified knn method on several experimental data from biological screens. The experimental results confirm the effectiveness of the proposed method.
Abstract: For complete support of Quality of Service, it is better that environment itself predicts resource requirements of a job by using special methods in the Grid computing. The exact and correct prediction causes exact matching of required resources with available resources. After the execution of each job, the used resources will be saved in the active database named "History". At first some of the attributes will be exploit from the main job and according to a defined similarity algorithm the most similar executed job will be exploited from "History" using statistic terms such as linear regression or average, resource requirements will be predicted. The new idea in this research is based on active database and centralized history maintenance. Implementation and testing of the proposed architecture results in accuracy percentage of 96.68% to predict CPU usage of jobs and 91.29% of memory usage and 89.80% of the band width usage.
Abstract: Competitive relationships among Bradyrhizobium
japonicum USDA serogroup 123, 122 and 138 were screened versus
the standard commercial soybean variety Williams and two
introductions P1 377578 "671" in a field trial. Displacement of strain
123 by an effective strain should improved N2 fixation. Root nodules
were collected and strain occupancy percentage was determined
using strain specific fluorescent antibodies technique. As anticipated
the strain USDA 123 dominated 92% of nodules due to the high
affinity between the host and the symbiont. This dominance was
consistent and not changed materially either by inoculation practice
or by introducing new strainan. The interrelationship between the
genotype Williams and serogroup 122 & 138 was found very weak
although the cell density of the strain in the rhizosphere area was
equal. On the other hand, the nodule occupancy of genotypes 671 and
166 with rhizobia serogroup 123 was almost diminished to zero. .
The data further exhibited that the genotypes P1 671 and P1 166 have
high affinity to colonize with strains 122 and 138 whereas Williams
was highly promiscuous to strain 123.
Abstract: In this paper is shown that the probability-statistic methods application, especially at the early stage of the aviation gas turbine engine (GTE) technical condition diagnosing, when the flight information has property of the fuzzy, limitation and uncertainty is unfounded. Hence is considered the efficiency of application of new technology Soft Computing at these diagnosing stages with the using of the Fuzzy Logic and Neural Networks methods. Training with high accuracy of fuzzy multiple linear and non-linear models (fuzzy regression equations) which received on the statistical fuzzy data basis is made. Thus for GTE technical condition more adequate model making are analysed dynamics of skewness and kurtosis coefficients' changes. Researches of skewness and kurtosis coefficients values- changes show that, distributions of GTE work parameters have fuzzy character. Hence consideration of fuzzy skewness and kurtosis coefficients is expedient. Investigation of the basic characteristics changes- dynamics of GTE work parameters allows to draw conclusion on necessity of the Fuzzy Statistical Analysis at preliminary identification of the engines' technical condition. Researches of correlation coefficients values- changes shows also on their fuzzy character. Therefore for models choice the application of the Fuzzy Correlation Analysis results is offered. For checking of models adequacy is considered the Fuzzy Multiple Correlation Coefficient of Fuzzy Multiple Regression. At the information sufficiency is offered to use recurrent algorithm of aviation GTE technical condition identification (Hard Computing technology is used) on measurements of input and output parameters of the multiple linear and non-linear generalised models at presence of noise measured (the new recursive Least Squares Method (LSM)). The developed GTE condition monitoring system provides stage-bystage estimation of engine technical conditions. As application of the given technique the estimation of the new operating aviation engine temperature condition was made.
Abstract: Web applications have become complex and crucial for many firms, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering). The scientific community has focused attention to Web application design, development, analysis, testing, by studying and proposing methodologies and tools. Static and dynamic techniques may be used to analyze existing Web applications. The use of traditional static source code analysis may be very difficult, for the presence of dynamically generated code, and for the multi-language nature of the Web. Dynamic analysis may be useful, but it has an intrinsic limitation, the low number of program executions used to extract information. Our reverse engineering analysis, used into our WAAT (Web Applications Analysis and Testing) project, applies mutational techniques in order to exploit server side execution engines to accomplish part of the dynamic analysis. This paper studies the effects of mutation source code analysis applied to Web software to build application models. Mutation-based generated models may contain more information then necessary, so we need a pruning mechanism.
Abstract: This paper presents a critical study about the
application of Neural Networks to ion-exchange process. Ionexchange
is a complex non-linear process involving many factors
influencing the ions uptake mechanisms from the pregnant solution.
The following step includes the elution. Published data presents
empirical isotherm equations with definite shortcomings resulting in
unreliable predictions. Although Neural Network simulation
technique encounters a number of disadvantages including its “black
box", and a limited ability to explicitly identify possible causal
relationships, it has the advantage to implicitly handle complex
nonlinear relationships between dependent and independent
variables. In the present paper, the Neural Network model based on
the back-propagation algorithm Levenberg-Marquardt was developed
using a three layer approach with a tangent sigmoid transfer function
(tansig) at hidden layer with 11 neurons and linear transfer function
(purelin) at out layer. The above mentioned approach has been used
to test the effectiveness in simulating ion exchange processes. The
modeling results showed that there is an excellent agreement between
the experimental data and the predicted values of copper ions
removed from aqueous solutions.
Abstract: With deep development of software reuse, componentrelated
technologies have been widely applied in the development of
large-scale complex applications. Component identification (CI) is
one of the primary research problems in software reuse, by analyzing
domain business models to get a set of business components with high
reuse value and good reuse performance to support effective reuse.
Based on the concept and classification of CI, its technical stack is
briefly discussed from four views, i.e., form of input business models,
identification goals, identification strategies, and identification
process. Then various CI methods presented in literatures are
classified into four types, i.e., domain analysis based methods,
cohesion-coupling based clustering methods, CRUD matrix based
methods, and other methods, with the comparisons between these
methods for their advantages and disadvantages. Additionally, some
insufficiencies of study on CI are discussed, and the causes are
explained subsequently. Finally, it is concluded with some
significantly promising tendency about research on this problem.
Abstract: This article is an extension and a practical application
approach of Wheeler-s NEBIC theory (Net Enabled Business
Innovation Cycle). NEBIC theory is a new approach in IS research
and can be used for dynamic environment related to new technology.
Firms can follow the market changes rapidly with support of the IT
resources. Flexible firms adapt their market strategies, and respond
more quickly to customers changing behaviors. When every leading
firm in an industry has access to the same IT resources, the way that
these IT resources are managed will determine the competitive
advantages or disadvantages of firm. From Dynamic Capabilities
Perspective and from newly introduced NEBIC theory by Wheeler,
we know that only IT resources cannot deliver customer value but
good configuration of those resources can guarantee customer value
by choosing the right emerging technology, grasping the economic
opportunities through business innovation and growth. We found
evidences in literature that SOA (Service Oriented Architecture) is a
promising emerging technology which can deliver the desired
economic opportunity through modularity, flexibility and loosecoupling.
SOA can also help firms to connect in network which can
open a new window of opportunity to collaborate in innovation and
right kind of outsourcing
Abstract: Product Data Management (PDM) systems for Computer
Aided Design (CAD) file management are widely established
in design processes. This management system is indispensable for
design collaboration or when design task distribution is present. It is
thus surprising that engineering design curricula has not paid much
attention in the education of PDM systems. This is also the case
for eduction of ecodesign and environmental evaluation of products.
With the rise of sustainability as a strategic aspect in companies,
environmental concerns are becoming a key issue in design. This
paper discusses the establishment of a PDM platform to be used
among technical and vocational schools in Austria. The PDM system
facilitates design collaboration among these schools. Further, it will
be discussed how the PDM system has been prepared in order to
facilitate environmental evaluation of parts, components and subassemblies
of a product. By integrating a Business Intelligence
solution, environmental Life Cycle Assessment and communication
of results is enabled.
Abstract: We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.
Abstract: Response to the public health-related emergencies is analysed here for a rural university in South Africa. The structure of the designated emergency plan covers all the phases of the disaster management cycle. The plan contains elements of the vulnerability model and the technocratic model of emergency management. The response structures are vertically and horizontally integrated, while the planning contains elements of scenario-based and functional planning. The available number of medical professionals at the Rhodes University, along with the medical insurance rates, makes the staff and students potentially more medically vulnerable than the South African population. The main improvements of the emergency management are required in the tornado response and the information dissemination during health emergencies. The latter should involve the increased use of social media and e-mails, following the Taylor model of communication. Infrastructure must be improved in the telecommunication sector in the face of unpredictable electricity outages.
Abstract: Nondestructive testing in engineering is an inverse
Cauchy problem for Laplace equation. In this paper the problem
of nondestructive testing is expressed by a Laplace-s equation with
third-kind boundary conditions. In order to find unknown values on
the boundary, the method of fundamental solution is introduced and
realized. Because of the ill-posedness of studied problems, the TSVD
regularization technique in combination with L-curve criteria and
Generalized Cross Validation criteria is employed. Numerical results
are shown that the TSVD method combined with L-curve criteria is
more efficient than the TSVD method combined with GCV criteria.
The abstract goes here.
Abstract: This paper will focus on modeling, analysis and simulation of a 42V/14V dc/dc converter based architecture. This architecture is considered to be technically a viable solution for automotive dual-voltage power system for passenger car in the near further. An interleaved dc/dc converter system is chosen for the automotive converter topology due to its advantages regarding filter reduction, dynamic response, and power management. Presented herein, is a model based on one kilowatt interleaved six-phase buck converter designed to operate in a Discontinuous Conduction Mode (DCM). The control strategy of the converter is based on a voltagemode- controlled Pulse Width Modulation (PWM) with a Proportional-Integral-Derivative (PID). The effectiveness of the interleaved step-down converter is verified through simulation results using control-oriented simulator, MatLab/Simulink.
Abstract: In this work a new platform for mobile-health systems is
presented. System target application is providing decision support to
rescue corps or military medical personnel in combat areas. Software
architecture relies on a distributed client-server system that manages a
wireless ad-hoc networks hierarchy in which several different types of
client operate. Each client is characterized for different hardware and
software requirements. Lower hierarchy levels rely in a network of
completely custom devices that store clinical information and patient
status and are designed to form an ad-hoc network operating in the
2.4 GHz ISM band and complying with the IEEE 802.15.4 standard
(ZigBee). Medical personnel may interact with such devices, that are
called MICs (Medical Information Carriers), by means of a PDA
(Personal Digital Assistant) or a MDA (Medical Digital Assistant),
and transmit the information stored in their local databases as well as
issue a service request to the upper hierarchy levels by using IEEE
802.11 a/b/g standard (WiFi). The server acts as a repository that
stores both medical evacuation forms and associated events (e.g., a
teleconsulting request). All the actors participating in the diagnostic
or evacuation process may access asynchronously to such repository
and update its content or generate new events. The designed system
pretends to optimise and improve information spreading and flow
among all the system components with the aim of improving both
diagnostic quality and evacuation process.
Abstract: This paper examines whether or not immigration has a positive influence on the duration of unemployment, in a macroeconomic perspective. We analyse also whether the degree of labor market integration can influence migration. The integration of immigrants into the labor market is a recurrence theme in the work on the economic consequences of immigration. However, to our knowledge, no researchers have studied the impact of immigration on unemployment duration, and vice versa. With two methodology of research (panel estimations (OLS and 2SLS) and panel cointegration techniques), we show that migration seems to influence positively the short-term unemployment and negatively long-term unemployment, for 14 OECD destination countries. In addition, immigration seems to be conditioned by the structural and institutional characteristics of the labour market.
Abstract: Classical Bose-Chaudhuri-Hocquenghem (BCH) codes C that contain their dual codes can be used to construct quantum stabilizer codes this chapter studies the properties of such codes. It had been shown that a BCH code of length n which contains its dual code satisfies the bound on weight of any non-zero codeword in C and converse is also true. One impressive difficulty in quantum communication and computation is to protect informationcarrying quantum states against undesired interactions with the environment. To address this difficulty, many good quantum errorcorrecting codes have been derived as binary stabilizer codes. We were able to shed more light on the structure of dual containing BCH codes. These results make it possible to determine the parameters of quantum BCH codes in terms of weight of non-zero dual codeword.
Abstract: The posteroanterior manipulation technique is usually include in the procedure of the lumbar spine to evaluate the intervertebral motion according to mechanical resistance. The mechanical device with visual feedback was proposed that allows one to analysis the lumbar segments mobility “in vivo" facilitating for the therapist to take its treatment evolution. The measuring system uses load cell and displacement sensor to estimate spine stiffness. In this work, the device was tested by 2 therapists, female, applying posteroanterior force techniques to 5 volunteers, female, with frequency of approximately 1.2-1.8 Hz. A test-retest procedure was used for 2 periods of day. The visual feedback results small variation of forces and cycle time during 6 cycles rhythmic application. The stiffness values showed good agreement between test-retest procedures when used same order of maximum forces.
Abstract: The Sensor Network consists of densely deployed
sensor nodes. Energy optimization is one of the most important
aspects of sensor application design. Data acquisition and aggregation
techniques for processing data in-network should be energy efficient.
Due to the cross-layer design, resource-limited and noisy nature
of Wireless Sensor Networks(WSNs), it is challenging to study
the performance of these systems in a realistic setting. In this
paper, we propose optimizing queries by aggregation of data and
data redundancy to reduce energy consumption without requiring
all sensed data and directed diffusion communication paradigm to
achieve power savings, robust communication and processing data
in-network. To estimate the per-node power consumption POWERTossim
mica2 energy model is used, which provides scalable and
accurate results. The performance analysis shows that the proposed
methods overcomes the existing methods in the aspects of energy
consumption in wireless sensor networks.
Abstract: Resistive switching of aluminum nitride (AlNx) thin film was demonstrated in a TaN/AlNx/TiN memory cell that was prepared by sputter deposition techniques. The memory cell showed bipolar switching of resistance between +3.5 V and –3.5 V. The resistance ratio of high resistance state (HRS) to low resistance state (HRS), RHRS/RLRS, was about 2 over 100 cycles of endurance test. Both the LRS and HRS of the memory cell exhibited ohmic conduction at low voltages and Poole-Frenkel emission at high voltages. The electrical conduction in the TaN/AlNx/TiN memory cell was possibly attributed to the interactions between charges and defects in the AlNx film.