Abstract: We study the problem of reconstructing a three dimensional binary matrices whose interiors are only accessible through few projections. Such question is prominently motivated by the demand in material science for developing tool for reconstruction of crystalline structures from their images obtained by high-resolution transmission electron microscopy. Various approaches have been suggested to reconstruct 3D-object (crystalline structure) by reconstructing slice of the 3D-object. To handle the ill-posedness of the problem, a priori information such as convexity, connectivity and periodicity are used to limit the number of possible solutions. Formally, 3Dobject (crystalline structure) having a priory information is modeled by a class of 3D-binary matrices satisfying a priori information. We consider 3D-binary matrices with periodicity constraints, and we propose a polynomial time algorithm to reconstruct 3D-binary matrices with periodicity constraints from two orthogonal projections.
Abstract: Currently, one of the main directions is developing of
development based on the clustering of economic operations of
Kazakhstan, providing for the organization and concentration of
production capacity in one region or the most optimal system. In the
modern economic literature clustering is regarded as one of the most
effective tools to ensure competitive businesses, and improve their
business itself.
Abstract: Since supply chains highly impact the financial
performance of companies, it is important to optimize and analyze
their Key Performance Indicators (KPI). The synergistic combination
of Particle Swarm Optimization (PSO) and Monte Carlo simulation is
applied to determine the optimal reorder point of warehouses in
supply chains. The goal of the optimization is the minimization of the
objective function calculated as the linear combination of holding and
order costs. The required values of service levels of the warehouses
represent non-linear constraints in the PSO. The results illustrate that
the developed stochastic simulator and optimization tool is flexible
enough to handle complex situations.
Abstract: Quality costs are the costs associated with preventing,
finding, and correcting defective work. Since the main language of
corporate management is money, quality-related costs act as means of
communication between the staff of quality engineering departments
and the company managers. The objective of quality engineering is to
minimize the total quality cost across the life of product. Quality
costs provide a benchmark against which improvement can be
measured over time. It provides a rupee-based report on quality
improvement efforts. It is an effective tool to identify, prioritize and
select quality improvement projects. After reviewing through the
literature it was noticed that a simplified methodology for data
collection of quality cost in a manufacturing industry was required.
The quantified standard methodology is proposed for collecting data
of various elements of quality cost categories for manufacturing
industry. Also in the light of research carried out so far, it is felt
necessary to standardise cost elements in each of the prevention,
appraisal, internal failure and external failure costs. . Here an attempt
is made to standardise the various cost elements applicable to
manufacturing industry and data is collected by using the proposed
quantified methodology. This paper discusses the case study carried
in luggage manufacturing industry.
Abstract: Nowadays social media are important tools for web
resource discovery. The performance and capabilities of web searches
are vital, especially search results from social research paper
bookmarking. This paper proposes a new algorithm for ranking
method that is a combination of similarity ranking with paper posted
time or CSTRank. The paper posted time is static ranking for
improving search results. For this particular study, the paper posted
time is combined with similarity ranking to produce a better ranking
than other methods such as similarity ranking or SimRank. The
retrieval performance of combination rankings is evaluated using
mean values of NDCG. The evaluation in the experiments implies
that the chosen CSTRank ranking by using weight score at ratio 90:10
can improve the efficiency of research paper searching on social
bookmarking websites.
Abstract: Web applications have become very complex and crucial, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering), the scientific community has focused attention to Web applications design, development, analysis, and testing, by studying and proposing methodologies and tools. This paper proposes an approach to automatic multi-dimensional concern mining for Web Applications, based on concepts analysis, impact analysis, and token-based concern identification. This approach lets the user to analyse and traverse Web software relevant to a particular concern (concept, goal, purpose, etc.) via multi-dimensional separation of concerns, to document, understand and test Web applications. This technique was developed in the context of WAAT (Web Applications Analysis and Testing) project. A semi-automatic tool to support this technique is currently under development.
Abstract: A new analytical method to predict the torsional
capacity and behavior of R.C multi-cell box girders strengthened with
carbon fiber reinforced polymer (CFRP) sheets is presented.
Modification was done on the Softened Truss Model (STM) in the
proposed method; the concrete torsional problem is solved by
combining the equilibrium conditions, compatibility conditions and
constitutive laws of materials by taking into account the confinement
of concrete with CFRP sheets. A specific algorithm is developed to
predict the torsional behavior of reinforced concrete multi-cell box
girders with or without strengthening by CFRP sheets. Applications
of the developed method as an assessment tool to strengthened multicell
box girders with CFRP and first analytical example that
demonstrate the contribution of the CFRP materials on the torsional
response is also included.
Abstract: Despite various methods that exist in software risk management, software projects have a high rate of failure. When complexity and size of the projects are increased, managing software development becomes more difficult. In these projects the need for more analysis and risk assessment is vital. In this paper, a classification for software risks is specified. Then relations between these risks using risk tree structure are presented. Analysis and assessment of these risks are done using probabilistic calculations. This analysis helps qualitative and quantitative assessment of risk of failure. Moreover it can help software risk management process. This classification and risk tree structure can apply to some software tools.
Abstract: Mobile Ad hoc Network (MANET) is a wireless ad hoc self-configuring network of mobile routers (and associated hosts) connected by wireless links, the union of which forms an arbitrary topology, cause of the random mobility of the nodes. In this paper, an attempt has been made to compare these three protocols DSDV, AODV and DSR on the performance basis under different traffic protocols namely CBR and TCP in a large network. The simulation tool is NS2, the scenarios are made to see the effect of pause times. The results presented in this paper clearly indicate that the different protocols behave differently under different pause times. Also, the results show the main characteristics of different traffic protocols operating on MANETs and thus select the best protocol on each scenario.
Abstract: Symbolic Circuit Analysis (SCA) is a technique used
to generate the symbolic expression of a network. It has become a
well-established technique in circuit analysis and design. The
symbolic expression of networks offers excellent way to perform
frequency response analysis, sensitivity computation, stability
measurements, performance optimization, and fault diagnosis. Many
approaches have been proposed in the area of SCA offering different
features and capabilities. Numerical Interpolation methods are very
common in this context, especially by using the Fast Fourier
Transform (FFT). The aim of this paper is to present a method for
SCA that depends on the use of Wavelet Transform (WT) as a
mathematical tool to generate the symbolic expression for large
circuits with minimizing the analysis time by reducing the number of
computations.
Abstract: A method is presented for using thermo-mechanical fatigue analysis as a tool in the design of automotive heat exchangers. Use of infra-red thermography to measure the real thermal history in the heat exchanger reduces the time necessary for calculating design parameters and improves prediction accuracy. Thermal shocks are the primary cause of heat exchanger damage. Thermo-mechanical simulation is based on the mean behavior of the aluminum tubes used in the heat exchanger. An energetic fatigue criterion is used to detect critical zones.
Abstract: In biological and biomedical research motif finding tools are important in locating regulatory elements in DNA sequences. There are many such motif finding tools available, which often yield position weight matrices and significance indicators. These indicators, p-values and E-values, describe the likelihood that a motif alignment is generated by the background process, and the expected number of occurrences of the motif in the data set, respectively. The various tools often estimate these indicators differently, making them not directly comparable. One approach for comparing motifs from different tools, is computing the E-value as the product of the p-value and the number of possible alignments in the data set. In this paper we explore the combinatorics of the motif alignment models OOPS, ZOOPS, and ANR, and propose a generic algorithm for computing the number of possible combinations accurately. We also show that using the wrong alignment model can give E-values that significantly diverge from their true values.
Abstract: The modern queueing theory is one of the powerful
tools for a quantitative and qualitative analysis of communication systems, computer networks, transportation systems, and many other technical systems. The paper is designated to the analysis of queueing
systems, arising in the networks theory and communications theory
(called open queueing network). The authors of this research in the
sphere of queueing theory present the theorem about the law of the iterated logarithm (LIL) for the queue length of a customers in open
queueing network and its application to the mathematical model of
the open message switching system.
Abstract: The main objective developed in this paper is to find a
graphic technique for modeling, simulation and diagnosis of the
industrial systems. This importance is much apparent when it is about
a complex system such as the nuclear reactor with pressurized water
of several form with various several non-linearity and time scales. In
this case the analytical approach is heavy and does not give a fast
idea on the evolution of the system. The tool Bond Graph enabled us
to transform the analytical model into graphic model and the
software of simulation SYMBOLS 2000 specific to the Bond Graphs
made it possible to validate and have the results given by the
technical specifications. We introduce the analysis of the problem
involved in the faults localization and identification in the complex
industrial processes. We propose a method of fault detection applied
to the diagnosis and to determine the gravity of a detected fault. We
show the possibilities of application of the new diagnosis approaches
to the complex system control. The industrial systems became
increasingly complex with the faults diagnosis procedures in the
physical systems prove to become very complex as soon as the
systems considered are not elementary any more. Indeed, in front of
this complexity, we chose to make recourse to Fault Detection and
Isolation method (FDI) by the analysis of the problem of its control
and to conceive a reliable system of diagnosis making it possible to
apprehend the complex dynamic systems spatially distributed applied
to the standard pressurized water nuclear reactor.
Abstract: Annotation of a protein sequence is pivotal for the understanding of its function. Accuracy of manual annotation provided by curators is still questionable by having lesser evidence strength and yet a hard task and time consuming. A number of computational methods including tools have been developed to tackle this challenging task. However, they require high-cost hardware, are difficult to be setup by the bioscientists, or depend on time intensive and blind sequence similarity search like Basic Local Alignment Search Tool. This paper introduces a new method of assigning highly correlated Gene Ontology terms of annotated protein sequences to partially annotated or newly discovered protein sequences. This method is fully based on Gene Ontology data and annotations. Two problems had been identified to achieve this method. The first problem relates to splitting the single monolithic Gene Ontology RDF/XML file into a set of smaller files that can be easy to assess and process. Thus, these files can be enriched with protein sequences and Inferred from Electronic Annotation evidence associations. The second problem involves searching for a set of semantically similar Gene Ontology terms to a given query. The details of macro and micro problems involved and their solutions including objective of this study are described. This paper also describes the protein sequence annotation and the Gene Ontology. The methodology of this study and Gene Ontology based protein sequence annotation tool namely extended UTMGO is presented. Furthermore, its basic version which is a Gene Ontology browser that is based on semantic similarity search is also introduced.
Abstract: Considering complexity of products, new geometrical
design and investment tolerances that are necessary, measuring and
dimensional controlling involve modern and more precise methods.
Photo digitizing method using two cameras to record pictures and
utilization of conventional method named “cloud points" and data
analysis by the use of ATOUS software, is known as modern and
efficient in mentioned context. In this paper, benefits of photo
digitizing method in evaluating sampling of machining processes
have been put forward. For example, assessment of geometrical
integrity surface in 5-axis milling process and measurement of
carbide tool wear in turning process, can be can be brought forward.
Advantages of this method comparing to conventional methods have
been expressed.
Abstract: The objective of this research seeks to transmit a distance training model to the community in the upper northeastern region. The group sampling consists of 60 community leaders in the municipality of sub-district Kumphawapi, Kumphawapi Disrict, Udonthani Province. The research tools rely on the following instruments, they are : 1) the achievement test of community leaders- training and 2) the satisfaction questionnaires of community leaders. The statistics used in data analysis takes the statistical mean, percentage, standard deviation, and statistical T-test. The resulted findings reveal : 1) the efficiency of the distance training developed by the researcher for the community leaders joining in the training received the average score between in-training and post-training period higher than the setup criterion, 2) the two groups of participants in the training achieved higher knowledge than their pre-training state, 3) the comparison of the achievements between the two group presented no different results, 4) the community leaders obtained the high-to-highest satisfaction.
Abstract: 16-Mercaptohexadecanoic acid (MHDA) and tripeptide glutathione conjugated with gold nanoparticles (Au-NPs) are characterized by Fourier Transform InfaRared (FTIR) spectroscopy combined with Surface-enhanced Raman scattering (SERS) spectroscopy. Surface Plasmon Resonance (SPR) technique based on FTIR spectroscopy has become an important tool in biophysics, which is perspective for the study of organic compounds. FTIR-spectra of MHDA shows the line at 2500 cm-1 attributed to thiol group which is modified by presence of Au-NPs, suggesting the formation of bond between thiol group and gold. We also can observe the peaks originate from characteristic chemical group. A Raman spectrum of the same sample is also promising. Our preliminary experiments confirm that SERS-effect takes place for MHDA connected with Au-NPs and enable us to detected small number (less than 106 cm-2) of MHDA molecules. Combination of spectroscopy methods: FTIR and SERS – enable to study optical properties of Au- NPs and immobilized bio-molecules in context of a bio-nano-sensors.
Abstract: This contribution deals with the relationship between
communication effectiveness and the extent of communication
among organizational units. To facilitate communication between
employees and to increase the level of understanding, the knowledge
of communication tools is necessary. Recent experience has shown
that personal communication is critical for smooth running of
companies and cannot be fully replaced by any form of technical
communication devices.
Below are presented the outcomes of the research on the
relationship between the extent of communication among
organisational units and its efficiency.
Abstract: The demand for new telecommunication services requiring higher capacities, data rates and different operating modes have motivated the development of new generation multi-standard wireless transceivers. A multi-standard design often involves extensive system level analysis and architectural partitioning, typically requiring extensive calculations. In this research, a decimation filter design tool for wireless communication standards consisting of GSM, WCDMA, WLANa, WLANb, WLANg and WiMAX is developed in MATLAB® using GUIDE environment for visual analysis. The user can select a required wireless communication standard, and obtain the corresponding multistage decimation filter implementation using this toolbox. The toolbox helps the user or design engineer to perform a quick design and analysis of decimation filter for multiple standards without doing extensive calculation of the underlying methods.