Abstract: The Yazd-Ardakan basin in Central Iran has two separated aquifers. The shallow unconfined aquifer is supplies 40 Qanats. The deep saturated confined aquifer is the main water storage. Due to over-withdrawal, water table has been decreasing during last 25 years. Recent study shows that the shortage of the aquifer is about 16 meters and land subsidence is 0.5 - 1.2 meters. Long deep cracks are found just above the aquifer and devour the irrigation water and floods. Although the most cracks direction is NW-SE and could be compared to the main direction of YA basin, there is no direct evidence for relation between land subsidence and the huge cracks. Large-scale water pumping has been decreased the water pressure in aquifer. The pressure decline disturbed the balance and increased the pressure of overlying sediments. So porosity decreased and compaction started. Then, sediments compaction developed and made land subsidence and some huge cracks slowly.
Abstract: Large-scale systems such as Grids offer
infrastructures for both data distribution and parallel processing. The
use of Grid infrastructures is a more recent issue that is already
impacting the Distributed Database Management System industry. In
DBMS, distributed query processing has emerged as a fundamental
technique for ensuring high performance in distributed databases.
Database placement is particularly important in large-scale systems
because it reduces communication costs and improves resource
usage. In this paper, we propose a dynamic database placement
policy that depends on query patterns and Grid sites capabilities. We
evaluate the performance of the proposed database placement policy
using simulations. The obtained results show that dynamic database
placement can significantly improve the performance of distributed
query processing.
Abstract: With deep development of software reuse, componentrelated
technologies have been widely applied in the development of
large-scale complex applications. Component identification (CI) is
one of the primary research problems in software reuse, by analyzing
domain business models to get a set of business components with high
reuse value and good reuse performance to support effective reuse.
Based on the concept and classification of CI, its technical stack is
briefly discussed from four views, i.e., form of input business models,
identification goals, identification strategies, and identification
process. Then various CI methods presented in literatures are
classified into four types, i.e., domain analysis based methods,
cohesion-coupling based clustering methods, CRUD matrix based
methods, and other methods, with the comparisons between these
methods for their advantages and disadvantages. Additionally, some
insufficiencies of study on CI are discussed, and the causes are
explained subsequently. Finally, it is concluded with some
significantly promising tendency about research on this problem.
Abstract: Recently, genetic algorithms (GA) and particle swarm optimization (PSO) technique have attracted considerable attention among various modern heuristic optimization techniques. The GA has been popular in academia and the industry mainly because of its intuitiveness, ease of implementation, and the ability to effectively solve highly non-linear, mixed integer optimization problems that are typical of complex engineering systems. PSO technique is a relatively recent heuristic search method whose mechanics are inspired by the swarming or collaborative behavior of biological populations. In this paper both PSO and GA optimization are employed for finding stable reduced order models of single-input- single-output large-scale linear systems. Both the techniques guarantee stability of reduced order model if the original high order model is stable. PSO method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. Both the methods are illustrated through numerical example from literature and the results are compared with recently published conventional model reduction technique.
Abstract: Shadow detection is still considered as one of the
potential challenges for intelligent automated video surveillance
systems. A pre requisite for reliable and accurate detection and
tracking is the correct shadow detection and classification. In such a
landscape of conditions, privacy issues add more and more
complexity and require reliable shadow detection.
In this work the intertwining between security, accuracy,
reliability and privacy is analyzed and, accordingly, a novel
architecture for Privacy Enhancing Video Surveillance (PEVS) is
introduced. Shadow detection and masking are dealt with through the
combination of two different approaches simultaneously. This results
in a unique privacy enhancement, without affecting security.
Subsequently, the methodology was employed successfully in a
large-scale wireless video surveillance system; privacy relevant
information was stored and encrypted on the unit, without
transferring it over an un-trusted network.
Abstract: There has been gradual progress of late in construction projects, particularly in big-scale megaprojects. Due to the long-term construction period, however, with large-scale budget investment, lack of construction management technologies, and increase in the incomplete elements of project schedule management, a plan to conduct efficient operations and to ensure business safety is required. In particular, as the project management information system (PMIS) is meant for managing a single project centering on the construction phase, there is a limitation in the management of program-scale businesses like megaprojects. Thus, a program management information system (PgMIS) that includes program-level management technologies is needed to manage multiple projects. In this study, a support tool was developed for managing the cost and schedule information occurring in the construction phase, at the program level. In addition, a case study on the developed support tool was conducted to verify the usability of the system. With the use of the developed support tool program, construction managers can monitor the progress of the entire project and of the individual subprojects in real time.
Abstract: Many new experimental films which were free from conventional movie forms have appeared since Nubellbak Movement in the late 1950s. Forty years after the movement started, on March 13th, 1995, on the 100th anniversary of the birth of film, the declaration called Dogme 95, was issued in Copenhagen, Denmark. It aimed to create a new style of avant-garde film, and showed a tendency toward being anti-Hollywood and anti-genre, which were against the highly popular Hollywood trend of movies based on large-scale investment. The main idea of Dogme 95 is the opposition to 'the writer's doctrine' that a film should be the artist's individual work and to 'the overuse of technology' in film. The key figures declared ten principles called 'Vow of Chastity', by which new movie forms were to be produced. Interview (2000), directed by Byunhyuk, was made in 2000, five years after Dogme 95 was declared. This movie was dedicated as the first Asian Dogme. This study will survey the relationship between Korean film and the Vow of Chastity through the Korean films released in theaters from a viewpoint of technology and content. It also will call attention to its effects on and significance to Korean film in modern society.
Abstract: The batch nature limits the standard kernel principal component analysis (KPCA) methods in numerous applications, especially for dynamic or large-scale data. In this paper, an efficient adaptive approach is presented for online extraction of the kernel principal components (KPC). The contribution of this paper may be divided into two parts. First, kernel covariance matrix is correctly updated to adapt to the changing characteristics of data. Second, KPC are recursively formulated to overcome the batch nature of standard KPCA.This formulation is derived from the recursive eigen-decomposition of kernel covariance matrix and indicates the KPC variation caused by the new data. The proposed method not only alleviates sub-optimality of the KPCA method for non-stationary data, but also maintains constant update speed and memory usage as the data-size increases. Experiments for simulation data and real applications demonstrate that our approach yields improvements in terms of both computational speed and approximation accuracy.
Abstract: Computing and maintaining network structures for efficient
data aggregation incurs high overhead for dynamic events
where the set of nodes sensing an event changes with time. Moreover,
structured approaches are sensitive to the waiting time that is used
by nodes to wait for packets from their children before forwarding
the packet to the sink. An optimal routing and data aggregation
scheme for wireless sensor networks is proposed in this paper. We
propose Tree on DAG (ToD), a semistructured approach that uses
Dynamic Forwarding on an implicitly constructed structure composed
of multiple shortest path trees to support network scalability. The key
principle behind ToD is that adjacent nodes in a graph will have
low stretch in one of these trees in ToD, thus resulting in early
aggregation of packets. Based on simulations on a 2,000-node Mica2-
based network, we conclude that efficient aggregation in large-scale
networks can be achieved by our semistructured approach.
Abstract: With the drastically growth in optical communication
technology, a lossless, low-crosstalk and multifunction optical switch
is most desirable for large-scale photonic network. To realize such a
switch, we have introduced the new architecture of optical switch
that embedded many functions on single device. The asymmetrical
architecture of OXADM consists of 3 parts; selective port, add/drop
operation, and path routing. Selective port permits only the interest
wavelength pass through and acts as a filter. While add and drop
function can be implemented in second part of OXADM architecture.
The signals can then be re-routed to any output port or/and perform
an accumulation function which multiplex all signals onto single path
and then exit to any interest output port. This will be done by path
routing operation. The unique features offered by OXADM has
extended its application to Fiber to-the Home Technology (FTTH),
here the OXADM is used as a wavelength management element in
Optical Line Terminal (OLT). Each port is assigned specifically with
the operating wavelengths and with the dynamic routing management
to ensure no traffic combustion occurs in OLT.
Abstract: This paper presents a reliability-based approach to select appropriate wind turbine types for a wind farm considering site-specific wind speed patterns. An actual wind farm in the northern region of Iran with the wind speed registration of one year is studied in this paper. An analytic approach based on total probability theorem is utilized in this paper to model the probabilistic behavior of both turbines- availability and wind speed. Well-known probabilistic reliability indices such as loss of load expectation (LOLE), expected energy not supplied (EENS) and incremental peak load carrying capability (IPLCC) for wind power integration in the Roy Billinton Test System (RBTS) are examined. The most appropriate turbine type achieving the highest reliability level is chosen for the studied wind farm.
Abstract: Investment in a constructed facility represents a cost in
the short term that returns benefits only over the long term use of the
facility. Thus, the costs occur earlier than the benefits, and the owners
of facilities must obtain the capital resources to finance the costs of
construction. A project cannot proceed without an adequate
financing, and the cost of providing an adequate financing can be
quite large. For these reasons, the attention to the project finance is an
important aspect of project management. Finance is also a concern to
the other organizations involved in a project such as the general
contractor and material suppliers. Unless an owner immediately and
completely covers the costs incurred by each participant, these
organizations face financing problems of their own. At a more
general level, the project finance is the only one aspect of the general
problem of corporate finance. If numerous projects are considered
and financed together, then the net cash flow requirements constitute
the corporate financing problem for capital investment. Whether
project finance is performed at the project or at the corporate level
does not alter the basic financing problem .In this paper, we will first
consider facility financing from the owner's perspective, with due
consideration for its interaction with other organizations involved in a
project. Later, we discuss the problems of construction financing
which are crucial to the profitability and solvency of construction
contractors. The objective of this paper is to present the steps utilized
to determine the best combination of minimum project financing.
The proposed model considers financing; schedule and maximum net
area .The proposed model is called Project Financing and Schedule
Integration using Genetic Algorithms "PFSIGA". This model
intended to determine more steps (maximum net area) for any project
with a subproject. An illustrative example will demonstrate the
feature of this technique. The model verification and testing are put
into consideration.
Abstract: The dynamics of User Datagram Protocol (UDP) traffic
over Ethernet between two computers are analyzed using nonlinear
dynamics which shows that there are two clear regimes in the data
flow: free flow and saturated. The two most important variables
affecting this are the packet size and packet flow rate. However,
this transition is due to a transcritical bifurcation rather than phase
transition in models such as in vehicle traffic or theorized large-scale
computer network congestion. It is hoped this model will help lay
the groundwork for further research on the dynamics of networks,
especially computer networks.
Abstract: By introducing the concept of Oracle we propose an approach for improving the performance of genetic algorithms for large-scale asymmetric Traveling Salesman Problems. The results have shown that the proposed approach allows overcoming some traditional problems for creating efficient genetic algorithms.
Abstract: This paper presents a mathematical model and a
methodology to analyze the losses in transmission expansion
planning (TEP) under uncertainty in demand. The methodology is
based on discrete particle swarm optimization (DPSO). DPSO is a
useful and powerful stochastic evolutionary algorithm to solve the
large-scale, discrete and nonlinear optimization problems like TEP.
The effectiveness of the proposed idea is tested on an actual
transmission network of the Azerbaijan regional electric company,
Iran. The simulation results show that considering the losses even for
transmission expansion planning of a network with low load growth
is caused that operational costs decreases considerably and the
network satisfies the requirement of delivering electric power more
reliable to load centers.
Abstract: In the present study, a heterogeneous and
homogeneous gas flow dispersion model for simulation and
optimisation of a large-scale catalytic slurry reactor for the direct
synthesis of dimethyl ether (DME) from syngas and CO2, using a
churn-turbulent regime was developed. In the heterogeneous gas flow
model the gas phase was distributed into two bubble phases: small
and large, however in the homogeneous one, the gas phase was
distributed into only one large bubble phase. The results indicated
that the heterogeneous gas flow model was in more agreement with
experimental pilot plant data than the homogeneous one.
Abstract: Understanding the cell's large-scale organization is an
interesting task in computational biology. Thus, protein-protein
interactions can reveal important organization and function of the
cell. Here, we investigated the correspondence between protein
interactions and function for the yeast. We obtained the correlations
among the set of proteins. Then these correlations are clustered using
both the hierarchical and biclustering methods. The detailed analyses
of proteins in each cluster were carried out by making use of their
functional annotations. As a result, we found that some functional
classes appear together in almost all biclusters. On the other hand, in
hierarchical clustering, the dominancy of one functional class is
observed. In brief, from interaction data to function, some correlated
results are noticed about the relationship between interaction and
function which might give clues about the organization of the
proteins.
Abstract: Decrease in hardware costs and advances in computer
networking technologies have led to increased interest in the use of
large-scale parallel and distributed computing systems. One of the
biggest issues in such systems is the development of effective
techniques/algorithms for the distribution of the processes/load of a
parallel program on multiple hosts to achieve goal(s) such as
minimizing execution time, minimizing communication delays,
maximizing resource utilization and maximizing throughput.
Substantive research using queuing analysis and assuming job
arrivals following a Poisson pattern, have shown that in a multi-host
system the probability of one of the hosts being idle while other host
has multiple jobs queued up can be very high. Such imbalances in
system load suggest that performance can be improved by either
transferring jobs from the currently heavily loaded hosts to the lightly
loaded ones or distributing load evenly/fairly among the hosts .The
algorithms known as load balancing algorithms, helps to achieve the
above said goal(s). These algorithms come into two basic categories -
static and dynamic. Whereas static load balancing algorithms (SLB)
take decisions regarding assignment of tasks to processors based on
the average estimated values of process execution times and
communication delays at compile time, Dynamic load balancing
algorithms (DLB) are adaptive to changing situations and take
decisions at run time.
The objective of this paper work is to identify qualitative
parameters for the comparison of above said algorithms. In future this
work can be extended to develop an experimental environment to
study these Load balancing algorithms based on comparative
parameters quantitatively.
Abstract: The authors present a mixed method for reducing the order of the large-scale dynamic systems. In this method, the denominator polynomial of the reduced order model is obtained by using the modified pole clustering technique while the coefficients of the numerator are obtained by Pade approximations. This method is conceptually simple and always generates stable reduced models if the original high-order system is stable. The proposed method is illustrated with the help of the numerical examples taken from the literature.
Abstract: In this paper, real-coded genetic algorithm (RCGA) optimization technique has been applied for large-scale linear dynamic multi-input-multi-output (MIMO) system. The method is based on error minimization technique where the integral square error between the transient responses of original and reduced order models has been minimized by RCGA. The reduction procedure is simple computer oriented and the approach is comparable in quality with the other well-known reduction techniques. Also, the proposed method guarantees stability of the reduced model if the original high-order MIMO system is stable. The proposed approach of MIMO system order reduction is illustrated with the help of an example and the results are compared with the recently published other well-known reduction techniques to show its superiority.