Abstract: In this paper, the classical bearing capacity problem is re-considered from discrete element analysis. In the discrete element approach, the bearing capacity problem is considered from the elastic stage to plastic stage to rupture stage (large displacement). The bearing capacity failure mechanism of a strip footing on soil is investigated, and the influence of micro-parameters on the bearing capacity of soil is also observed. It is found that the distinct element method (DEM) gives very good visualized results, and basically coincides well with that derived by the classical methods.
Abstract: In this paper, numerical approximate Laplace transform inversion algorithm based on Chebyshev polynomial of second kind is developed using odd cosine series. The technique has been tested for three different functions to work efficiently. The illustrations show that the new developed numerical inverse Laplace transform is very much close to the classical analytic inverse Laplace transform.
Abstract: In the current paper, a domain independent conduction model compatible for multi-physical system dynamic investigations is suggested. By means of a port-based approach, a classical nonlinear conduction model containing physical states is first represented. A compatible discrete configuration of the thermal domain in line with the elastic domain is then generated through the enhancement of the configuration of the conventional thermal element. The presented simulation results of a sample structure indicate that the suggested conductive model can cover a wide range of dynamic behavior of the thermal domain.
Abstract: Philip Morris International (PMI) is developing a range
of novel tobacco products with the potential to reduce individual
risk and population harm in comparison to smoking cigarettes.
One of these products is the Tobacco Heating System 2.2 (THS
2.2), (named as the Electrically Heated Tobacco System (EHTS) in
this paper), already commercialized in a number of countries (e.g.,
Japan, Italy, Switzerland, Russia, Portugal and Romania). During use,
the patented EHTS heats a specifically designed tobacco product
(Electrically Heated Tobacco Product (EHTP)) when inserted into
a Holder (heating device). The EHTP contains tobacco material in
the form of a porous plug that undergoes a controlled heating process
to release chemical compounds into vapors, from which an aerosol
is formed during cooling. The aim of this work was to investigate
the aerosol formation characteristics for realistic operating conditions
of the EHTS as well as for relevant gas mixture compositions
measured in the EHTP aerosol consisting mostly of water, glycerol
and nicotine, but also other compounds at much lower concentrations.
The nucleation process taking place in the EHTP during use when
operated in the Holder has therefore been modeled numerically using
an extended Classical Nucleation Theory (CNT) for multicomponent
gas mixtures. Results from the performed simulations demonstrate
that aerosol droplets are formed only in the presence of an aerosol
former being mainly glycerol. Minor compounds in the gas mixture
were not able to reach a supersaturated state alone and therefore
could not generate aerosol droplets from the multicomponent gas
mixture at the operating conditions simulated. For the analytically
characterized aerosol composition and estimated operating conditions
of the EHTS and EHTP, glycerol was shown to be the main aerosol
former triggering the nucleation process in the EHTP. This implies
that according to the CNT, an aerosol former, such as glycerol
needs to be present in the gas mixture for an aerosol to form
under the tested operating conditions. To assess if these conclusions
are sensitive to the initial amount of the minor compounds and to
include and represent the total mass of the aerosol collected during
the analytical aerosol characterization, simulations were carried out
with initial masses of the minor compounds increased by as much
as a factor of 500. Despite this extreme condition, no aerosol
droplets were generated when glycerol, nicotine and water were
treated as inert species and therefore not actively contributing to the
nucleation process. This implies that according to the CNT, an aerosol
cannot be generated without the help of an aerosol former, from
the multicomponent gas mixtures at the compositions and operating
conditions estimated for the EHTP, even if all minor compounds are
released or generated in a single puff.
Abstract: This paper presents a subband adaptive filter (SAF)
for a system identification where an impulse response is sparse
and disturbed with an impulsive noise. Benefiting from the uses
of l1-norm optimization and l0-norm penalty of the weight vector
in the cost function, the proposed l0-norm sign SAF (l0-SSAF)
achieves both robustness against impulsive noise and much improved
convergence behavior than the classical adaptive filters. Simulation
results in the system identification scenario confirm that the proposed
l0-norm SSAF is not only more robust but also faster and more
accurate than its counterparts in the sparse system identification in
the presence of impulsive noise.
Abstract: This paper presents a normalized subband adaptive
filtering (NSAF) algorithm to cope with the sparsity condition of
an underlying system in the context of compressive sensing. By
regularizing a weighted l1-norm of the filter taps estimate onto the
cost function of the NSAF and utilizing a subgradient analysis,
the update recursion of the l1-norm constraint NSAF is derived.
Considering two distinct weighted l1-norm regularization cases, two
versions of the l1-norm constraint NSAF are presented. Simulation
results clearly indicate the superior performance of the proposed
l1-norm constraint NSAFs comparing with the classical NSAF.
Abstract: The gas holdup fluctuations in a bubble column (0.15 m in ID) have been recorded by means of a conductivity wire-mesh sensor in order to extract information about the main transition velocities. These parameters are very important for bubble column design, operation and scale-up. For this purpose, the classical definition of the Shannon entropy was modified and used to identify both the onset (at UG=0.034 m/s) of the transition flow regime and the beginning (at UG=0.089 m/s) of the churn-turbulent flow regime. The results were compared with the Kolmogorov entropy (KE) results. A slight discrepancy was found, namely the transition velocities identified by means of the KE were shifted to somewhat higher (0.045 and 0.101 m/s) superficial gas velocities UG.
Abstract: The present study focuses on the environmental performance of the companies in the electricity-producing sector and its relationship with their financial performance. We will review the major studies that examined the relationship between the environmental and financial performance of firms in various industries. While the classical economic debates consider the environmental friendly activities costly and harmful to a firm’s profitability, it is claimed that firms will be rewarded with higher profitability in long run through the investments in environmental friendly activities. In this context, prior studies have examined the relationship between the environmental and financial performance of firms operating in different industry sectors. Our study will employ an environmental indicator to increase the accuracy of the results and be employed as an independent variable in our developed econometric model to evaluate the impact of the financial performance of the firms on their environmental friendly activities in the context of companies operating in the Australian electricity-producing sector. As a result, we expect our methodology to contribute to the literature and the findings of the study will help us to provide recommendations and policy implications to the electricity producers.
Abstract: This paper reviews a number of theoretical aspects
for implementing an explicit spatial perspective in econometrics
for modelling non-continuous data, in general, and count data, in
particular. It provides an overview of the several spatial econometric
approaches that are available to model data that are collected with
reference to location in space, from the classical spatial econometrics
approaches to the recent developments on spatial econometrics to
model count data, in a Bayesian hierarchical setting. Considerable
attention is paid to the inferential framework, necessary for
structural consistent spatial econometric count models, incorporating
spatial lag autocorrelation, to the corresponding estimation and
testing procedures for different assumptions, to the constrains and
implications embedded in the various specifications in the literature. This review combines insights from the classical spatial
econometrics literature as well as from hierarchical modeling and
analysis of spatial data, in order to look for new possible directions
on the processing of count data, in a spatial hierarchical Bayesian
econometric context.
Abstract: We present a new subband adaptive filter (R-SAF)
which is robust against impulsive noise in system identification. To
address the vulnerability of adaptive filters based on the L2-norm
optimization criterion against impulsive noise, the R-SAF comes from
the L1-norm optimization criterion with a constraint on the energy
of the weight update. Minimizing L1-norm of the a posteriori error
in each subband with a constraint on minimum disturbance gives
rise to the robustness against the impulsive noise and the capable
convergence performance. Experimental results clearly demonstrate
that the proposed R-SAF outperforms the classical adaptive filtering
algorithms when impulsive noise as well as background noise exist.
Abstract: Green chemistry for plant extraction of active principles is the main interest of many researchers concerned with climate change. While classical organic solvents are detrimental to our environment, greener alternatives to ionic liquids are very promising for sustainable organic chemistry. This study focused on the determination of functional groups observed in the main constituents from the ionic liquid extracts of Coleus aromaticus Benth leaves using FT-IR Spectroscopy. Moreover, this research aimed to determine the best ionic liquid that can separate functionalized plant constituents from the leaves Coleus aromaticus Benth using Fourier Transform Infrared Spectroscopy. Coleus aromaticus Benth leaf extract in different ionic liquids, elucidated pharmacologically important functional groups present in major constituents of the plant, namely, rosmarinic acid, caffeic acid and chlorogenic acid. In connection to distinctive appearance of functional groups in the spectrum and highest % transmittance, potassium chloride-glycerol is the best ionic liquid for green extraction.
Abstract: The very well-known stacked sets of numbers referred
to as Pascal’s triangle present the coefficients of the binomial
expansion of the form (x+y)n. This paper presents an approach (the
Staircase Horizontal Vertical, SHV-method) to the generalization of
planar Pascal’s triangle for polynomial expansion of the form
(x+y+z+w+r+⋯)n. The presented generalization of Pascal’s triangle
is different from other generalizations of Pascal’s triangles given in
the literature. The coefficients of the generalized Pascal’s triangles,
presented in this work, are generated by inspection, using embedded
Pascal’s triangles. The coefficients of I-variables expansion are
generated by horizontally laying out the Pascal’s elements of (I-1)
variables expansion, in a staircase manner, and multiplying them with
the relevant columns of vertically laid out classical Pascal’s elements,
hence avoiding factorial calculations for generating the coefficients
of the polynomial expansion. Furthermore, the classical Pascal’s
triangle has some pattern built into it regarding its odd and even
numbers. Such pattern is known as the Sierpinski’s triangle. In this
study, a presentation of Sierpinski-like patterns of the generalized
Pascal’s triangles is given. Applications related to those coefficients
of the binomial expansion (Pascal’s triangle), or polynomial
expansion (generalized Pascal’s triangles) can be in areas of
combinatorics, and probabilities.
Abstract: In this paper, according to the classical algorithm
LSQR for solving the least-squares problem, an iterative method is
proposed for least-squares solution of constrained matrix equation. By
using the Kronecker product, the matrix-form LSQR is presented to
obtain the like-minimum norm and minimum norm solutions in a
constrained matrix set for the symmetric arrowhead matrices. Finally,
numerical examples are also given to investigate the performance.
Abstract: The modelling of physical phenomena, such as the
earth’s free oscillations, the vibration of strings, the interaction of
atomic particles, or the steady state flow in a bar give rise to Sturm-
Liouville (SL) eigenvalue problems. The boundary applications of
some systems like the convection-diffusion equation, electromagnetic
and heat transfer problems requires the combination of Dirichlet and
Neumann boundary conditions. Hence, the incorporation of Robin
boundary condition in the analyses of Sturm-Liouville problem. This
paper deals with the computation of the eigenvalues and
eigenfunction of generalized Sturm-Liouville problems with Robin
boundary condition using the finite element method. Numerical
solution of classical Sturm–Liouville problem is presented. The
results show an agreement with the exact solution. High results
precision is achieved with higher number of elements.
Abstract: Graphene, a single-atom sheet, has been considered as
the most promising material for making future nanoelectromechanical
systems as well as purely electrical switching with graphene
transistors. Graphene-based devices have advantages in scaled-up
device fabrication due to the recent progress in large area graphene
growth and lithographic patterning of graphene nanostructures. Here
we investigated its mechanical responses of circular graphene
nanoflake under the nanoindentation using classical molecular
dynamics simulations. A correlation between the load and the
indentation depth was constructed. The nanoindented force in this
work was applied to the center point of the circular graphene nanoflake
and then, the resonance frequency could be tuned by a nanoindented
depth. We found the hardening or the softening of the graphene
nanoflake during its nanoindented-deflections, and such properties
were recognized by the shift of the resonance frequency. The
calculated mechanical parameters in the force-vs-deflection plot were
in good agreement with previous experimental and theoretical works.
This proposed schematics can detect the pressure via the deflection
change or/and the resonance frequency shift, and also have great
potential for versatile applications in nanoelectromechanical systems.
Abstract: The McEliece cryptosystem is an asymmetric type of
cryptography based on error correction code. The classical McEliece
used irreducible binary Goppa code which considered unbreakable
until now especially with parameter [1024, 524, and 101], but it is
suffering from large public key matrix which leads to be difficult to
be used practically. In this work Irreducible and Separable Goppa
codes have been introduced. The Irreducible and Separable Goppa
codes used are with flexible parameters and dynamic error vectors. A
Comparison between Separable and Irreducible Goppa code in
McEliece Cryptosystem has been done. For encryption stage, to get
better result for comparison, two types of testing have been chosen;
in the first one the random message is constant while the parameters
of Goppa code have been changed. But for the second test, the
parameters of Goppa code are constant (m=8 and t=10) while the
random message have been changed. The results show that the time
needed to calculate parity check matrix in separable are higher than
the one for irreducible McEliece cryptosystem, which is considered
expected results due to calculate extra parity check matrix in
decryption process for g2(z) in separable type, and the time needed to
execute error locator in decryption stage in separable type is better
than the time needed to calculate it in irreducible type. The proposed
implementation has been done by Visual studio C#.
Abstract: Tamil handwritten document is taken as a key source
of data to identify the writer. Tamil is a classical language which has
247 characters include compound characters, consonants, vowels and
special character. Most characters of Tamil are multifaceted in
nature. Handwriting is a unique feature of an individual. Writer may
change their handwritings according to their frame of mind and this
place a risky challenge in identifying the writer. A new
discriminative model with pooled features of handwriting is proposed
and implemented using support vector machine. It has been reported
on 100% of prediction accuracy by RBF and polynomial kernel based
classification model.
Abstract: Tamil handwritten document is taken as a key source of data to identify the writer. Tamil is a classical language which has 247 characters include compound characters, consonants, vowels and special character. Most characters of Tamil are multifaceted in nature. Handwriting is a unique feature of an individual. Writer may change their handwritings according to their frame of mind and this place a risky challenge in identifying the writer. A new discriminative model with pooled features of handwriting is proposed and implemented using support vector machine. It has been reported on 100% of prediction accuracy by RBF and polynomial kernel based classification model.
Abstract: Facility location is a complex real-world problem
which needs a strategic management decision. This paper provides a
general review on studies, efforts and developments in Facility
Location Problems which are classical optimization problems having
a wide-spread applications in various areas such as transportation,
distribution, production, supply chain decisions and
telecommunication. Our goal is not to review all variants of different
studies in FLPs or to describe very detailed computational techniques
and solution approaches, but rather to provide a broad overview of
major location problems that have been studied, indicating how they
are formulated and what are proposed by researchers to tackle the
problem. A brief, elucidative table based on a grouping according to
“General Problem Type” and “Methods Proposed” used in the studies
is also presented at the end of the work.
Abstract: This paper treats different aspects of entropy measure
in classical information theory and statistical quantum mechanics, it
presents the possibility of extending the definition of Von Neumann
entropy to image and array processing. In the first part, we generalize
the quantum entropy using singular values of arbitrary rectangular
matrices to measure the randomness and the quality of denoising
operation, this new definition of entropy can be implemented to
compare the performance analysis of filtering methods. In the second
part, we apply the concept of pure state in quantum formalism
to generalize the maximum entropy method for narrowband and
farfield source localization problem. Several computer simulation
results are illustrated to demonstrate the effectiveness of the proposed
techniques.