Abstract: Ultrasound (US) is widely used in medical field for
a variety diagnostic techniques but, in recent years, it has also
been creating great interest for therapeutic aims. Regarding drug
delivery, the use of US as an activation source provides better
spatial delivery confinement and limits the undesired side effects.
However, at present there is no complete characterization
at a fundamental level of the different signals produced by
sono-activated nanocarriers. Therefore, the aim of this study
is to obtain a metrological characterization of the cavitation
phenomena induced by US through three parallel investigation
approaches. US was focused into a channel of a customized
phantom in which a solution with oxygen-loaded nanodroplets
(OLNDs) was led to flow and the cavitation activity was
monitored. Both quantitative and qualitative real-time analysis
were performed giving information about the dynamics of
bubble formation, oscillation and final implosion with respect
to the working acoustic pressure and the type of nanodroplets,
compared with pure water. From this analysis a possible
interpretation of the observed results is proposed.
Abstract: Chain Ladder (CL) method, Expected Loss Ratio (ELR) method and Bornhuetter-Ferguson (BF) method, in addition to more complex transition-rate modeling, are commonly used actuarial reserving methods in general insurance. There is limited published research about their relative performance in the context of Mortgage Insurance (MI). In our experience, these traditional techniques pose unique challenges and do not provide stable claim estimates for medium to longer term liabilities. The relative strengths and weaknesses among various alternative approaches revolve around: stability in the recent loss development pattern, sufficiency and reliability of loss development data, and agreement/disagreement between reported losses to date and ultimate loss estimate. CL method results in volatile reserve estimates, especially for accident periods with little development experience. The ELR method breaks down especially when ultimate loss ratios are not stable and predictable. While the BF method provides a good tradeoff between the loss development approach (CL) and ELR, the approach generates claim development and ultimate reserves that are disconnected from the ever-to-date (ETD) development experience for some accident years that have more development experience. Further, BF is based on subjective a priori assumption. The fundamental shortcoming of these methods is their inability to model exogenous factors, like the economy, which impact various cohorts at the same chronological time but at staggered points along their life-time development. This paper proposes an alternative approach of parametrizing the loss development curve and using logistic regression to generate the ultimate loss estimate for each homogeneous group (accident year or delinquency period). The methodology was tested on an actual MI claim development dataset where various cohorts followed a sigmoidal trend, but levels varied substantially depending upon the economic and operational conditions during the development period spanning over many years. The proposed approach provides the ability to indirectly incorporate such exogenous factors and produce more stable loss forecasts for reserving purposes as compared to the traditional CL and BF methods.
Abstract: A vibration isolation technology for precise position
control of a rotary system powered by two permanent magnet DC
(PMDC) motors is proposed, where this system is mounted on an
oscillatory frame. To achieve vibration isolation for this system,
active damping and disturbance rejection (ADDR) technology
is presented which introduces a cooperation of a main and
an auxiliary PMDC, controlled by discrete-time sliding mode
control (DTSMC) based schemes. The controller of the main
actuator tracks a desired position and the auxiliary actuator
simultaneously isolates the induced vibration, as its controller
follows a torque trend. To determine this torque trend, a
combination of two algorithms is introduced by the ADDR
technology. The first torque-trend producing algorithm rejects
the disturbance by counteracting the perturbation, estimated
using a model-based observer. The second torque trend applies
active variable damping to minimize the oscillation of the output
shaft. In this practice, the presented technology is implemented
on a rotary system with a pendulum attached, mounted on a
linear actuator simulating an oscillation-transmitting structure.
In addition, the obtained results illustrate the functionality of the
Abstract: Parkinson's disease (PD) is a heterogeneous disorder with common age of onset, symptoms, and progression levels. In this paper we will solve analytically the PD model as a non-linear delay differential equation using the steps method. The step method transforms a system of delay differential equations (DDEs) into systems of ordinary differential equations (ODEs). On some numerical examples, the analytical solution will be difficult. So we will approximate the analytical solution using Picard method and Taylor method to ODEs.
Abstract: An option is defined as a financial contract that provides the holder the right but not the obligation to buy or sell a specified quantity of an underlying asset in the future at a fixed price (called a strike price) on or before the expiration date of the option. This paper examined two approaches for derivation of Partial Differential Equation (PDE) options price valuation formula for the Heston stochastic volatility model. We obtained various PDE option price valuation formulas using the riskless portfolio method and the application of Feynman-Kac theorem respectively. From the results obtained, we see that the two derived PDEs for Heston model are distinct and non-unique. This establishes the fact of incompleteness in the model for option price valuation.
Abstract: The use of mathematical models for solving biological problems varies from simple to complex analyses, depending on the nature of the research problems and applicability of the models. The method is more common nowadays. Many complex models become impractical when transmitted analytically. However, alternative approach such as numerical method can be employed. It appropriateness in solving linear and non-linear model equation in Differential Transformation Method (DTM) which depends on Taylor series make it applicable. Hence this study investigates the application of DTM to solve dynamic transmission of Lassa fever model in a population. The mathematical model was formulated using first order differential equation. Firstly, existence and uniqueness of the solution was determined to establish that the model is mathematically well posed for the application of DTM. Numerically, simulations were conducted to compare the results obtained by DTM and that of fourth-order Runge-Kutta method. As shown, DTM is very effective in predicting the solution of epidemics of Lassa fever model.
Abstract: Recent development of AI and edge computing plays a critical role to capture meaningful events such as detection of an unattended bag. One of the core problems is re-identification across multiple CCTVs. Immediately following the detection of a meaningful event is to track and trace the objects related to the event. In an extensive environment, the challenge becomes severe when the number of CCTVs increases substantially, imposing difficulties in achieving high accuracy while maintaining real-time performance. The algorithm that re-identifies cross-boundary objects for extensive tracking is referred to Extensive Re-Identification, which emphasizes the issues related to the complexity behind a great number of CCTVs. The Spatial-Temporal Awareness approach challenges the conventional thinking and concept of operations which is labor intensive and time consuming. The ability to perform Extensive Re-Identification through a multi-sensory network provides the next-level insights – creating value beyond traditional risk management.
Abstract: The emergence of digital twin technology, a digital replica of physical world, has improved the real-time access to data from sensors about the performance of buildings. This digital transformation has opened up many opportunities to improve the management of the building by using the data collected to help monitor consumption patterns and energy leakages. One example is the integration of predictive models for anomaly detection. In this paper, we use the GAM (Generalised Additive Model) for the anomaly detection of Air Handling Units (AHU) power consumption pattern. There is ample research work on the use of GAM for the prediction of power consumption at the office building and nation-wide level. However, there is limited illustration of its anomaly detection capabilities, prescriptive analytics case study, and its integration with the latest development of digital twin technology. In this paper, we applied the general GAM modelling framework on the historical data of the AHU power consumption and cooling load of the building between Jan 2018 to Aug 2019 from an education campus in Singapore to train prediction models that, in turn, yield predicted values and ranges. The historical data are seamlessly extracted from the digital twin for modelling purposes. We enhanced the utility of the GAM model by using it to power a real-time anomaly detection system based on the forward predicted ranges. The magnitude of deviation from the upper and lower bounds of the uncertainty intervals is used to inform and identify anomalous data points, all based on historical data, without explicit intervention from domain experts. Notwithstanding, the domain expert fits in through an optional feedback loop through which iterative data cleansing is performed. After an anomalously high or low level of power consumption detected, a set of rule-based conditions are evaluated in real-time to help determine the next course of action for the facilities manager. The performance of GAM is then compared with other approaches to evaluate its effectiveness. Lastly, we discuss the successfully deployment of this approach for the detection of anomalous power consumption pattern and illustrated with real-world use cases.
Abstract: A warrant is a financial contract that confers the right but not the obligation, to buy or sell a security at a certain price before expiration. The standard procedure to value equity warrants using call option pricing models such as the Black–Scholes model had been proven to contain many flaws, such as the assumption of constant interest rate and constant volatility. In fact, existing alternative models were found focusing more on demonstrating techniques for pricing, rather than empirical testing. Therefore, a mathematical model for pricing and analyzing equity warrants which comprises stochastic interest rate and stochastic volatility is essential to incorporate the dynamic relationships between the identified variables and illustrate the real market. Here, the aim is to develop dynamic pricing formulations for hybrid equity warrants by incorporating stochastic interest rates from the Cox-Ingersoll-Ross (CIR) model, along with stochastic volatility from the Heston model. The development of the model involves the derivations of stochastic differential equations that govern the model dynamics. The resulting equations which involve Cauchy problem and heat equations are then solved using partial differential equation approaches. The analytical pricing formulas obtained in this study comply with the form of analytical expressions embedded in the Black-Scholes model and other existing pricing models for equity warrants. This facilitates the practicality of this proposed formula for comparison purposes and further empirical study.
Abstract: The following assumptions of the Big Bang theory are challenged and found to be false: the cosmological principle, the assumption that all matter formed at the same time and the assumption regarding the cause of the cosmic microwave background radiation. The evolution of the universe is described based on the conclusion that the universe is finite with a space boundary. This conclusion is reached by ruling out the possibility of an infinite universe or a universe which is finite with no boundary. In a finite universe, the centre of the universe can be located with reference to our home galaxy (The Milky Way) using the speed relative to the Cosmic Microwave Background (CMB) rest frame and Hubble's law. This places our home galaxy at a distance of approximately 26 million light years from the centre of the universe. Because we are making observations from a point relatively close to the centre of the universe, the universe appears to be isotropic and homogeneous but this is not the case. The CMB is coming from a source located within the event horizon of the universe. There is sufficient mass in the universe to create an event horizon at the Schwarzschild radius. Galaxies form over time due to the energy released by the expansion of space. Conservation of energy must consider total energy which is mass (+ve) plus energy (+ve) plus spacetime curvature (-ve) so that the total energy of the universe is always zero. The predominant position of galaxy formation moves over time from the centre of the universe towards the boundary so that today the majority of new galaxy formation is taking place beyond our horizon of observation at 14 billion light years.
Abstract: The generalization of relativistic theory of gravity based essentially on the principle of equivalence stipulates that for all bodies, the grave mass is equal to the inert mass which leads us to believe that gravitation is not a property of the bodies themselves, but of space, and the conclusion that the gravitational field must curved space-time what allows the abandonment of Minkowski space (because Minkowski space-time being nonetheless null curvature) to adopt Riemannian geometry as a mathematical framework in order to determine the curvature. Therefore the work presented in this paper begins with the evolution of the concept of gravity then tensor field which manifests by Riemannian geometry to formulate the general equation of the gravitational field.
Abstract: The goal of option pricing theory is to help the investors
to manage their money, enhance returns and control their financial
future by theoretically valuing their options. However, most of the
option pricing models have no analytical solution. Furthermore,
not all the numerical methods are efficient to solve these models
because they have nonsmoothing payoffs or discontinuous derivatives
at the exercise price. In this paper, we solve the American option
under jump diffusion models by using efficient time-dependent
numerical methods. several techniques are integrated to reduced
the overcome the computational complexity. Fast Fourier Transform
(FFT) algorithm is used as a matrix-vector multiplication solver,
which reduces the complexity from O(M2) into O(M logM).
Partial fraction decomposition technique is applied to rational
approximation schemes to overcome the complexity of inverting
polynomial of matrices. The proposed method is easy to implement
on serial or parallel versions. Numerical results are presented to prove
the accuracy and efficiency of the proposed method.
Abstract: The goal of option pricing theory is to help the
investors to manage their money, enhance returns and control their
financial future by theoretically valuing their options. Modeling
option pricing by Black-School models with jumps guarantees to
consider the market movement. However, only numerical methods
can solve this model. Furthermore, not all the numerical methods
are efficient to solve these models because they have nonsmoothing
payoffs or discontinuous derivatives at the exercise price. In this
paper, the exponential time differencing (ETD) method is applied
for solving partial integrodifferential equations arising in pricing
European options under Merton’s and Kou’s jump-diffusion models.
Fast Fourier Transform (FFT) algorithm is used as a matrix-vector
multiplication solver, which reduces the complexity from O(M2)
into O(M logM). A partial fraction form of Pad`e schemes is used
to overcome the complexity of inverting polynomial of matrices.
These two tools guarantee to get efficient and accurate numerical
solutions. We construct a parallel and easy to implement a version
of the numerical scheme. Numerical experiments are given to show
how fast and accurate is our scheme.
Abstract: Given H a Hilbert space and B(H) the algebra of
bounded linear operator in H, let δAB denote the generalized
derivation defined by A and B. The main objective of this article
is to study Weyl type theorems for generalized derivation for (A,B)
satisfying a couple of Fuglede.
Abstract: Continuously differentiable radial basis functions (RBFs) are meshfree, converge faster as the dimensionality increases, and is theoretically spectrally convergent. When implemented on current single and double precision computers, such RBFs can suffer from ill-conditioning because the systems of equations needed to be solved to find the expansion coefficients are full. However, the Advanpix extended precision software package allows computer mathematics to resemble asymptotically ideal Platonic mathematics. Additionally, full systems with extended precision execute faster graphical processors units and field-programmable gate arrays because no branching is needed. Sparse equation systems are fast for iterative solvers in a very limited number of cases.
Abstract: Nowadays, most universities use the course enrollment system considering students’ registration orders. However, the students’ preference level to certain courses is also one important factor to consider. In this research, the possibility of applying a preference-first system has been discussed and analyzed compared to the order-first system. A bipartite graph is applied to resemble the relationship between students and courses they tend to register. With the graph set up, we apply Ford-Fulkerson (F.F.) Algorithm to maximize parings between two sets of nodes, in our case, students and courses. Two models are proposed in this paper: the one considered students’ order first, and the one considered students’ preference first. By comparing and contrasting the two models, we highlight the usability of models which potentially leads to better designs for school course registration systems.
Abstract: We present a discrete game theoretical model with
homogeneous individuals who make simultaneous decisions. In
this model the strategy space of all individuals is a discrete
and dichotomous set which consists of two strategies. We fully
characterize the coherent, split and mixed strategies that form Nash
equilibria and we determine the corresponding Nash domains for all
individuals. We find all strategic thresholds in which individuals can
change their mind if small perturbations in the parameters of the
Abstract: This paper introduces the Nested Vortexes conception of the universe structure and interprets all the physical phenomena according this conception. The paper first reviews recent physics theories, either in microscopic scale or macroscopic scale, to collect evidence that the space is not empty. But, these theories describe the property of the space medium without determining its structure. Determining the structure of space medium is essential to understand the mechanism that leads to its properties. Without determining the space medium structure, many phenomena; such as electric and magnetic fields, gravity, or wave-particle duality remain uninterpreted. Thus, this paper introduces a conception about the structure of the universe. It assumes that the universe is a medium of ultra-tiny homogeneous particles which are still undiscovered. Like any medium with certain movements, possibly because of a great asymmetric explosion, vortexes have occurred. A vortex condenses the ultra-tiny particles in its center forming a bigger particle, the bigger particles, in turn, could be trapped in a bigger vortex and condense in its center forming a much bigger particle and so on. This conception describes galaxies, stars, protons as particles at different levels. Existing of the particle’s vortexes make the consistency of the speed of light postulate is not true. This conception shows that the vortex motion dynamic agrees with the motion of all the universe particles at any level. An experiment has been carried out to detect the orbiting effect of aggregated vortexes of aligned atoms of a permanent magnet. Based on the described particle’s structure, the gravity force of a particle and attraction between particles as well as charge, electric and magnetic fields and quantum mechanics characteristics are interpreted. All augmented physics phenomena are solved.
Abstract: Customer relationship analysis is vital for retail stores, especially for supermarkets. The point of sale (POS) systems make it possible to record the daily purchasing behaviors of customers as an identification point of sale (ID-POS) database, which can be used to analyze customer behaviors of a supermarket. The customer value is an indicator based on ID-POS database for detecting the customer loyalty of a store. In general, there are many supermarkets in a city, and other nearby competitor supermarkets significantly affect the customer value of customers of a supermarket. However, it is impossible to get detailed ID-POS databases of competitor supermarkets. This study firstly focused on the customer value and distance between a customer's home and supermarkets in a city, and then constructed the models based on logistic regression analysis to analyze correlations between distance and purchasing behaviors only from a POS database of a supermarket chain. During the modeling process, there are three primary problems existed, including the incomparable problem of customer values, the multicollinearity problem among customer value and distance data, and the number of valid partial regression coefficients. The improved customer value, Huff’s gravity model, and inverse attractiveness frequency are considered to solve these problems. This paper presents three types of models based on these three methods for loyal customer classification and competitors’ influence analysis. In numerical experiments, all types of models are useful for loyal customer classification. The type of model, including all three methods, is the most superior one for evaluating the influence of the other nearby supermarkets on customers' purchasing of a supermarket chain from the viewpoint of valid partial regression coefficients and accuracy.
Abstract: Unsteady effects of MHD free convection flow past in an infinite vertical plate with heat source in presence of radiation with reference to all critical parameters that appear in field equations are studied in this paper. The governing equations are developed by usual Boussinesq’s approximation. The problem is solved by using perturbation technique. The results are obtained for velocity, temperature, Nusselt number and skin-friction. The effects of magnetic parameter, prandtl number, Grashof number, permeability parameter, heat source/sink parameter and radiation parameter are discussed on flow characteristics and shown by means of graphs and tables.