Forecasting 24-Hour Ahead Electricity Load Using Time Series Models

Forecasting electricity load is important for various purposes like planning, operation and control. Forecasts can save operating and maintenance costs, increase the reliability of power supply and delivery systems, and correct decisions for future development. This paper compares various time series methods to forecast 24 hours ahead of electricity load. The methods considered are the Holt-Winters smoothing, SARIMA Modeling, LSTM Network, Fbprophet and Tensorflow probability. The performance of each method is evaluated by using the forecasting accuracy criteria namely, the Mean Absolute Error and Root Mean Square Error. The National Renewable Energy Laboratory (NREL) residential energy consumption data are used to train the models. The results of this study show that SARIMA model is superior to the others for 24 hours ahead forecasts. Furthermore, a Bagging technique is used to make the predictions more robust. The obtained results show that by Bagging multiple time-series forecasts we can improve the robustness of the models for 24 hour ahead electricity load forecasting.

Economic Model of Sustainable Value Chain in Passenger Waterway Transportation Service

The service of passenger waterway transportation lacks economic models that help in designing and implementing strategies to ensure its sustainability in several aspects (economic, social and environmental). The size of costs, though not the only one, is of particular importance in these models. However, traditionally, cost management has been focused only on reducing production costs, for the purpose of companies to keep prices low and gain market competitiveness. Although, with all the technological advances, and other restrictions imposed by the market in terms of service, in the case of passengers waterway transportation: intermodal competition; quality of service; or by regulatory environment for public concession and; in the aspect of business: to stay in the market with natural, demand and institutional restrictions, this view is not enough. Thus, there is an evolution of a traditional cost accounting to strategic cost management. On the other hand, it is important to consider other important dimensions and recognize that companies no longer exist in isolation, but they are part of highly integrated value and supplies chains. Therefore, this work will explore and analyze the sustainable value chain of passenger waterway transportation service using the tools of strategic cost management. The method will start from three components of analysis: (1) definition of basic elements of sustainable value chain; (2) identification of main restrictions to the chain development and aspects critical for service sustainability; (3) development of a cost model and propositions to overcome the bottlenecks found, to add value. Whether in the internal cost structure of the company; operational cost reduction strategies; in search of new markets, or to establish new partnerships or even; in the broadest level, in terms of investments in infrastructure or recommendations involving governance decisions to improve the current institutional environment. The case study will be developed in passenger transport companies located in the Lower Amazon, consolidated in this market, with defined enterprise structure of business sustainability, and who have already been willing to collaborate with the investigation. As results, it is expected to understand the cost structures that support sustainable value chains, namely, costs of activities and relevant cost objects in order to determine the cost drivers, profitability margins, cost reduction opportunities and conditions conducive to competitive advantages related to the different strategic options to cost leadership, differentiation or approach. Finally, in the model to be developed, the proper characterization of cost structure and value creation in transport processes under study may constitute reference points for future more sophisticated applied works of optimizing the resources involved and supporting the decision making, in particular with regard to operations research and quantitative methods more robust.

An Improved Adaptive Dot-Shape Beamforming Algorithm Research on Frequency Diverse Array

Frequency diverse array (FDA) beamforming is a technology developed in recent years, and its antenna pattern has a unique angle-distance-dependent characteristic. However, the beam is always required to have strong concentration, high resolution and low sidelobe level to form the point-to-point interference in the concentrated set. In order to eliminate the angle-distance coupling of the traditional FDA and to make the beam energy more concentrated, this paper adopts a multi-carrier FDA structure based on proposed power exponential frequency offset to improve the array structure and frequency offset of the traditional FDA. The simulation results show that the beam pattern of the array can form a dot-shape beam with more concentrated energy, and its resolution and sidelobe level performance are improved. However, the covariance matrix of the signal in the traditional adaptive beamforming algorithm is estimated by the finite-time snapshot data. When the number of snapshots is limited, the algorithm has an underestimation problem, which leads to the estimation error of the covariance matrix to cause beam distortion, so that the output pattern cannot form a dot-shape beam. And it also has main lobe deviation and high sidelobe level problems in the case of limited snapshot. Aiming at these problems, an adaptive beamforming technique based on exponential correction for multi-carrier FDA is proposed to improve beamforming robustness. The steps are as follows: first, the beamforming of the multi-carrier FDA is formed under linear constrained minimum variance (LCMV) criteria. Then the eigenvalue decomposition of the covariance matrix is ​​performed to obtain the diagonal matrix composed of the interference subspace, the noise subspace and the corresponding eigenvalues. Finally, the correction index is introduced to exponentially correct the small eigenvalues ​​of the noise subspace, improve the divergence of small eigenvalues ​​in the noise subspace, and improve the performance of beamforming. The theoretical analysis and simulation results show that the proposed algorithm can make the multi-carrier FDA form a dot-shape beam at limited snapshots, reduce the sidelobe level, improve the robustness of beamforming, and have better performance.

A Speeded up Robust Scale-Invariant Feature Transform Currency Recognition Algorithm

All currencies around the world look very different from each other. For instance, the size, color, and pattern of the paper are different. With the development of modern banking services, automatic methods for paper currency recognition become important in many applications like vending machines. One of the currency recognition architecture’s phases is Feature detection and description. There are many algorithms that are used for this phase, but they still have some disadvantages. This paper proposes a feature detection algorithm, which merges the advantages given in the current SIFT and SURF algorithms, which we call, Speeded up Robust Scale-Invariant Feature Transform (SR-SIFT) algorithm. Our proposed SR-SIFT algorithm overcomes the problems of both the SIFT and SURF algorithms. The proposed algorithm aims to speed up the SIFT feature detection algorithm and keep it robust. Simulation results demonstrate that the proposed SR-SIFT algorithm decreases the average response time, especially in small and minimum number of best key points, increases the distribution of the number of best key points on the surface of the currency. Furthermore, the proposed algorithm increases the accuracy of the true best point distribution inside the currency edge than the other two algorithms.

Image Dehazing Using Dark Channel Prior and Fast Guided Filter in Daubechies Lifting Wavelet Transform Domain

In this paper a method for image dehazing is proposed in lifting wavelet transform domain. Lifting Daubechies (D4) wavelet has been used to obtain the approximate image and detail images.  As the haze is contained in low frequency part, only the approximate image is used for further processing. This region is processed by dehazing algorithm based on dark channel prior (DCP). The dehazed approximate image is then recombined with the detail images using inverse lifting wavelet transform. Implementation of lifting wavelet transform has the advantage of auxiliary memory saving, fast implementation and simplicity. Also, the proposed method deals with near white scene problem, blue horizon issue and localized light sources in a way to enhance image quality and makes the algorithm robust. Simulation results present improvement in terms of visual quality, parameters such as root mean square (RMS) contrast, structural similarity index (SSIM), entropy and execution time.

GPS and Discrete Kalman Filter for Indoor Robot Navigation

This paper discusses the implementation of the Kalman Filter along with the Global Positioning System (GPS) for indoor robot navigation. Two dimensional coordinates is used for the map building, and refers to the global coordinate which is attached to the reference landmark for position and direction information the robot gets. The Discrete Kalman Filter is used to estimate the robot position, project the estimated current state ahead in time through time update and adjust the projected estimated state by an actual measurement at that time via the measurement update. The navigation test has been performed and has been found to be robust.

Acquiring Contour Following Behaviour in Robotics through Q-Learning and Image-based States

In this work a visual and reactive contour following behaviour is learned by reinforcement. With artificial vision the environment is perceived in 3D, and it is possible to avoid obstacles that are invisible to other sensors that are more common in mobile robotics. Reinforcement learning reduces the need for intervention in behaviour design, and simplifies its adjustment to the environment, the robot and the task. In order to facilitate its generalisation to other behaviours and to reduce the role of the designer, we propose a regular image-based codification of states. Even though this is much more difficult, our implementation converges and is robust. Results are presented with a Pioneer 2 AT on a Gazebo 3D simulator.

An Improved Resource Discovery Approach Using P2P Model for Condor: A Grid Middleware

Resource Discovery in Grids is critical for efficient resource allocation and management. Heterogeneous nature and dynamic availability of resources make resource discovery a challenging task. As numbers of nodes are increasing from tens to thousands, scalability is essentially desired. Peer-to-Peer (P2P) techniques, on the other hand, provide effective implementation of scalable services and applications. In this paper we propose a model for resource discovery in Condor Middleware by using the four axis framework defined in P2P approach. The proposed model enhances Condor to incorporate functionality of a P2P system, thus aim to make Condor more scalable, flexible, reliable and robust.

Sensor Fusion Based Discrete Kalman Filter for Outdoor Robot Navigation

The objective of the presented work is to implement the Kalman Filter into an application that reduces the influence of the environmental changes over the robot expected to navigate over a terrain of varying friction properties. The Discrete Kalman Filter is used to estimate the robot position, project the estimated current state ahead at time through time update and adjust the projected estimated state by an actual measurement at that time via the measurement update using the data coming from the infrared sensors, ultrasonic sensors and the visual sensor respectively. The navigation test has been performed in a real world environment and has been found to be robust.

Kernel Matching versus Inverse Probability Weighting: A Comparative Study

Recent quasi-experimental evaluation of the Canadian Active Labour Market Policies (ALMP) by Human Resources and Skills Development Canada (HRSDC) has provided an opportunity to examine alternative methods to estimating the incremental effects of Employment Benefits and Support Measures (EBSMs) on program participants. The focus of this paper is to assess the efficiency and robustness of inverse probability weighting (IPW) relative to kernel matching (KM) in the estimation of program effects. To accomplish this objective, the authors compare pairs of 1,080 estimates, along with their associated standard errors, to assess which type of estimate is generally more efficient and robust. In the interest of practicality, the authorsalso document the computationaltime it took to produce the IPW and KM estimates, respectively.

Reentry Trajectory Optimization Based on Differential Evolution

Reentry trajectory optimization is a multi-constraints optimal control problem which is hard to solve. To tackle it, we proposed a new algorithm named CDEN(Constrained Differential Evolution Newton-Raphson Algorithm) based on Differential Evolution( DE) and Newton-Raphson.We transform the infinite dimensional optimal control problem to parameter optimization which is finite dimensional by discretize control parameter. In order to simplify the problem, we figure out the control parameter-s scope by process constraints. To handle constraints, we proposed a parameterless constraints handle process. Through comprehensive analyze the problem, we use a new algorithm integrated by DE and Newton-Raphson to solve it. It is validated by a reentry vehicle X-33, simulation results indicated that the algorithm is effective and robust.

Some Third Order Methods for Solving Systems of Nonlinear Equations

Based on Traub-s methods for solving nonlinear equation f(x) = 0, we develop two families of third-order methods for solving system of nonlinear equations F(x) = 0. The families include well-known existing methods as special cases. The stability is corroborated by numerical results. Comparison with well-known methods shows that the present methods are robust. These higher order methods may be very useful in the numerical applications requiring high precision in their computations because these methods yield a clear reduction in number of iterations.