Abstract: Over the past decade, there have been promising developments in Natural Language Processing (NLP) with several investigations of approaches focusing on Recognizing Textual Entailment (RTE). These models include models based on lexical similarities, models based on formal reasoning, and most recently deep neural models. In this paper, we present a sentence encoding model that exploits the sentence-to-sentence relation information for RTE. In terms of sentence modeling, Convolutional neural network (CNN) and recurrent neural networks (RNNs) adopt different approaches. RNNs are known to be well suited for sequence modeling, whilst CNN is suited for the extraction of n-gram features through the filters and can learn ranges of relations via the pooling mechanism. We combine the strength of RNN and CNN as stated above to present a unified model for the RTE task. Our model basically combines relation vectors computed from the phrasal representation of each sentence and final encoded sentence representations. Firstly, we pass each sentence through a convolutional layer to extract a sequence of higher-level phrase representation for each sentence from which the first relation vector is computed. Secondly, the phrasal representation of each sentence from the convolutional layer is fed into a Bidirectional Long Short Term Memory (Bi-LSTM) to obtain the final sentence representations from which a second relation vector is computed. The relations vectors are combined and then used in then used in the same fashion as attention mechanism over the Bi-LSTM outputs to yield the final sentence representations for the classification. Experiment on the Stanford Natural Language Inference (SNLI) corpus suggests that this is a promising technique for RTE.
Abstract: Several meteorological parameters were used for the
prediction of monthly average daily global solar radiation on
horizontal using recurrent neural networks (RNNs). Climatological
data and measures, mainly air temperature, humidity, sunshine
duration, and wind speed between 1995 and 2007 were used to design
and validate a feed forward and recurrent neural network based
prediction systems. In this paper we present our reference system
based on a feed-forward multilayer perceptron (MLP) as well as the
proposed approach based on an RNN model. The obtained results
were promising and comparable to those obtained by other existing
empirical and neural models. The experimental results showed the
advantage of RNNs over simple MLPs when we deal with time series
solar radiation predictions based on daily climatological data.
Abstract: Microstrip lines, widely used for good reason, are
broadband in frequency and provide circuits that are compact and
light in weight. They are generally economical to produce since they
are readily adaptable to hybrid and monolithic integrated circuit (IC)
fabrication technologies at RF and microwave frequencies. Although,
the existing EM simulation models used for the synthesis and
analysis of microstrip lines are reasonably accurate, they are
computationally intensive and time consuming. Neural networks
recently gained attention as fast and flexible vehicles to microwave
modeling, simulation and optimization. After learning and
abstracting from microwave data, through a process called training,
neural network models are used during microwave design to provide
instant answers to the task learned.This paper presents simple and
accurate ANN models for the synthesis and analysis of Microstrip
lines to more accurately compute the characteristic parameters and
the physical dimensions respectively for the required design
specifications.
Abstract: Among neural models the Support Vector Machine
(SVM) solutions are attracting increasing attention, mostly because
they eliminate certain crucial questions involved by neural network
construction. The main drawback of standard SVM is its high
computational complexity, therefore recently a new technique, the
Least Squares SVM (LS–SVM) has been introduced. In this paper we
present an extended view of the Least Squares Support Vector
Regression (LS–SVR), which enables us to develop new
formulations and algorithms to this regression technique. Based on
manipulating the linear equation set -which embodies all information
about the regression in the learning process- some new methods are
introduced to simplify the formulations, speed up the calculations
and/or provide better results.