Abstract: In conjunction with the problem of the target selection on a clutter background, the analysis of the scanning rate influence on the spatial-temporal signal structure, the generalized multivariate correlation function and the quality of the resolution with the increase pulse repetition frequency is made. The possibility of the object space-distance resolution, which is conditioned by the range-to-angle conversion with an increased scanning rate, is substantiated. The calculations for the real cylindrical array at high scanning rate are presented. The high scanning rate let to get the signal to noise improvement of the order of 10 dB for the space-time signal processing.
Abstract: Networks are often presented as containing a “core”
and a “periphery.” The existence of a core suggests that some
vertices are central and form the skeleton of the network, to which
all other vertices are connected. An alternative view of graphs is
through communities. Multiple measures have been proposed for
dense communities in graphs, the most classical being k-cliques,
k-cores, and k-plexes, all presenting groups of tightly connected
vertices. We here show that the edge number thresholds for such
communities to emerge and for their percolation into a single dense
connectivity component are very close, in all networks studied. These
percolating cliques produce a natural core and periphery structure.
This result is generic and is tested in configuration models and in
real-world networks. This is also true for k-cores and k-plexes. Thus,
the emergence of this connectedness among communities leading to
a core is not dependent on some specific mechanism but a direct
result of the natural percolation of dense communities.
Abstract: Environmental issues and the ever-increasing in demand of electrical energy make it necessary to have distributed generation (DG) resources in the power system. In this research, in order to realize the goals of reducing losses and improving the voltage profile in a microgrid, the allocation and sizing of DGs have been used. The proposed Genetic Algorithm (GA) is described from the array of artificial intelligence methods for solving the problem. The algorithm is implemented on the IEEE 33 buses network. This study is presented in two scenarios, primarily to illustrate the effect of location and determination of DGs has been done to reduce losses and improve the voltage profile. On the other hand, decisions made with the one-level assumptions of load are not universally accepted for all levels of load. Therefore, in this study, load modelling is performed and the results are presented for multi-levels load state.
Abstract: This paper presents the techno-economic evaluation of run-of-river small-hydropower plants. In this regard, a multi-objective optimization procedure is proposed for the optimal sizing of the hydropower plants, and NSGAII is employed as the optimization algorithm. Annual generated energy and investment cost are considered as the objective functions, and number of generator units (n) and nominal turbine flow rate (QT) constitute the decision variables. Site of Yeripao in Benin is considered as the case study. We have categorized the river of this site using its environmental characteristics: gross head, and first quartile, median, third quartile and mean of flow. Effects of each decision variable on the objective functions are analysed. The results gave Pareto Front which represents the trade-offs between annual energy generation and the investment cost of hydropower plants, as well as the recommended optimal solutions. We noted that with the increase of the annual energy generation, the investment cost rises. Thus, maximizing energy generation is contradictory with minimizing the investment cost. Moreover, we have noted that the solutions of Pareto Front are grouped according to the number of generator units (n). The results also illustrate that the costs per kWh are grouped according to the n and rise with the increase of the nominal turbine flow rate. The lowest investment costs per kWh are obtained for n equal to one and are between 0.065 and 0.180 €/kWh. Following the values of n (equal to 1, 2, 3 or 4), the investment cost and investment cost per kWh increase almost linearly with increasing the nominal turbine flowrate while annual generated. Energy increases logarithmically with increasing of the nominal turbine flowrate. This study made for the Yeripao river can be applied to other rivers with their own characteristics.
Abstract: Given the fact that the pharmaceutical industry is a commonly studied sector in the context of innovation, the majority of innovation research is devoted to the developed markets known by high research and development (R&D) assets and intensive innovation. In contrast, in developing countries where R&D assets are very low, there is relatively little research to mention in the area of pharmaceutical sector innovation, characterized mainly by two principal elements which are the presence of foreign-owned firms and licensed manufacturing agreements between local firms and multinationals. With the scarcity of research in this field, this paper attempts to study the effect of these two elements on the firms’ innovation tendencies. Other traditional factors that influence innovation, which are the age and the size of the firm, the R&D activities and the market structure, revealed in the literature review, will be included in the study in order to try to make this work more exhaustive. The study starts by examining innovation tendency in pharmaceutical firms located in developing countries before analyzing the effect of foreign-owned firms and licensed manufacturing agreements between local firms and multinationals on technological, organizational and marketing innovation. Based on the related work and on the theoretical framework developed, there is a probability that foreign-owned firms and licensed manufacturing agreements between local firms and multinationals have a negative influence on technological innovation. The opposite effect is possible in the case of organizational and marketing innovation.
Abstract: This paper reviews the different concepts that have been investigated concerning the mechanization of fruit thinning as well as multiple working principles and solutions that have been developed for feature extraction of horticultural products, both in the field and industrial environments. The research should be committed towards selective methods, which inevitably need to incorporate some kinds of sensor technology. Computer vision often comes out as an obvious solution for unstructured detection problems, although leaves despite the chosen point of view frequently occlude fruits. Further research on non-traditional sensors that are capable of object differentiation is needed. Ultrasonic and Near Infrared (NIR) technologies have been investigated for applications related to horticultural produce and show a potential to satisfy this need while simultaneously providing spatial information as time of flight sensors. Light Detection and Ranging (LIDAR) technology also shows a huge potential but it implies much greater costs and the related equipment is usually much larger, making it less suitable for portable devices, which may serve a purpose on smaller unstructured orchards. Portable devices may serve a purpose on these types of orchards. In what concerns sensor methods, on-tree fruit detection, major challenge is to overcome the problem of fruits’ occlusion by leaves and branches. Hence, nontraditional sensors capable of providing some type of differentiation should be investigated.
Abstract: Alternative electrode materials for optoelectronic devices have been widely investigated in recent years. Since indium tin oxide (ITO) is the most preferred transparent conductive electrode, producing ITO films by simple and cost-effective solution-based techniques with enhanced optical and electrical properties has great importance. In this study, single- and multi-walled carbon nanotubes (SWCNT and MWCNT) incorporated into the ITO structure to increase electrical conductivity, mechanical strength, and chemical stability. Carbon nanotubes (CNTs) were firstly functionalized by acid treatment (HNO3:H2SO4), and the thermal resistance of CNTs after functionalization was determined by thermogravimetric analysis (TGA). Thin films were then prepared by spin coating technique and characterized by scanning electron microscopy (SEM), X-ray diffraction (XRD), four-point probe measurement system and UV-Vis spectrophotometer. The effects of process parameters were compared for ITO, MWCNT-ITO, and SWCNT-ITO films. Two factors including CNT concentration and annealing temperature were considered. The UV-Vis measurements demonstrated that the transmittance of ITO films was 83.58% at 550 nm, which was decreased depending on the concentration of CNT dopant. On the other hand, both CNT dopants provided an enhancement in the crystalline structure and electrical conductivity. Due to compatible diameter and better dispersibility of SWCNTs in the ITO solution, the best result in terms of electrical conductivity was obtained by SWCNT-ITO films with the 0.1 g/L SWCNT dopant concentration and heat-treatment at 550 °C for 1 hour.
Abstract: The using of finite element programs in analyzing and designing buildings are becoming very popular, but there are many engineers still using the tributary area method (TAM) in designing the structural members such as columns. This study is an attempt to investigate the accuracy of the TAM results with different load condition (gravity and lateral load), different floors numbers, and different columns stiffness's. To conduct this study, linear elastic analysis in ETABS program is used. The results from finite element method are compared to those obtained from TAM. According to the analysis of the data obtained, it can be seen that there is significance difference between the real load carried by columns and the load which is calculated by using the TAM. Thus, using 3-D models are the best choice to calculate the real load effected on columns and design these columns according to this load.
Abstract: To understand the risk for dry storage concrete casks in the cask loading, transfer, and storage phase, the purpose of this research is to establish the probabilistic risk assessment (PRA) analysis methodology for dry storage concrete casks by using SAPHIRE 8 code. This analysis methodology is used to perform the study of Taiwan nuclear power plants (NPPs) dry storage system. The process of research has three steps. First, the data of the concrete casks and Taiwan NPPs are collected. Second, the PRA analysis methodology is developed by using SAPHIRE 8. Third, the PRA analysis is performed by using this methodology. According to the analysis results, the maximum risk is the multipurpose canister (MPC) drop case.
Abstract: Autonomous vehicles offer the possibility of significant benefits to social welfare. However, fully automated cars might not be going to happen in the near further. To speed the adoption of the self-driving technologies, many governments worldwide are passing laws requiring data recorders for the testing of autonomous vehicles. Currently, the self-driving vehicle, (e.g., shuttle bus) has to be monitored from a remote control center. When an autonomous vehicle encounters an unexpected driving environment, such as road construction or an obstruction, it should request assistance from a remote operator. Nevertheless, large amounts of data, including images, radar and lidar data, etc., have to be transmitted from the vehicle to the remote center. Therefore, this paper proposes a data compression method of in-vehicle networks for remote monitoring of autonomous vehicles. Firstly, the time-series data are rearranged into a multi-dimensional signal space. Upon the arrival, for controller area networks (CAN), the new data are mapped onto a time-data two-dimensional space associated with the specific CAN identity. Secondly, the data are sampled based on differential sampling. Finally, the whole set of data are encoded using existing algorithms such as Huffman, arithmetic and codebook encoding methods. To evaluate system performance, the proposed method was deployed on an in-house built autonomous vehicle. The testing results show that the amount of data can be reduced as much as 1/7 compared to the raw data.
Abstract: In the case of the proposed method, the problem is
parallelized by considering multiple possible mode of operation
profiles, which determine the range in which the generators operate
in each period. For each of these profiles, the optimization is carried
out independently, and the best resulting dispatch is chosen. For each
such profile, the resulting problem is a quadratic programming (QP)
problem with a potentially negative definite Q quadratic term, and
constraints depending on the actual operation profile. In this paper we
analyze the performance of available MATLAB optimization methods
and solvers for the corresponding QP.
Abstract: Data on various aspects of education are collected at the institutional and government level regularly. In Australia, for example, students at various levels of schooling undertake examinations in numeracy and literacy as part of NAPLAN testing, enabling longitudinal assessment of such data as well as comparisons between schools and states within Australia. Another source of educational data collected internationally is via the PISA study which collects data from several countries when students are approximately 15 years of age and enables comparisons in the performance of science, mathematics and English between countries as well as ranking of countries based on performance in these standardised tests. As well as student and school outcomes based on the tests taken as part of the PISA study, there is a wealth of other data collected in the study including parental demographics data and data related to teaching strategies used by educators. Overall, an abundance of educational data is available which has the potential to be used to help improve educational attainment and teaching of content in order to improve learning outcomes. A multivariate assessment of such data enables multiple variables to be considered simultaneously and will be used in the present study to help develop profiles of students based on performance in mathematics using data obtained from the PISA study.
Abstract: Link adaptation is an important strategy for achieving robust wireless multimedia communications based on quality of service (QoS) demand. Scheme switching in multiple-input multiple-output (MIMO) systems is an aspect of link adaptation, and it involves selecting among different MIMO transmission schemes or modes so as to adapt to the varying radio channel conditions for the purpose of achieving QoS delivery. However, finding the most appropriate switching method in MIMO links is still a challenge as existing methods are either computationally complex or not always accurate. This paper presents an intelligent switching method for the MIMO system consisting of two schemes - transmit diversity (TD) and spatial multiplexing (SM) - using fuzzy logic technique. In this method, two channel quality indicators (CQI) namely average received signal-to-noise ratio (RSNR) and received signal strength indicator (RSSI) are measured and are passed as inputs to the fuzzy logic system which then gives a decision – an inference. The switching decision of the fuzzy logic system is fed back to the transmitter to switch between the TD and SM schemes. Simulation results show that the proposed fuzzy logic – based switching technique outperforms conventional static switching technique in terms of bit error rate and spectral efficiency.
Abstract: Machine Learning and Data Mining are the two important tools for extracting useful information and knowledge from large datasets. In machine learning, classification is a wildly used technique to predict qualitative variables and is generally preferred over regression from an operational point of view. Due to the enormous increase in air pollution in various countries especially China, Air Quality Classification has become one of the most important topics in air quality research and modelling. This study aims at introducing a hybrid classification model based on information theory and Support Vector Machine (SVM) using the air quality data of four cities in China namely Beijing, Guangzhou, Shanghai and Tianjin from Jan 1, 2014 to April 30, 2016. China's Ministry of Environmental Protection has classified the daily air quality into 6 levels namely Serious Pollution, Severe Pollution, Moderate Pollution, Light Pollution, Good and Excellent based on their respective Air Quality Index (AQI) values. Using the information theory, information gain (IG) is calculated and feature selection is done for both categorical features and continuous numeric features. Then SVM Machine Learning algorithm is implemented on the selected features with cross-validation. The final evaluation reveals that the IG and SVM hybrid model performs better than SVM (alone), Artificial Neural Network (ANN) and K-Nearest Neighbours (KNN) models in terms of accuracy as well as complexity.
Abstract: Chinese overall urban design faces a large number of problems such as the neglect of urban characteristics, generalization of content, and difficulty in implementation. Focusing on these issues, this paper proposes the main points of shaping urban characteristics in overall urban design: focuses on core problems in city function and scale, landscape pattern, historical culture, social resources and modern city style and digs the urban characteristic genes. Then, we put forward “core problem location and characteristic gene enhancement” as a kind of overall urban design technical method. Firstly, based on the main problems in urban space as a whole, for the operability goal, the method extracts the key genes and integrates into the multi-dimension system in a targeted manner. Secondly, hierarchical management and guidance system is established which may be in line with administrative management. Finally, by converting the results, action plan is drawn up that can be dynamically implemented. Based on the above idea and method, a practical exploration has been performed in the case of Xiangyang central city.
Abstract: The quality of press-fit assembly is closely related to
reliability and safety of product. The paper proposed a keypoint
detection method based on convolutional neural network to improve
the accuracy of keypoint detection in press-fit curve. It would
provide an auxiliary basis for judging quality of press-fit assembly.
The press-fit curve is a curve of press-fit force and displacement.
Both force data and distance data are time-series data. Therefore,
one-dimensional convolutional neural network is used to process
the press-fit curve. After the obtained press-fit data is filtered, the
multi-layer one-dimensional convolutional neural network is used to
perform the automatic learning of press-fit curve features, and then
sent to the multi-layer perceptron to finally output keypoint of the
curve. We used the data of press-fit assembly equipment in the actual
production process to train CNN model, and we used different data
from the same equipment to evaluate the performance of detection.
Compared with the existing research result, the performance of
detection was significantly improved. This method can provide a
reliable basis for the judgment of press-fit quality.
Abstract: Computacional recognition of sign languages aims to
allow a greater social and digital inclusion of deaf people through
interpretation of their language by computer. This article presents
a model of recognition of two of global parameters from sign
languages; hand configurations and hand movements. Hand motion
is captured through an infrared technology and its joints are built
into a virtual three-dimensional space. A Multilayer Perceptron
Neural Network (MLP) was used to classify hand configurations and
Dynamic Time Warping (DWT) recognizes hand motion. Beyond
of the method of sign recognition, we provide a dataset of
hand configurations and motion capture built with help of fluent
professionals in sign languages. Despite this technology can be
used to translate any sign from any signs dictionary, Brazilian
Sign Language (Libras) was used as case study. Finally, the model
presented in this paper achieved a recognition rate of 80.4%.
Abstract: The challenge for everyone involved in preserving the ecosystem is to find creative ways to protect and restore the remaining ecosystems while accommodating and enhancing the country social and economic well-being. Frequent fires of anthropogenic origin have been affecting the ecosystems in many countries adversely. Hence adopting ways of decision making such as Multicriteria Decision Making (MCDM) is appropriate since it will enhance the evaluation and analysis of fire risk and hazard of the ecosystem. In this paper, fire risk and hazard data from the West Gonja area of Ghana were used in some of the methods (Analytical Hierarchy Process, Compromise Programming, and Grey Relational Analysis (GRA) for MCDM evaluation and analysis to determine the optimal weight method for fire risk and hazard. Ranking of the land cover types was carried out using; Fire Hazard, Fire Fighting Capacity and Response Risk Criteria. Pairwise comparison under Analytic Hierarchy Process (AHP) was used to determine the weight of the various criteria. Weights for sub-criteria were also obtained by the pairwise comparison method. The results were optimised using GRA and Compromise Programming (CP). The results from each method, hybrid GRA and CP, were compared and it was established that all methods were satisfactory in terms of optimisation of weight. The most optimal method for spatial multicriteria evaluation was the hybrid GRA method. Thus, a hybrid AHP and GRA method is more effective method for ranking alternatives in MCDM than the hybrid AHP and CP method.
Abstract: In the course of recent decades, medical imaging has
been dominated by the use of costly film media for review and
archival of medical investigation, however due to developments in
networks technologies and common acceptance of a standard digital
imaging and communication in medicine (DICOM) another approach
in light of World Wide Web was produced. Web technologies
successfully used in telemedicine applications, the combination of
web technologies together with DICOM used to design a web-based
and open source DICOM viewer. The Web server allowance to
inquiry and recovery of images and the images viewed/manipulated
inside a Web browser without need for any preinstalling software.
The dynamic site page for medical images visualization and
processing created by using JavaScript and HTML5 advancements.
The XAMPP ‘apache server’ is used to create a local web server for
testing and deployment of the dynamic site. The web-based viewer
connected to multiples devices through local area network (LAN) to
distribute the images inside healthcare facilities. The system offers a
few focal points over ordinary picture archiving and communication
systems (PACS): easy to introduce, maintain and independently
platforms that allow images to display and manipulated efficiently,
the system also user-friendly and easy to integrate with an existing
system that have already been making use of web technologies. The
wavelet-based image compression technique on which 2-D discrete
wavelet transform used to decompose the image then wavelet
coefficients are transmitted by entropy encoding after threshold to
decrease transmission time, stockpiling cost and capacity. The
performance of compression was estimated by using images quality
metrics such as mean square error ‘MSE’, peak signal to noise ratio
‘PSNR’ and compression ratio ‘CR’ that achieved (83.86%) when
‘coif3’ wavelet filter is used.
Abstract: The quantum communication technology is an evolving
design which connects multiple quantum enabled devices to internet
for secret communication or sensitive information exchange. In
future, the number of these compact quantum enabled devices
will increase immensely making them an integral part of present
communication systems. Therefore, safety and security of such
devices is also a major concern for us. To ensure the customer
sensitive information will not be eavesdropped or deciphered, we
need a strong authentications and encryption mechanism. In this
paper, we propose a mutual authentication scheme between these
smart quantum devices and server based on the secure exchange of
information through quantum channel which gives better solutions
for symmetric key exchange issues. An important part of this
work is to propose a secure mutual authentication protocol over
the quantum channel. We show that our approach offers robust
authentication protocol and further our solution is lightweight,
scalable, cost-effective with optimized computational processing
overheads.