Fighter Aircraft Evaluation and Selection Process Based on Triangular Fuzzy Numbers in Multiple Criteria Decision Making Analysis Using the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS)

This article presents a multiple criteria evaluation approach to uncertainty, vagueness, and imprecision analysis for ranking alternatives with fuzzy data for decision making using the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS). The fighter aircraft evaluation and selection decision making problem is modeled in a fuzzy environment with triangular fuzzy numbers. The fuzzy decision information related to the fighter aircraft selection problem is taken into account in ordering the alternatives and selecting the best candidate. The basic fuzzy TOPSIS procedure steps transform fuzzy decision matrices into matrices of alternatives evaluated according to all decision criteria. A practical numerical example illustrates the proposed approach to the fighter aircraft selection problem.

Military Combat Aircraft Selection Using Trapezoidal Fuzzy Numbers with the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS)

This article presents a new approach to uncertainty, vagueness, and imprecision analysis for ranking alternatives with fuzzy data for decision making using the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS). In the proposed approach, fuzzy decision information related to the aircraft selection problem is taken into account in ranking the alternatives and selecting the best one. The basic procedural step is to transform the fuzzy decision matrices into matrices of alternatives evaluated according to all decision criteria. A numerical example illustrates the proposed approach for the military combat aircraft selection problem.

Development of Electrospun Membranes with Defined Polyethylene Collagen and Oxide Architectures Reinforced with Medium and High Intensity Statins

Cardiovascular diseases (CVD) are related to affectations of the heart and blood vessels, within these are pathologies such as coronary or peripheral heart disease, caused by the narrowing of the vessel wall (atherosclerosis), which is related to the accumulation of Low-Density Lipoproteins (LDL) in the arterial walls that leads to a progressive reduction of the lumen of the vessel and alterations in blood perfusion. Currently, the main therapeutic strategy for this type of alteration is drug treatment with statins, which inhibit the enzyme 3-hydroxy-3-methyl-glutaryl-CoA reductase (HMG-CoA reductase), responsible for modulating the rate of cholesterol production and other isoprenoids in the mevalonate pathway. This enzyme induces the expression of LDL receptors in the liver, increasing their number on the surface of liver cells, reducing the plasma concentration of cholesterol. On the other hand, when the blood vessel presents stenosis, a surgical procedure with vascular implants is indicated, which are used to restore circulation in the arterial or venous bed. Among the materials used for the development of vascular implants are Dacron® and Teflon®, which perform the function of re-waterproofing the circulatory circuit, but due to their low biocompatibility, they do not have the ability to promote remodeling and tissue regeneration processes. Based on this, the present research proposes the development of a hydrolyzed collagen and polyethylene oxide electrospun membrane reinforced with medium and high-intensity statins, so that in future research it can favor tissue remodeling processes from its microarchitecture.

A Comprehensive Survey on Machine Learning Techniques and User Authentication Approaches for Credit Card Fraud Detection

With the increase of credit card usage, the volume of credit card misuse also has significantly increased, which may cause appreciable financial losses for both credit card holders and financial organizations issuing credit cards. As a result, financial organizations are working hard on developing and deploying credit card fraud detection methods, in order to adapt to ever-evolving, increasingly sophisticated defrauding strategies and identifying illicit transactions as quickly as possible to protect themselves and their customers. Compounding on the complex nature of such adverse strategies, credit card fraudulent activities are rare events compared to the number of legitimate transactions. Hence, the challenge to develop fraud detection that are accurate and efficient is substantially intensified and, as a consequence, credit card fraud detection has lately become a very active area of research. In this work, we provide a survey of current techniques most relevant to the problem of credit card fraud detection. We carry out our survey in two main parts. In the first part, we focus on studies utilizing classical machine learning models, which mostly employ traditional transnational features to make fraud predictions. These models typically rely on some static physical characteristics, such as what the user knows (knowledge-based method), or what he/she has access to (object-based method). In the second part of our survey, we review more advanced techniques of user authentication, which use behavioral biometrics to identify an individual based on his/her unique behavior while he/she is interacting with his/her electronic devices. These approaches rely on how people behave (instead of what they do), which cannot be easily forged. By providing an overview of current approaches and the results reported in the literature, this survey aims to drive the future research agenda for the community in order to develop more accurate, reliable and scalable models of credit card fraud detection.

Neutrosophic Multiple Criteria Decision Making Analysis Method for Selecting Stealth Fighter Aircraft

In this paper, a neutrosophic multiple criteria decision analysis method is proposed to select stealth fighter aircraft. Neutrosophic multiple criteria decision analysis methods are used to analyze the neutrosophic environment and give results under uncertainty and incompleteness. Neutrosophic numbers are used to evaluate alternatives over a set of evaluation criteria in decision making problems. Finally, the proposed model is applied to a practical decision problem for selecting stealth fighter aircraft.

Fundamentals of Performance Management in the World of Public Service Organisations

The examination of the Public Service Organization’s performance evaluation includes several steps that help public organizations to develop a more efficient system. Public sector organizations have different characteristics than the competitive sector, so it can be stated that other/new elements become more important in their performance processes. The literature in this area is diverse, so highlighting an indicator system can be useful for introducing a system, but it is also worthwhile to measure the specific elements of the organization. In the case of a public service organization, due to the service obligation, it is usually possible to talk about a high number of users, so compliance is more difficult. For the organization, it is an important target to place great emphasis on the increase of service standards and the development of related processes. In this research, the health sector is given a prominent role, as it is a sensitive area where both organizational and individual performance is important for all participants. As a primary step, the content of the strategy is decisive, as this is important for the efficient structure of the process. When designing any system, it is important to review the expectations of the stakeholders, as this is primary when considering the design. The goal of this paper is to build the foundations of a performance management and indexing framework that can help a hospital to provide effective feedback and a direction that is important in assessing and developing a service and can become a management philosophy.

An Overview of Technology Availability to Support Remote Decentralized Clinical Trials

Developing new medicine and health solutions and improving patient health currently rely on the successful execution of clinical trials, which generate relevant safety and efficacy data. For their success, recruitment and retention of participants are some of the most challenging aspects of protocol adherence. Main barriers include: i) lack of awareness of clinical trials; ii) long distance from the clinical site; iii) the burden on participants, including the duration and number of clinical visits, and iv) high dropout rate. Most of these aspects could be addressed with a new paradigm, namely the Remote Decentralized Clinical Trials (RDCTs). Furthermore, the COVID-19 pandemic has highlighted additional advantages and challenges for RDCTs in practice, allowing participants to join trials from home and not depending on site visits, etc. Nevertheless, RDCTs should follow the process and the quality assurance of conventional clinical trials, which involve several processes. For each part of the trial, the Building Blocks, existing software and technologies were assessed through a systematic search. The technology needed to perform RDCTs is widely available and validated but is yet segmented and developed in silos, as different software solutions address different parts of the trial and at various levels. The current paper is analyzing the availability of technology to perform RDCTs, identifying gaps and providing an overview of Basic Building Blocks and functionalities that need to be covered to support the described processes.

Software Product Quality Evaluation Model with Multiple Criteria Decision Making Analysis

This paper presents a software product quality evaluation model based on the ISO/IEC 25010 quality model. The evaluation characteristics and sub characteristics were identified from the ISO/IEC 25010 quality model. The multidimensional structure of the quality model is based on characteristics such as functional suitability, performance efficiency, compatibility, usability, reliability, security, maintainability, and portability, and associated sub characteristics. Random numbers are generated to establish the decision maker’s importance weights for each sub characteristics. Also, random numbers are generated to establish the decision matrix of the decision maker’s final scores for each software product against each sub characteristics. Thus, objective criteria importance weights and index scores for datasets were obtained from the random numbers. In the proposed model, five different software product quality evaluation datasets under three different weight vectors were applied to multiple criteria decision analysis method, preference analysis for reference ideal solution (PARIS) for comparison, and sensitivity analysis procedure. This study contributes to provide a better understanding of the application of MCDMA methods and ISO/IEC 25010 quality model guidelines in software product quality evaluation process.

An E-Maintenance IoT Sensor Node Designed for Fleets of Diverse Heavy-Duty Vehicles

E-maintenance is a relatively recent concept, generally referring to maintenance management by monitoring assets over the Internet. One of the key links in the chain of an e-maintenance system is data acquisition and transmission. Specifically for the case of a fleet of heavy-duty vehicles, where the main challenge is the diversity of the vehicles and vehicle-embedded self-diagnostic/reporting technologies, the design of the data acquisition and transmission unit is a demanding task. This is clear if one takes into account that a heavy-vehicles fleet assortment may range from vehicles with only a limited number of analog sensors monitored by dashboard light indicators and gauges to vehicles with plethora of sensors monitored by a vehicle computer producing digital reporting. The present work proposes an adaptable internet of things (IoT) sensor node that is capable of addressing this challenge. The proposed sensor node architecture is based on the increasingly popular single-board computer – expansion boards approach. In the proposed solution, the expansion boards undertake the tasks of position identification, cellular connectivity, connectivity to the vehicle computer, and connectivity to analog and digital sensors by means of a specially targeted design of expansion board. Specifically, the latter offers a number of adaptability features to cope with the diverse sensor types employed in different vehicles. In standard mode, the IoT sensor node communicates to the data center through cellular network, transmitting all digital/digitized sensor data, IoT device identity and position. Moreover, the proposed IoT sensor node offers connectivity, through WiFi and an appropriate application, to smart phones or tablets allowing the registration of additional vehicle- and driver-specific information and these data are also forwarded to the data center. All control and communication tasks of the IoT sensor node are performed by dedicated firmware.

Effect of Needle Height on Discharge Coefficient and Cavitation Number

Cavitation inside diesel injector nozzle is investigated using Reynolds-Stress-Navier stokes equations. Schnerr-Sauer cavitation model is used for modeling cavitation inside diesel injector nozzle. The carrying fluid utilized in the current study is diesel fuel. The flow is verified at the beginning by comparing with the previous experimental data and it was found that K-Epsilon turbulent model could lead to a better accuracy comparing to K-Omega turbulent model. Moreover, mass flow rate obtained numerically is compared with the experimental value and discrepancy was found to be less than 5% - which shows the accuracy of the current results. Finally, a real-size four-hole nozzle is investigated and the flow inside it is visualized based on velocity profile, discharge coefficient and cavitation number. It was found that the mesh density could be reduced significantly by utilizing periodic boundary condition. Velocity contour at the mid nozzle showed that maximum value of velocity occurs at the end of the needle before entering the orifice area. Last but not least, at the same boundary conditions, when different needle heights were utilized, it was found that as needle height increases with an increase in cavitation number, discharge coefficient increases, while the mentioned increases is more tangible at smaller values of needle heights.

Machine Learning Methods for Flood Hazard Mapping

This paper proposes a neural network approach for assessing flood hazard mapping. The core of the model is a machine learning component fed by frequency ratios, namely statistical correlations between flood event occurrences and a selected number of topographic properties. The classification capability was compared with the flood hazard mapping River Basin Plans (Piani Assetto Idrogeologico, acronimed as PAI) designed by the Italian Institute for Environmental Research and Defence, ISPRA (Istituto Superiore per la Protezione e la Ricerca Ambientale), encoding four different increasing flood hazard levels. The study area of Piemonte, an Italian region, has been considered without loss of generality. The frequency ratios may be used as a standalone block to model the flood hazard mapping. Nevertheless, the mixture with a neural network improves the classification power of several percentage points, and may be proposed as a basic tool to model the flood hazard map in a wider scope.

OILU Tag: A Projective Invariant Fiducial System

This paper presents the development of a 2D visual marker, derived from a recent patented work in the field of numbering systems. The proposed fiducial uses a group of projective invariant straight-line patterns, easily detectable and remotely recognizable. Based on an efficient data coding scheme, the developed marker enables producing a large panel of unique real time identifiers with highly distinguishable patterns. The proposed marker Incorporates simultaneously decimal and binary information, making it readable by both humans and machines. This important feature opens up new opportunities for the development of efficient visual human-machine communication and monitoring protocols. Extensive experiment tests validate the robustness of the marker against acquisition and geometric distortions.

Influence of Laser Treatment on the Growth of Sprouts of Different Wheat Varieties

Cereals are considered as a strategic product in human life and their demand is increasing with the growth of world population. Increasing wheat production is important for the country. One of the ways to solve the problem is to develop and implement new, environmentally and economically acceptable technologies. Such technologies include pre-sowing treatment of seed with a laser and associative nitrogen-fixing bacteria - Azospirillum brasilense. In the region there are the wheat varieties - Dika and Lomtagora, which are among the most common in Georgia. Dika is a frost-resistant wheat, with a high ability to adapt to the environment, resistant to falling and it is sown in highlands. Lomtagora 126 differs with its winter and drought resistance, and it has a great ability to germinate. Lomtagora is characterized by a strong root system and a high budding capacity. It is an early variety, fall-resistant, easy to thresh and suitable for mechanized harvesting with large and red grains. This paper presents some preliminary experimental results where a continuous CO2 laser with a power of 25-40 W was used to radiate grains at a flow rate of 10 and 15 cm/sec. The treatment was carried out on grains of the Triticum aestivum L. var. Lutescens (local variety name - Lomtagora 126), and Triticum carthlicum Nevski (local variety name - Dika). Here the grains were treated with A. brasilense isolate (108-109 CFU/ml), which was isolated from the rhizosphere of wheat. It was observed that the germination of the wheat was not significantly influenced by either laser or bacteria treatment. The results of our research show that combined treatment with laser and A. brasilense significantly influenced the germination of wheat. In the case of the Lomtagora 126 variety, grains were exposed to the beam on a speed of 10 cm/sec, only slightly improved the growth for 38-day seedlings, in case of exposition of grains with a speed of 15 cm/sec - by 23%. Treatment of seeds with A. brasilense in both exposed and non-exposed variants led to an improvement in the growth of seedlings, with A. brasilense alone - by 22%, and with combined treatment of grains - by 29%. In the case of the Dika variety, only exposure led to growth by 8-9%, and the combined treatment - by 10-15%, in comparison with the control variant. Superior effect on growth of seedlings of different varieties was achieved with the combinations of laser treatment on grains in a beam of 15 cm/sec (radiation power 30-40 W) and in addition of A. brasilense - nitrogen fixing bacteria. Therefore, this is a promising application of A. brasilense as active agents of bacterial fertilizers due to their ability of molecular nitrogen fixation in cereals in combination with laser irradiation: choosing a proper strain gives a good ability to colonize roots of agricultural crops, providing a high nitrogen-fixing ability and the ability to mobilize soil phosphorus, and laser treatment stimulates natural processes occurring in plant cells, will increase the yield.

Review of Innovation Management Frameworks and Assessment Tools

Research studies are highly fragmented when an Innovation Management Framework is being discussed. With the aim to identify an Innovation Management Framework/Assessment Tool suitable for Small & Medium Enterprises (SMEs) in the service industry, this researcher critically reviewed existing innovation management frameworks and assessment models/tools and discovered a number of literature gaps. It is established that the existing literature lacks generally agreed innovation management dimensions, commonly accepted knowledge creation through empirical studies on innovation management in SMEs, effective innovation management performance measurements, suitable innovation management framework in SMEs, and studies on innovation management in the service industry, in particular in retail SMEs. As such, there is a dire need to develop an appropriate firm-level innovation management framework suitable for SMEs in the service industry for future research projects and further studies. In addition, this researcher also discussed the significance of establishing such an innovation management framework.

Aircraft Selection Using Multiple Criteria Decision Making Analysis Method with Different Data Normalization Techniques

This paper presents an original application of multiple criteria decision making analysis theory to the evaluation of aircraft selection problem. The selection of an optimal, efficient and reliable fleet, network and operations planning policy is one of the most important factors in aircraft selection problem. Given that decision making in aircraft selection involves the consideration of a number of opposite criteria and possible solutions, such a selection can be considered as a multiple criteria decision making analysis problem. This study presents a new integrated approach to decision making by considering the multiple criteria utility theory and the maximal regret minimization theory methods as well as aircraft technical, economical, and environmental aspects. Multiple criteria decision making analysis method uses different normalization techniques to allow criteria to be aggregated with qualitative and quantitative data of the decision problem. Therefore, selecting a suitable normalization technique for the model is also a challenge to provide data aggregation for the aircraft selection problem. To compare the impact of different normalization techniques on the decision problem, the vector, linear (sum), linear (max), and linear (max-min) data normalization techniques were identified to evaluate aircraft selection problem. As a logical implication of the proposed approach, it enhances the decision making process through enabling the decision maker to: (i) use higher level knowledge regarding the selection of criteria weights and the proposed technique, (ii) estimate the ranking of an alternative, under different data normalization techniques and integrated criteria weights after a posteriori analysis of the final rankings of alternatives. A set of commercial passenger aircraft were considered in order to illustrate the proposed approach. The obtained results of the proposed approach were compared using Spearman's rho tests. An analysis of the final rank stability with respect to the changes in criteria weights was also performed so as to assess the sensitivity of the alternative rankings obtained by the application of different data normalization techniques and the proposed approach.

Drug Abuse among Immigrant Youth in Canada

There has been an increased number of immigrants arriving in Canada and a concurrent rise in the number of immigrant youth suffering from drug abuse. Immigrant youths’ drug abuse has become a significant social and public health concern for researchers. This paper explores the nature of immigrant youths’ drug abuse by examining the factors influencing the onset of substance misuse, the barriers that discourage youth to seek out treatment, and how to resolve addictions amidst immigrant youth. Findings demonstrate that diminished parental supervision, acculturation challenges, peer conformity, discrimination, and ethnic marginalization are all significant factors influencing youth to use drugs as an outlet for their pain, while culturally incompetent care and fear of family and culture-based addiction stigma act as barriers discouraging youth from seeking out addiction support. To resolve addiction challenges amidst immigrant youth, future research should focus on promoting and implementing culturally sensitive practices and psychoeducational initiatives into immigrant communities and within public health policies.

Scholar Index for Research Performance Evaluation Using Multiple Criteria Decision Making Analysis

This paper aims to present an objective quantitative methodology on how to evaluate individual’s scholarly research output using multiple criteria decision analysis. A multiple criteria decision making analysis (MCDMA) methodological process is adopted to build a multiple criteria evaluation model. With the introduction of the scholar index, which gives significant information about a researcher's productivity and the scholarly impact of his or her publications in a single number (s is the number of publications with at least s citations); cumulative research citation index; the scholar index is included in the citation databases to cover the multidimensional complexity of scholarly research performance and to undertake objective evaluations with scholar index. The scholar index, one of publication activity indexes, is analyzed by considering it to be the most appropriate sciencemetric indicator which allows to smooth over many drawbacks of scholarly output assessment by mere calculation of the number of publications (quantity) and citations (quality). Hence, this study includes a set of indicators-based scholar index to be used for evaluating scholarly researchers. Google Scholar open science database was used to assess and discuss scholarly productivity and impact of researchers. Based on the experiment of computing the scholar index, and its derivative indexes for a set of researchers on open research database platform, quantitative methods of assessing scholarly research output were successfully considered to rank researchers. The proposed methodology considers the ranking, and the selection of data on which a scholarly research performance evaluation was based, the analysis of the data, and the presentation of the multiple criteria analysis results.

Reducing the Need for Multi-Input Multi-Output in Multi-Beam Base Transceiver Station Antennas Using Orthogonally-Polarized Feeds with an Arbitrary Number of Ports

A multi-beam BTS (Base Transceiver Station) antenna has been developed using dual parabolic cylindrical reflectors. The ±45° polarization feeds are used in spatial diversity MIMO (Multi-Input Multi-Output). They can be replaced by single-port orthogonally polarized feeds. Then, with two sets of beams generated above each other, the ± 45° polarization ports of any conventional transceiver can be connected to two of these beam sets. Thus, with two-port transceivers, the system will be equivalent to 4x4 MIMO, instead of 2x2. Radio Frequency (RF) power combiners/splitters can also be used to combine the multiple beams into a single beam or any arbitrary number of beams/ports. The gain of the combined-beam will be more than 20-24 dBi instead of 17-18 dBi of conventional wide-beam antennas. Furthermore, the gain of the combined beam will be high over the whole beam angle. Moreover, the users will always be close to the peak gain value of the combined beam regardless of their location within the combined beam angle. The frequency bands of all the combined beams are adjusted such that they all have the same frequency band. Different configurations of RF power splitter/combiners can be used to provide any arbitrary number of beams/ports according to the requirements of any existing base station configuration.

Research on the Protection and Reuse Model of Historical Buildings in Chinese Airports

China had constructed a large number of military and civilian airports before and after World War II, and then began large-scale repairs, reconstructions or relocation of airports after the baptism of wars after World War I and World War II. The airport's historical area and its historical buildings such as terminals, hangars, and towers have adopted different protection strategies and reuse application strategies. This paper is based on the judgment of the value of airport historical buildings to study different protection and reuse strategies. The protection and reuse models of historical buildings are classified in three dimensions: the airport historical area, the airport historical building complex and its individual buildings, and combined with specific examples to discuss and summarize the technical characteristics, protection strategies and successful experiences of different modes of protection and reuse of historical areas and historical buildings of airports.

Index t-SNE: Tracking Dynamics of High-Dimensional Datasets with Coherent Embeddings

t-SNE is an embedding method that the data science community has widely used. It helps two main tasks: to display results by coloring items according to the item class or feature value; and for forensic, giving a first overview of the dataset distribution. Two interesting characteristics of t-SNE are the structure preservation property and the answer to the crowding problem, where all neighbors in high dimensional space cannot be represented correctly in low dimensional space. t-SNE preserves the local neighborhood, and similar items are nicely spaced by adjusting to the local density. These two characteristics produce a meaningful representation, where the cluster area is proportional to its size in number, and relationships between clusters are materialized by closeness on the embedding. This algorithm is non-parametric. The transformation from a high to low dimensional space is described but not learned. Two initializations of the algorithm would lead to two different embedding. In a forensic approach, analysts would like to compare two or more datasets using their embedding. A naive approach would be to embed all datasets together. However, this process is costly as the complexity of t-SNE is quadratic, and would be infeasible for too many datasets. Another approach would be to learn a parametric model over an embedding built with a subset of data. While this approach is highly scalable, points could be mapped at the same exact position, making them indistinguishable. This type of model would be unable to adapt to new outliers nor concept drift. This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved. The optimization process minimizes two costs, one relative to the embedding shape and the second relative to the support embedding’ match. The embedding with the support process can be repeated more than once, with the newly obtained embedding. The successive embedding can be used to study the impact of one variable over the dataset distribution or monitor changes over time. This method has the same complexity as t-SNE per embedding, and memory requirements are only doubled. For a dataset of n elements sorted and split into k subsets, the total embedding complexity would be reduced from O(n2) to O(n2/k), and the memory requirement from n2 to 2(n/k)2 which enables computation on recent laptops. The method showed promising results on a real-world dataset, allowing to observe the birth, evolution and death of clusters. The proposed approach facilitates identifying significant trends and changes, which empowers the monitoring high dimensional datasets’ dynamics.