Generative Adversarial Network Based Fingerprint Anti-Spoofing Limitations

Fingerprint Anti-Spoofing approaches have been actively developed and applied in real-world applications. One of the main problems for Fingerprint Anti-Spoofing is not robust to unseen samples, especially in real-world scenarios. A possible solution will be to generate artificial, but realistic fingerprint samples and use them for training in order to achieve good generalization. This paper contains experimental and comparative results with currently popular GAN based methods and uses realistic synthesis of fingerprints in training in order to increase the performance. Among various GAN models, the most popular StyleGAN is used for the experiments. The CNN models were first trained with the dataset that did not contain generated fake images and the accuracy along with the mean average error rate were recorded. Then, the fake generated images (fake images of live fingerprints and fake images of spoof fingerprints) were each combined with the original images (real images of live fingerprints and real images of spoof fingerprints), and various CNN models were trained. The best performances for each CNN model, trained with the dataset of generated fake images and each time the accuracy and the mean average error rate, were recorded. We observe that current GAN based approaches need significant improvements for the Anti-Spoofing performance, although the overall quality of the synthesized fingerprints seems to be reasonable. We include the analysis of this performance degradation, especially with a small number of samples. In addition, we suggest several approaches towards improved generalization with a small number of samples, by focusing on what GAN based approaches should learn and should not learn.

Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Strategic Investment in Infrastructure Development to Facilitate Economic Growth in the United States

The COVID-19 pandemic is unprecedented in terms of its global reach and economic impacts. Historically, investment in infrastructure development projects has been touted to boost the economic growth of a nation. The State and Local governments responsible for delivering infrastructure assets work under tight budgets. Therefore, it is important to understand which infrastructure projects have the highest potential of boosting economic growth in the post-pandemic era. This paper presents relationships between infrastructure projects and economic growth. Statistical relationships between investment in different types of infrastructure projects (transit, water and wastewater, highways, power, manufacturing etc.) and indicators of economic growth are presented using historic data between 2002 and 2020 from the U.S. Census Bureau and U.S. Bureau of Economic Analysis (BEA). The outcome of the paper is the comparison of statistical correlations between investment in different types of infrastructure projects and indicators of economic growth. The comparison of the statistical correlations is useful in ranking the types of infrastructure projects based on their ability to influence economic prosperity. Therefore, investment in the infrastructures with the higher rank will have a better chance of boosting the economic growth. Once, the ranks are derived, they can be used by the decision-makers in infrastructure investment related decision-making process.

Vulnerability Analysis for Risk Zones Boundary Definition to Support a Decision Making Process at CBRNE Operations

An effective emergency response to accidents with chemical, biological, radiological, nuclear, or explosive materials (CBRNE) that represent highly dynamic situations needs immediate actions within limited time, information and resources. The aim of the study is to provide the foundation for division of unsafe area into risk zones according to the impact of hazardous parameters (heat radiation, thermal dose, overpressure, chemical concentrations). A decision on the boundary values for three risk zones is based on the vulnerability analysis that covered a variety of accident scenarios containing the release of a toxic or flammable substance which either evaporates, ignites and/or explodes. Critical values are selected for the boundary definition of the Red, Orange and Yellow risk zones upon the examination of harmful effects that are likely to cause injuries of varying severity to people and different levels of damage to structures. The obtained results provide the basis for creating a comprehensive real-time risk map for a decision support at CBRNE operations.

Electrical Effects during the Wetting-Drying Cycle of Porous Brickwork: Electrical Aspects of Rising Damp

Rising damp is an extremely complex phenomenon that is of great practical interest to the field of building conservation due to the irreversible damages it can make to old and historic structures. The electrical effects occurring in damp masonry have been scarcely researched and are a largely unknown aspect of rising damp. Present paper describes the typical electrical patterns occurring in porous brickwork during a wetting and drying cycle. It has been found that in contrast with dry masonry, where electrical phenomena are virtually non-existent, damp masonry exhibits a wide array of electrical effects. Long-term real-time measurements performed in the lab on small-scale brick structures, using an array of embedded micro-sensors, revealed significant voltage, current, capacitance and resistance variations which can be linked to the movement of moisture inside porous materials. The same measurements performed on actual old buildings revealed a similar behaviour, the electrical effects being more significant in areas of the brickwork affected by rising damp. Understanding these electrical phenomena contributes to a better understanding of the driving mechanisms of rising damp, potentially opening new avenues of dealing with it in a less invasive manner.

Impact of Climate Change on Sea Level Rise along the Coastline of Mumbai City, India

Sea-level rise being one of the most important impacts of anthropogenic induced climate change resulting from global warming and melting of icebergs at Arctic and Antarctic, the investigations done by various researchers both on Indian Coast and elsewhere during the last decade has been reviewed in this paper. The paper aims to ascertain the propensity of consistency of different suggested methods to predict the near-accurate future sea level rise along the coast of Mumbai. Case studies at East Coast, Southern Tip and West and South West coast of India have been reviewed. Coastal Vulnerability Index of several important international places has been compared, which matched with Intergovernmental Panel on Climate Change forecasts. The application of Geographic Information System mapping, use of remote sensing technology, both Multi Spectral Scanner and Thematic Mapping data from Landsat classified through Iterative Self-Organizing Data Analysis Technique for arriving at high, moderate and low Coastal Vulnerability Index at various important coastal cities have been observed. Instead of data driven, hindcast based forecast for Significant Wave Height, additional impact of sea level rise has been suggested. Efficacy and limitations of numerical methods vis-à-vis Artificial Neural Network has been assessed, importance of Root Mean Square error on numerical results is mentioned. Comparing between various computerized methods on forecast results obtained from MIKE 21 has been opined to be more reliable than Delft 3D model.

A Comparison of YOLO Family for Apple Detection and Counting in Orchards

In agricultural production and breeding, implementing automatic picking robot in orchard farming to reduce human labour and error is challenging. The core function of it is automatic identification based on machine vision. This paper focuses on apple detection and counting in orchards and implements several deep learning methods. Extensive datasets are used and a semi-automatic annotation method is proposed. The proposed deep learning models are in state-of-the-art YOLO family. In view of the essence of the models with various backbones, a multi-dimensional comparison in details is made in terms of counting accuracy, mAP and model memory, laying the foundation for realising automatic precision agriculture.

Improving Subjective Bias Detection Using Bidirectional Encoder Representations from Transformers and Bidirectional Long Short-Term Memory

Detecting subjectively biased statements is a vital task. This is because this kind of bias, when present in the text or other forms of information dissemination media such as news, social media, scientific texts, and encyclopedias, can weaken trust in the information and stir conflicts amongst consumers. Subjective bias detection is also critical for many Natural Language Processing (NLP) tasks like sentiment analysis, opinion identification, and bias neutralization. Having a system that can adequately detect subjectivity in text will boost research in the above-mentioned areas significantly. It can also come in handy for platforms like Wikipedia, where the use of neutral language is of importance. The goal of this work is to identify the subjectively biased language in text on a sentence level. With machine learning, we can solve complex AI problems, making it a good fit for the problem of subjective bias detection. A key step in this approach is to train a classifier based on BERT (Bidirectional Encoder Representations from Transformers) as upstream model. BERT by itself can be used as a classifier; however, in this study, we use BERT as data preprocessor as well as an embedding generator for a Bi-LSTM (Bidirectional Long Short-Term Memory) network incorporated with attention mechanism. This approach produces a deeper and better classifier. We evaluate the effectiveness of our model using the Wiki Neutrality Corpus (WNC), which was compiled from Wikipedia edits that removed various biased instances from sentences as a benchmark dataset, with which we also compare our model to existing approaches. Experimental analysis indicates an improved performance, as our model achieved state-of-the-art accuracy in detecting subjective bias. This study focuses on the English language, but the model can be fine-tuned to accommodate other languages.

Bayesian Geostatistical Modelling of COVID-19 Datasets

The COVID-19 dataset is obtained by extracting weather, longitude, latitude, ISO3666, cases and death of coronavirus patients across the globe. The data were extracted for a period of eight day choosing uniform time within the specified period. Then mapping of cases and deaths with reverence to continents were obtained. Bayesian Geostastical modelling was carried out on the dataset. The study found out that countries in the tropical region suffered less deaths/attacks compared to countries in the temperate region, this is due to high temperature in the tropical region.

Transformation of Industrial Policy towards Industry 4.0 and Its Impact on Firms' Competition

Although Europe is on the threshold of a new industrial revolution called Industry 4.0, many believe that this will increase the flexibility of production, the mass adaptation of products to consumers and the speed of their service; it will also improve product quality and dramatically increase productivity. However, as expected, all the benefits of Industry 4.0 face many of the inevitable changes and challenges they pose. One of them is the inevitable transformation of current competition and business models. This article examines the possible results of competitive conversion from the classic Bertrand and Cournot models to qualitatively new competition based on innovation. Ability to deliver a new product quickly and the possibility to produce the individual design (through flexible and quickly configurable factories) by reducing equipment failures and increasing process automation and control is highly important. This study shows that the ongoing transformation of the competition model is changing the game. This, together with the creation of complex value networks, means huge investments that make it particularly difficult for small and medium-sized enterprises. In addition, the ongoing digitalization of data raises new concerns regarding legal obligations, intellectual property, and security.

Modified Genome-Scale Metabolic Model of Escherichia coli by Adding Hyaluronic Acid Biosynthesis-Related Enzymes (GLMU2 and HYAD) from Pasteurella multocida

Hyaluronic acid (HA) consists of linear heteropolysaccharides repeat of D-glucuronic acid and N-acetyl-D-glucosamine. HA has various useful properties to maintain skin elasticity and moisture, reduce inflammation, and lubricate the movement of various body parts without causing immunogenic allergy. HA can be found in several animal tissues as well as in the capsule component of some bacteria including Pasteurella multocida. This study aimed to modify a genome-scale metabolic model of Escherichia coli using computational simulation and flux analysis methods to predict HA productivity under different carbon sources and nitrogen supplement by the addition of two enzymes (GLMU2 and HYAD) from P. multocida to improve the HA production under the specified amount of carbon sources and nitrogen supplements. Result revealed that threonine and aspartate supplement raised the HA production by 12.186%. Our analyses proposed the genome-scale metabolic model is useful for improving the HA production and narrows the number of conditions to be tested further.

The SOCI Strategy as a Method to Meet the Innovation Challenges of COVID-19

The COVID-19 causes a worldwide crisis and has an impact in every dimension of the economy. Organizations with the ability to adapt to new developments and which innovate solutions for the disrupted world during and after the Corona crises have the opportunity to not only survive the crisis but rather to use new trends to implement new business models and gain advantage. In this context, startups seem to have better opportunities to manage the Corona crisis through their innovation-based nature. The main result of this paper is the understanding that by applying a startup orientated innovation (SOCI) strategy, established companies can be motivated to meet the challenge of COVID-19 in a similar way like startups. This result can be achieved by describing the role of innovation and a SOCI strategy as helpful methods for organizations to meet the coming challenges during and after the COVID-19 epidemics. In addition to this, this paper presents a practical application of SOCI through the PANDA approach of the Fresenius University of Applied Sciences in Germany and discuss it in the context of COVID-19 as an exemplary successful real-world implementation of SOCI strategy.

Investigation of the Effects of Biodiesel Blend on Particulate-Phase Exhaust Emissions from a Light Duty Diesel Vehicle

This study presents an investigation of diesel vehicle particulate-phase emissions with neat ultralow sulphur diesel (B0, ULSD) and 5% waste cooking oil-based biodiesel blend (B5) in Hong Kong. A Euro VI light duty diesel vehicle was tested under transient (New European Driving Cycle (NEDC)), steady-state and idling on a chassis dynamometer. Chemical analyses including organic carbon (OC), elemental carbon (EC), as well as 30 polycyclic aromatic hydrocarbons (PAHs) and 10 oxygenated PAHs (oxy-PAHs) were conducted. The OC fuel-based emission factors (EFs) for B0 ranged from 2.86 ± 0.33 to 7.19 ± 1.51 mg/kg, and those for B5 ranged from 4.31 ± 0.64 to 15.36 ± 3.77 mg/kg, respectively. The EFs of EC were low for both fuel blends (0.25 mg/kg or below). With B5, the EFs of total PAHs were decreased as compared to B0. Specifically, B5 reduced total PAH emissions by 50.2%, 30.7%, and 15.2% over NEDC, steady-state and idling, respectively. It was found that when B5 was used, PAHs and oxy-PAHs with lower molecular weight (2 to 3 rings) were reduced whereas PAHs/oxy-PAHs with medium or high molecular weight (4 to 7 rings) were increased. Our study suggests the necessity of taking atmospheric and health factors into account for biodiesel application as an alternative motor fuel.

The Reproducibility and Repeatability of Modified Likelihood Ratio for Forensics Handwriting Examination

The forensic use of handwriting depends on the analysis, comparison, and evaluation decisions made by forensic document examiners. When using biometric technology in forensic applications, it is necessary to compute Likelihood Ratio (LR) for quantifying strength of evidence under two competing hypotheses, namely the prosecution and the defense hypotheses wherein a set of assumptions and methods for a given data set will be made. It is therefore important to know how repeatable and reproducible our estimated LR is. This paper evaluated the accuracy and reproducibility of examiners' decisions. Confidence interval for the estimated LR were presented so as not get an incorrect estimate that will be used to deliver wrong judgment in the court of Law. The estimate of LR is fundamentally a Bayesian concept and we used two LR estimators, namely Logistic Regression (LoR) and Kernel Density Estimator (KDE) for this paper. The repeatability evaluation was carried out by retesting the initial experiment after an interval of six months to observe whether examiners would repeat their decisions for the estimated LR. The experimental results, which are based on handwriting dataset, show that LR has different confidence intervals which therefore implies that LR cannot be estimated with the same certainty everywhere. Though the LoR performed better than the KDE when tested using the same dataset, the two LR estimators investigated showed a consistent region in which LR value can be estimated confidently. These two findings advance our understanding of LR when used in computing the strength of evidence in handwriting using forensics.

An Effort at Improving Reliability of Laboratory Data in Titrimetric Analysis for Zinc Sulphate Tablets Using Validated Spreadsheet Calculators

The requirement for maintaining data integrity in laboratory operations is critical for regulatory compliance. Automation of procedures reduces incidence of human errors. Quality control laboratories located in low-income economies may face some barriers in attempts to automate their processes. Since data from quality control tests on pharmaceutical products are used in making regulatory decisions, it is important that laboratory reports are accurate and reliable. Zinc Sulphate (ZnSO4) tablets is used in treatment of diarrhea in pediatric population, and as an adjunct therapy for COVID-19 regimen. Unfortunately, zinc content in these formulations is determined titrimetrically; a manual analytical procedure. The assay for ZnSO4 tablets involves time-consuming steps that contain mathematical formulae prone to calculation errors. To achieve consistency, save costs, and improve data integrity, validated spreadsheets were developed to simplify the two critical steps in the analysis of ZnSO4 tablets: standardization of 0.1M Sodium Edetate (EDTA) solution, and the complexometric titration assay procedure. The assay method in the United States Pharmacopoeia was used to create a process flow for ZnSO4 tablets. For each step in the process, different formulae were input into two spreadsheets to automate calculations. Further checks were created within the automated system to ensure validity of replicate analysis in titrimetric procedures. Validations were conducted using five data sets of manually computed assay results. The acceptance criteria set for the protocol were met. Significant p-values (p < 0.05, α = 0.05, at 95% Confidence Interval) were obtained from students’ t-test evaluation of the mean values for manual-calculated and spreadsheet results at all levels of the analysis flow. Right-first-time analysis and principles of data integrity were enhanced by use of the validated spreadsheet calculators in titrimetric evaluations of ZnSO4 tablets. Human errors were minimized in calculations when procedures were automated in quality control laboratories. The assay procedure for the formulation was achieved in a time-efficient manner with greater level of accuracy. This project is expected to promote cost savings for laboratory business models.

Management Prospects of Winery By-Products Based on Phenolic Compounds and Antioxidant Activity of Grape Skins: The Case of Greek Ionian Islands

The aim of this work was to recover phenolic compounds from grape skins produced in Greek varieties of the Ionian Islands in order to form the basis of calculations for their further utilization in the context of the circular economy. Isolation and further utilization of phenolic compounds is an important issue in winery by-products. For this purpose, 37 samples were collected, extracted, and analyzed in an attempt to provide the appropriate basis for their sustainable exploitation. Extraction of the bioactive compounds was held using an eco-friendly, non-toxic, and highly effective water-glycerol solvent system. Then, extracts were analyzed using UV-Vis, liquid chromatography-mass spectrometry (LC-MS), FTIR, and Raman spectroscopy. Also, total phenolic content and antioxidant activity were measured. LC-MS chromatography showed qualitative differences between different varieties. Peaks were attributed to monomeric 3-flavanols as well as monomeric, dimeric, and trimeric proanthocyanidins. The FT-IR and Raman spectra agreed with the chromatographic data and contributed to identifying phenolic compounds. Grape skins exhibited high total phenolic content (TPC), and it was proved that during vinification, a large number of polyphenols remained in the pomace. This study confirmed that grape skins from Ionian Islands are a promising source of bioactive compounds, suggesting their utilization under a bio-economic and environmental strategic framework.

An Exploratory Study on the Difference between Online and Offline Conformity Behavior among Chinese College Students

Conformity is defined as one in a social group changing his or her behavior to match the others’ behavior in the group. It is used to find that people show a higher level of online conformity behavior than offline. However, as anonymity can decrease the level of online conformity behavior, the difference between online and offline conformity behavior among Chinese college students still needs to be tested. In this study, college students (N = 60) have been randomly assigned into three groups: control group, offline experimental group, and online experimental group. Through comparing the results of offline experimental group and online experimental group with the Mann-Whitney U test, this study verified the results of Asch’s experiment, and found out that people show a lower level of online conformity behavior than offline, which contradicted the previous finding found in China. These results can be used to explain why some people make a lot of vicious remarks and radical ideas on the Internet but perform normally in their real life: the anonymity of the network makes the online group pressure less than offline, so people are less likely to conform to social norms and public opinions on the Internet. What is more, these results support the importance and relevance of online voting, because fewer online group pressures make it easier for people to expose their true ideas, thus gathering more comprehensive and truthful views and opinions.

Geometrical Based Unequal Droplet Splitting Using Microfluidic Y-Junction

Among different droplet manipulations, controlled droplet-splitting is of great significance due to its ability to increase throughput and operational capability. Furthermore, unequal droplet-splitting can provide greater flexibility and a wider range of dilution factors. In this study, we developed two-dimensional, time-dependent complex fluid dynamics simulations to model droplet formation in a flow focusing device, followed by splitting in a Y-shaped junction with sub-channels of unequal widths. From the results obtained from the numerical study, we correlated the diameters of the droplets in the sub-channels to the Weber number, thereby demarcating the droplet splitting and non-splitting regimes.

Users’ Information Disclosure Determinants in Social Networking Sites: A Systematic Literature Review

The privacy paradox describes a phenomenon whereby there is no connection between stated privacy concerns and privacy behaviours. We need to understand the underlying reasons for this paradox if we are to help users to preserve their privacy more effectively. In particular, the Social Networking System (SNS) domain offers a rich area of investigation due to the risks of unwise information disclosure decisions. Our study thus aims to untangle the complicated nature and underlying mechanisms of online privacy-related decisions in SNSs. In this paper, we report on the findings of a Systematic Literature Review (SLR) that revealed a number of factors that are likely to influence online privacy decisions. Our deductive analysis approach was informed by Communicative Privacy Management (CPM) theory. We uncovered a lack of clarity around privacy attitudes and their link to behaviours, which makes it challenging to design privacy-protecting SNS platforms and to craft legislation to ensure that users’ privacy is preserved.

Using Analytical Hierarchy Process and TOPSIS Approaches in Designing a Finite Element Analysis Automation Program

Sophisticated numerical simulations like finite element analysis (FEA) involve a complicated process from model setup to post-processing tasks that require replication of time-consuming steps. Utilizing FEA automation program simplifies the complexity of the involved steps while minimizing human errors in analysis set up, calculations, and results processing. One of the main challenges in designing FEA automation programs is to identify user requirements and link them to possible design alternatives. This paper presents a decision-making framework to design a Python based FEA automation program for modal analysis, frequency response analysis, and random vibration fatigue (RVF) analysis procedures. Analytical hierarchy process (AHP) and technique for order preference by similarity to ideal solution (TOPSIS) are applied to evaluate design alternatives considering the feedback received from experts and program users.