Speech Intelligibility Improvement Using Variable Level Decomposition DWT

Intelligibility is an essential characteristic of a speech signal, which is used to help in the understanding of information in speech signal. Background noise in the environment can deteriorate the intelligibility of a recorded speech. In this paper, we presented a simple variance subtracted - variable level discrete wavelet transform, which improve the intelligibility of speech. The proposed algorithm does not require an explicit estimation of noise, i.e., prior knowledge of the noise; hence, it is easy to implement, and it reduces the computational burden. The proposed algorithm decides a separate decomposition level for each frame based on signal dominant and dominant noise criteria. The performance of the proposed algorithm is evaluated with speech intelligibility measure (STOI), and results obtained are compared with Universal Discrete Wavelet Transform (DWT) thresholding and Minimum Mean Square Error (MMSE) methods. The experimental results revealed that the proposed scheme outperformed competing methods

Author Profiling: Prediction of Learners’ Gender on a MOOC Platform Based on Learners’ Comments

The more an educational system knows about a learner, the more personalised interaction it can provide, which leads to better learning. However, asking a learner directly is potentially disruptive, and often ignored by learners. Especially in the booming realm of MOOC Massive Online Learning platforms, only a very low percentage of users disclose demographic information about themselves. Thus, in this paper, we aim to predict learners’ demographic characteristics, by proposing an approach using linguistically motivated Deep Learning Architectures for Learner Profiling, particularly targeting gender prediction on a FutureLearn MOOC platform. Additionally, we tackle here the difficult problem of predicting the gender of learners based on their comments only – which are often available across MOOCs. The most common current approaches to text classification use the Long Short-Term Memory (LSTM) model, considering sentences as sequences. However, human language also has structures. In this research, rather than considering sentences as plain sequences, we hypothesise that higher semantic - and syntactic level sentence processing based on linguistics will render a richer representation. We thus evaluate, the traditional LSTM versus other bleeding edge models, which take into account syntactic structure, such as tree-structured LSTM, Stack-augmented Parser-Interpreter Neural Network (SPINN) and the Structure-Aware Tag Augmented model (SATA). Additionally, we explore using different word-level encoding functions. We have implemented these methods on Our MOOC dataset, which is the most performant one comparing with a public dataset on sentiment analysis that is further used as a cross-examining for the models' results.

Propane Dehydrogenation over Pt-Sn Supported on Magnesium Aluminate Material

Pt-Sn catalysts have been prepared using magnesium aluminate as a support with two different Mg/Al ratio. The supports/catalysts have been characterized by N2-adsorption, XRD, and temperature programmed desorption of NH3 and thermogravimetry analysis (TGA). The catalysts have been evaluated at 595 0C for the propane dehydrogenation reaction at 0.5 barg pressure using a feed containing pure propane with steam to hydrocarbon ratio of 1 mol/mol and weight hourly space velocity (WHSV) 0.9 h-1. Chlorine quantification studies have been developed using Carbon-Hydrogen-Nitrogen-Sulphur (CHNS) analyzer. The dechlorinated catalyst with higher alumina content showed better performance (38-43% propane conversion, 91-94% propylene selectivity) in propane conversion and propylene selectivity than Pt-Sn-MG-AL-DC-1 (30-18% propane conversion, 83-90% propylene selectivity).

Providing a Secure, Reliable and Decentralized Document Management Solution Using Blockchain by a Virtual Identity Card

In today's world, we need documents everywhere for a smooth workflow in the identification process or any other security aspects. The current system and techniques which are used for identification need one thing, that is ‘proof of existence’, which involves valid documents, for example, educational, financial, etc. The main issue with the current identity access management system and digital identification process is that the system is centralized in their network, which makes it inefficient. The paper presents the system which resolves all these cited issues. It is based on ‘blockchain’ technology, which is a 'decentralized system'. It allows transactions in a decentralized and immutable manner. The primary notion of the model is to ‘have everything with nothing’. It involves inter-linking required documents of a person with a single identity card so that a person can go anywhere without having the required documents with him/her. The person just needs to be physically present at a place wherein documents are necessary, and using a fingerprint impression and an iris scan print, the rest of the verification will progress. Furthermore, some technical overheads and advancements are listed. This paper also aims to layout its far-vision scenario of blockchain and its impact on future trends.

A Framework for Improving Trade Contractors’ Productivity Tracking Methods

Despite being one of the most significant economic contributors of the country, Canada’s construction industry is lagging behind other sectors when it comes to labor productivity improvements. The construction industry is very collaborative as a general contractor, will hire trade contractors to perform most of a project’s work; meaning low productivity from one contractor can have a domino effect on the shared success of a project. To address this issue and encourage trade contractors to improve their productivity tracking methods, an investigative study was done on the productivity views and tracking methods of various trade contractors. Additionally, an in-depth review was done on four standard tracking methods used in the construction industry: cost codes, benchmarking, the job productivity measurement (JPM) standard, and WorkFace Planning (WFP). The four tracking methods were used as a baseline in comparing the trade contractors’ responses, determining gaps within their current tracking methods, and for making improvement recommendations. 15 interviews were conducted with different trades to analyze how contractors value productivity. The results of these analyses indicated that there seem to be gaps within the construction industry when it comes to an understanding of the purpose and value in productivity tracking. The trade contractors also shared their current productivity tracking systems; which were then compared to the four standard tracking methods used in the construction industry. Gaps were identified in their various tracking methods and using a framework; recommendations were made based on the type of trade on how to improve how they track productivity.

Effect of Bentonite on Shear Strength of Bushehr Calcareous Sand

Calcareous sands are found most commonly in areas adjacent to crude oil and gas, and particularly around water. These types of soil have high compressibility due to high inter-granular porosity, irregularity, fragility, and especially crushing. Also, based on experience, it has been shown that the behavior of these types of soil is not similar to silica sand in loading. Since the destructive effects of cement on the environment are obvious, other alternatives such as bentonite are popular to be used. Bentonite has always been used commercially in civil engineering projects and according to its low hydraulic conductivity, it is used for landfills, cut-off walls, and nuclear wastelands. In the present study, unconfined compression tests in five ageing periods (1, 3, 7, 14, and 28 days) after mixing different percentages of bentonite (5%, 7.5% and 10%) with Bushehr calcareous sand were performed. The relative density considered for the specimens is 50%. Optimum water content was then added to each specimen accordingly (19%, 18.5%, and 17.5%). The sample preparation method was wet tamping and the specimens were compacted in five layers. It can be concluded from the results that as the bentonite content increases, the unconfined compression strength of the soil increases. Based on the obtained results, 3-day and 7-day ageing periods showed 30% and 50% increase in the shear strength of soil, respectively.

Optimization of Surface Finish in Milling Operation Using Live Tooling via Taguchi Method

The main objective of this research is to optimize the surface roughness of a milling operation on AISI 1018 steel using live tooling on a HAAS ST-20 lathe. In this study, Taguchi analysis is used to optimize the milling process by investigating the effect of different machining parameters on surface roughness. The L9 orthogonal array is designed with four controllable factors with three different levels each and an uncontrollable factor, resulting in 18 experimental runs. The optimal parameters determined from Taguchi analysis were feed rate – 76.2 mm/min, spindle speed 1150 rpm, depth of cut – 0.762 mm and 2-flute TiN coated high-speed steel as tool material. The process capability Cp and process capability index Cpk values were improved from 0.62 and -0.44 to 1.39 and 1.24 respectively. The average surface roughness values from the confirmation runs were 1.30 µ, decreasing the defect rate from 87.72% to 0.01%. The purpose of this study is to efficiently utilize the Taguchi design to optimize the surface roughness in a milling operation using live tooling.

Learning Programming for Hearing Impaired Students via an Avatar

Deaf and hearing-impaired students face many obstacles throughout their education, especially with learning applied sciences such as computer programming. In addition, there is no clear signs in the Arabic Sign Language that can be used to identify programming logic terminologies such as while, for, case, switch etc. However, hearing disabilities should not be a barrier for studying purpose nowadays, especially with the rapid growth in educational technology. In this paper, we develop an Avatar based system to teach computer programming to deaf and hearing-impaired students using Arabic Signed language with new signs vocabulary that is been developed for computer programming education. The system is tested on a number of high school students and results showed the importance of visualization in increasing the comprehension or understanding of concepts for deaf students through the avatar.

Motivational Antecedents that Influenced a Higher Education Institution in the Philippines to Adopt Enterprise Architecture

Technology is a recent prodigy in people’s everyday life that has taken off. It infiltrated almost every aspect of one’s lives, changing how people work, how people learn and how people perceive things. Academic Institutions, just like other organizations, have deeply modified its strategies to integrate technology into the institutional vision and corporate strategy that has never been greater. Information and Communications Technology (ICT) continues to be recognized as a major factor in organizations realizing its aims and objectives. Consequently, ICT has an important role in the mobilization of an academic institution’s strategy to support the delivery of operational, strategic or transformational objectives. This ICT strategy should align the institution with the radical changes of the ICT world through the use of Enterprise Architecture (EA). Hence, EA’s objective is to optimize the islands of legacy processes to be integrated that is receptive to change and supportive of the delivery of the strategy. In this paper, the focus is to explore the motivational antecedents during the adoption of EA in a Higher Education Institution in the Philippines for its ICT strategic plan. The seven antecedents (viewpoint, stakeholders, human traits, vision, revolutionary innovation, techniques and change components) provide understanding into EA adoption and the antecedents that influences the process of EA adoption.

Monitoring Blood Pressure Using Regression Techniques

Blood pressure helps the physicians greatly to have a deep insight into the cardiovascular system. The determination of individual blood pressure is a standard clinical procedure considered for cardiovascular system problems. The conventional techniques to measure blood pressure (e.g. cuff method) allows a limited number of readings for a certain period (e.g. every 5-10 minutes). Additionally, these systems cause turbulence to blood flow; impeding continuous blood pressure monitoring, especially in emergency cases or critically ill persons. In this paper, the most important statistical features in the photoplethysmogram (PPG) signals were extracted to estimate the blood pressure noninvasively. PPG signals from more than 40 subjects were measured and analyzed and 12 features were extracted. The features were fed to principal component analysis (PCA) to find the most important independent features that have the highest correlation with blood pressure. The results show that the stiffness index means and standard deviation for the beat-to-beat heart rate were the most important features. A model representing both features for Systolic Blood Pressure (SBP) and Diastolic Blood Pressure (DBP) was obtained using a statistical regression technique. Surface fitting is used to best fit the series of data and the results show that the error value in estimating the SBP is 4.95% and in estimating the DBP is 3.99%.

Chemical Leaching of Metals from Landfill’s Fine Fraction

Leaching of heavy metals (chromium, zinc, copper) from the fine fraction of the Torma landfill (Estonia) was investigated. The leaching kinetics studies have determined the dependence of some metal’s concentration on the leaching time. Metals were leached with Aqua Regia, distilled water and EDTA (Ethylenediaminetetraacetic acid); process was most intensive 2 hours after the start of the experiment, except for copper with EDTA (0.5 h) and lead with EDTA (4 h). During leaching, steady concentrations of Fe, Mn, Cd and Pb were fully stabilized after 8 h; however concentrations of Cu and Ni were not stabilized after 10 h.

Human Factors Considerations in New Generation Fighter Planes to Enhance Combat Effectiveness

Role of fighter planes in modern network centric military warfare scenarios has changed significantly in the recent past. New generation fighter planes have multirole capability of engaging both air and ground targets with high precision. Multirole aircraft undertakes missions such as Air to Air combat, Air defense, Air to Surface role (including Air interdiction, Close air support, Maritime attack, Suppression and Destruction of enemy air defense), Reconnaissance, Electronic warfare missions, etc. Designers have primarily focused on development of technologies to enhance the combat performance of the fighter planes and very little attention is given to human factor aspects of technologies. Unique physical and psychological challenges are imposed on the pilots to meet operational requirements during these missions. Newly evolved technologies have enhanced aircraft performance in terms of its speed, firepower, stealth, electronic warfare, situational awareness, and vulnerability reduction capabilities. This paper highlights the impact of emerging technologies on human factors for various military operations and missions. Technologies such as ‘cooperative knowledge-based systems’ to aid pilot’s decision making in military conflict scenarios as well as simulation technologies to enhance human performance is also studied as a part of research work. Current and emerging pilot protection technologies and systems which form part of the integrated life support systems in new generation fighter planes is discussed. System safety analysis application to quantify the human reliability in military operations is also studied.

Collaborative Reflexive/Reflective Teaching and Action Research in TESL

Teaching English as a Second Language (TESL) has become a very rich area of research. Practitioners or teachers of English as a foreign or a second language are now promoting both collaborative learning and collaborative teaching. Students learning a language collaboratively and cooperatively are learning in a better environment of team work where they learn from each other. Further, teaching English collaboratively also creates an enriching environment that is also very enriching to students’ and teachers’ experiences of learning and teaching. Moreover, action research stems from actual teacher concerns and students’ needs. Reflection in turn, on the experience of the material taught and the delivery of material is becoming an integral part of the teaching and learning experience self- evaluation and self-development. In this case, the concern of the research field in the area of TESL will be the development of teaching delivery, material and quality of learning. In the present research, the TESL module taught to year two students in the Faculty of Arts and Humanities, British University in Egypt (BUE) will be evaluated reflexively by the students and teachers. The module was taught to students in two different specialisms. It was taught and delivered through collaborative teaching and was evaluated by both teachers and students as very successful and enjoyable. The reflections of both teachers and students as well as student results confirm that it was a success.

SVID: Structured Vulnerability Intelligence for Building Deliberated Vulnerable Environment

The diversity and complexity of modern IT systems make it almost impossible for internal teams to find vulnerabilities in all software before the software is officially released. The emergence of threat intelligence and vulnerability reporting policy has greatly reduced the burden on software vendors and organizations to find vulnerabilities. However, to prove the existence of the reported vulnerability, it is necessary but difficult for security incident response team to build a deliberated vulnerable environment from the vulnerability report with limited and incomplete information. This paper presents a structured, standardized, machine-oriented vulnerability intelligence format, that can be used to automate the orchestration of Deliberated Vulnerable Environment (DVE). This paper highlights the important role of software configuration and proof of vulnerable specifications in vulnerability intelligence, and proposes a triad model, which is called DIR (Dependency Configuration, Installation Configuration, Runtime Configuration), to define software configuration. Finally, this paper has also implemented a prototype system to demonstrate that the orchestration of DVE can be automated with the intelligence.

Advanced Materials Based on Ethylene-Propylene-Diene Terpolymers and Organically Modified Montmorillonite

This paper presents studies on the development and characterization of nanocomposites based on ethylene-propylene terpolymer rubber (EPDM), chlorobutyl rubber (IIR-Cl) and organically modified montmorillonite (OMMT). Mixtures were made containing 0, 3 and 6 phr (parts per 100 parts rubber) OMMT, respectively. They were obtained by melt intercalation in an internal mixer - Plasti-Corder Brabender, in suitable blending parameters, at high temperature for 11 minutes. Curing agents were embedded on a laboratory roller at 70-100 ºC, friction 1:1.1, processing time 5 minutes. Rubber specimens were obtained by compression, using a hydraulic press at 165 ºC and a pressing force of 300 kN. Curing time, determined using the Monsanto rheometer, decreases with the increased amount of OMMT in the mixtures. At the same time, it was noticed that mixtures containing OMMT show improvement in physical-mechanical properties. These types of nanocomposites may be used to obtain rubber seals for the space application or for other areas of application.

CybeRisk Management in Banks: An Italian Case Study

The financial sector is exposed to the risk of cyber-attacks like any other industrial sector. Furthermore, the topic of CybeRisk (cyber risk) has become particularly relevant given that Information Technology (IT) attacks have increased drastically in recent years, and cannot be stopped by single organizations requiring a response at international and national level. IT risk is never a matter purely for the IT manager, although he clearly plays a key role. A bank's risk management function requires a thorough understanding of the evolving risks as well as the tools and practical techniques available to address them. Upon the request of European and national legislation regarding CybeRisk in the financial system, banks are therefore called upon to strengthen the operational model for CybeRisk management. This will require an important change with a more intense collaboration with the structures that deal with information security for the development of an ad hoc system for the evaluation and control of this type of risk. The aim of the work is to propose a framework for the management and control of CybeRisk that will bridge the gap in the literature regarding the understanding and consideration of CybeRisk as an integral part of business management. The IT function has a strong relevance in the management of CybeRisk, which is perceived mainly as operational risk, but with a positive tendency on the part of risk management to the identification of CybeRisk assessment methods that are increasingly complete, quantitative and able to better describe the possible impacts on the business. The paper provides answers to the research questions: Is it possible to define a CybeRisk governance structure able to support the comparison between risk and security? How can the relationships between IT assets be integrated into a cyberisk assessment framework to guarantee a system of protection and risks control? From a methodological point of view, this research uses a case study approach. The choice of “Monte dei Paschi di Siena” was determined by the specific features of one of Italy’s biggest lenders. It is chosen to use an intensive research strategy: an in-depth study of reality. The case study methodology is an empirical approach to explore a complex and current phenomenon that develops over time. The use of cases has also the advantage of allowing the deepening of aspects concerning the "how" and "why" of contemporary events, on which the scholar has little control. The research bases on quantitative data and qualitative information obtained through semi-structured interviews of an open-ended nature and questionnaires to directors, members of the audit committee, risk, IT and compliance managers, and those responsible for internal audit function and anti-money laundering. The added value of the paper can be seen in the development of a framework based on a mapping of IT assets from which it is possible to identify their relationships for purposes of a more effective management and control of cyber risk.

Experimental Investigation on Geosynthetic-Reinforced Soil Sections via California Bearing Ratio Test

Loose soils normally are of weak bearing capacity due to their structural nature. Being exposed to heavy traffic loads, they would fail in most cases. To tackle the aforementioned issue, geotechnical engineers have come up with different approaches; one of which is making use of geosynthetic-reinforced soil-aggregate systems. As these polymeric reinforcements have highlighted economic and environmentally-friendly features, they have become widespread in practice during the last decades. The present research investigates the efficiency of four different types of these reinforcements in increasing the bearing capacity of two-layered soil sections using a series California Bearing Ratio (CBR) test. The studied sections are comprised of a 10 cm-thick layer of no. 161 Firouzkooh sand (weak subgrade) and a 10 cm-thick layer of compacted aggregate materials (base course) classified as SP and GW according to the United Soil Classification System (USCS), respectively. The aggregate layer was compacted to the relative density (Dr) of 95% at the optimum water content (Wopt) of 6.5%. The applied reinforcements were including two kinds of geocomposites (type A and B), a geotextile, and a geogrid that were embedded at the interface of the lower and the upper layers of the soil-aggregate system. As the standard CBR mold was not appropriate in height for this study, the mold used for soaked CBR tests were utilized. To make a comparison between the results of stress-settlement behavior in the studied specimens, CBR values pertinent to the penetrations of 2.5 mm and 5 mm were considered. The obtained results demonstrated 21% and 24.5% increments in the amount of CBR value in the presence of geocomposite type A and geogrid, respectively. On the other hand, the effect of both geotextile and geocomposite type B on CBR values was generally insignificant in this research.

Problems of Boolean Reasoning Based Biclustering Parallelization

Biclustering is the way of two-dimensional data analysis. For several years it became possible to express such issue in terms of Boolean reasoning, for processing continuous, discrete and binary data. The mathematical backgrounds of such approach — proved ability of induction of exact and inclusion–maximal biclusters fulfilling assumed criteria — are strong advantages of the method. Unfortunately, the core of the method has quite high computational complexity. In the paper the basics of Boolean reasoning approach for biclustering are presented. In such context the problems of computation parallelization are risen.

Lattice Boltzmann Simulation of MHD Natural Convection Heat Transfer of Cu-Water Nanofluid in a Linearly/Sinusoidally Heated Cavity

In this numerical study, natural convection of Cu–water nanofluid in a cavity submitted to different heating modes on its vertical walls is analyzed. Maxwell-Garnetts (MG) and Brinkman models have been utilized for calculating the effective thermal conductivity and dynamic viscosity of nanofluid, respectively. Influences of Rayleigh number (Ra = 103−106), nanoparticle volume concentration (f = 0-0.04) and Hartmann number (Ha = 0-90) on the flow and heat transfer characteristics have been examined. The results indicate that the Hartmann number influences the heat transfer at Ra = 106 more than other Raleigh numbers, as the least effect is observed at Ra = 103. Moreover, the results show that the solid volume fraction has a significant influence on heat transfer, depending on the value of Hartmann, heat generation or absorption coefficient and Rayleigh numbers.

Mechanical Properties of Cement Slurry by Partially Substitution of Industry Waste Natural Pozzolans

There have been many reports of the destructive effects of cement on the environment in recent years. In the present research, it has been attempted to reduce the destructive effects of cement by replacing silica fume as adhesive materials instead of cement. The present study has attempted to improve the mechanical properties of cement slurry by using waste material from a glass production factory, located in Qazvin city of Iran, in which accumulation volume has become an environmental threat. The chemical analysis of the waste material indicates that this material contains about 94% of SiO2 and AL2O3 and has a close structure to silica fume. Also, the particle grain size test was performed on the mentioned waste. Then, the unconfined compressive strength test of the slurry was performed by preparing a mixture of water and adhesives with different percentages of cement and silica fume. The water to an adhesive ratio of this mixture is 1:3, and the curing process last 28 days. It was found that the sample had an unconfined compressive strength of about 300 kg/cm2 in a mixture with equal proportions of cement and silica fume. Besides, the sample had a brittle fracture in the slurry sample made of pure cement, however, the fracture in cement-silica fume slurry mixture is flexible and the structure of the specimen remains coherent after fracture. Therefore, considering the flexibility that is achieved by replacing this waste, it can be used to stabilize soils with cracking potential.