Leveraging Hyperledger Iroha for the Issuance and Verification of Higher-Education Certificates

Higher Education is resisting the pull of technology, especially as this concerns the issuance and verification of degrees and certificates. It is widely known that education certificates are largely produced in paper form making them vulnerable to damage while holders of such certificates are dependent on the universities and other issuing organisations. QualiChain is an EU Horizon 2020 (H2020) research project aiming to transform and revolutionise the domain of public education and its ties with the job market by leveraging blockchain, analytics and decision support to develop a platform for the verification and sharing of education certificates. Blockchain plays an integral part in the QualiChain solution in providing a trustworthy environment to store, share and manage such accreditations. Under the context of this paper, three prominent blockchain platforms (Ethereum, Hyperledger Fabric, Hyperledger Iroha) were considered as a means of experimentation for creating a system with the basic functionalities that will be needed for trustworthy degree verification. The methodology and respective system developed and presented in this paper used Hyperledger Iroha and proved that this specific platform can be used to easily develop decentralize applications. Future papers will attempt to further experiment with other blockchain platforms and assess which has the best potential.

Shifting Paradigms of Culture: Rise of Secular Sensibility in Indian Literature

Burgeoning demand of ‘Secularism’ has shaken the pillars of cultural studies in the contemporary literature. The perplexity of the culturally estranged term ‘secular’ gives rise to temporal ideologies across the world. Hence, it is high time to scan this concept in the context of Indian lifestyle which is a blend of assimilated cultures woven in multiple religious fabrics. The infliction of such secular taste is depicted in literary productions like ‘Satanic Verses’ and ‘An Area of Darkness’. The paper conceptually makes a cross-cultural analysis of anti-religious Indian literary texts, assessing its revitalization in current times. Further, this paper studies the increasing popularity of secular sensibility in the contemporary times. The mushrooming elements of secularism such as abstraction, spirituality, liberation, individualism give rise to a seemingly newer idea i.e. ‘Plurality’ making the literature highly hybrid. This approach has been used to study Indian modernity reflected in its literature. Seminal works of stalwarts are used to understand the consequence of this cultural synthesis. Conclusively, this theoretical research inspects the efficiency of secular culture, intertwined with internal coherence and throws light on the plurality of texts in Indian literature.

Performance Evaluation of Distributed Deep Learning Frameworks in Cloud Environment

2016 has become the year of the Artificial Intelligence explosion. AI technologies are getting more and more matured that most world well-known tech giants are making large investment to increase the capabilities in AI. Machine learning is the science of getting computers to act without being explicitly programmed, and deep learning is a subset of machine learning that uses deep neural network to train a machine to learn  features directly from data. Deep learning realizes many machine learning applications which expand the field of AI. At the present time, deep learning frameworks have been widely deployed on servers for deep learning applications in both academia and industry. In training deep neural networks, there are many standard processes or algorithms, but the performance of different frameworks might be different. In this paper we evaluate the running performance of two state-of-the-art distributed deep learning frameworks that are running training calculation in parallel over multi GPU and multi nodes in our cloud environment. We evaluate the training performance of the frameworks with ResNet-50 convolutional neural network, and we analyze what factors that result in the performance among both distributed frameworks as well. Through the experimental analysis, we identify the overheads which could be further optimized. The main contribution is that the evaluation results provide further optimization directions in both performance tuning and algorithmic design.

Promoting Social Advocacy through Digital Storytelling: The Case of Ocean Acidification

Many chemical changes in the atmosphere and the ocean are invisible to the naked eye, but they have profound impacts. These changes not only confirm the phenomenon of global carbon pollution, but also forewarn that more changes are coming. The carbon dioxide gases emitted from the burning of fossil fuels dissolve into the ocean and chemically react with seawater to form carbonic acid, which increases the acidity of the originally alkaline seawater. This gradual acidification is occurring at an unprecedented rate and will affect the effective formation of carapace of some marine organisms such as corals and crustaceans, which are almost entirely composed of calcium carbonate. The carapace of these organisms will become more dissoluble. Acidified seawater not only threatens the survival of marine life, but also negatively impacts the global ecosystem via the food chain. Faced with the threat of ocean acidification, all humans are duty-bound. The industrial sector outputs the highest level of carbon dioxide emissions in Taiwan, and the petrochemical industry is the major contributor. Ever since the construction of Formosa Plastics Group's No. 6 Naphtha Cracker Plant in Yunlin County, there have been many environmental concerns such as air pollution and carbon dioxide emission. The marine life along the coast of Yunlin is directly affected by ocean acidification arising from the carbon emissions. Societal change demands our willingness to act, which is what social advocacy promotes. This study uses digital storytelling for social advocacy and ocean acidification as the subject of a visual narrative in visualization to demonstrate the subsequent promotion of social advocacy. Storytelling can transform dull knowledge into an engaging narrative of the crisis faced by marine life. Digital dissemination is an effective social-work practice. The visualization promoting awareness on ocean acidification disseminated via social media platforms, such as Facebook and Instagram. Social media enables users to compose their own messages and share information across different platforms, which helps disseminate the core message of social advocacy.

Continuous Plug Flow and Discrete Particle Phase Coupling Using Triangular Parcels

Various processes are modelled using a discrete phase, where particles are seeded from a source. Such particles can represent liquid water droplets, which are affecting the continuous phase by exchanging thermal energy, momentum, species etc. Discrete phases are typically modelled using parcel, which represents a collection of particles, which share properties such as temperature, velocity etc. When coupling the phases, the exchange rates are integrated over the cell, in which the parcel is located. This can cause spikes and fluctuating exchange rates. This paper presents an alternative method of coupling a discrete and a continuous plug flow phase. This is done using triangular parcels, which span between nodes following the dynamics of single droplets. Thus, the triangular parcels are propagated using the corner nodes. At each time step, the exchange rates are spatially integrated over the surface of the triangular parcels, which yields a smooth continuous exchange rate to the continuous phase. The results shows that the method is more stable, converges slightly faster and yields smooth exchange rates compared with the steam tube approach. However, the computational requirements are about five times greater, so the applicability of the alternative method should be limited to processes, where the exchange rates are important. The overall balances of the exchanged properties did not change significantly using the new approach.

Six Sigma-Based Optimization of Shrinkage Accuracy in Injection Molding Processes

This paper focuses on using six sigma methodologies to reach the desired shrinkage of a manufactured high-density polyurethane (HDPE) part produced by the injection molding machine. It presents a case study where the correct shrinkage is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for an injection molding process. To improve this process and keep the product within specifications, the six sigma methodology, design, measure, analyze, improve, and control (DMAIC) approach, was implemented in this study. The six sigma approach was paired with the Taguchi methodology to identify the optimized processing parameters that keep the shrinkage rate within the specifications by our customer. An L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of the cooling time, melt temperature, holding time, and metering stroke. The noise factor is the difference between material brand 1 and material brand 2. After the confirmation run was completed, measurements verify that the new parameter settings are optimal. With the new settings, the process capability index has improved dramatically. The purpose of this study is to show that the six sigma and Taguchi methodology can be efficiently used to determine important factors that will improve the process capability index of the injection molding process.

An Algorithm of Regulation of Glucose-Insulin Concentration in the Blood

The pancreas is an elongated organ that extends across the abdomen, below the stomach. In addition, it secretes certain enzymes that aid in food digestion. The pancreas also manufactures hormones responsible for regulating blood glucose levels. In the present paper, we propose a mathematical model to study the homeostasis of glucose and insulin in healthy human, and a simulation of this model, which depicts the physiological events after a meal, will be represented in ordinary humans. The aim of this paper is to design an algorithm which regulates the level of glucose in the blood. The algorithm applied the concept of expert system for performing an algorithm control in the form of an "active" used to prescribe the rate of insulin infusion. By decomposing the system into subsystems, we have developed parametric models of each subsystem by using a forcing function strategy. The results showed a performance of the control system.

Modeling Nanomechanical Behavior of ZnO Nanowires as a Function of Nano-Diameter

Elastic performances, as an essential property of nanowires (NWs), play a significant role in the design and fabrication of modern nanodevices. In this paper, our interest is focused on ZnO NWs to investigate wire diameter (Dwire ≤ 400 nm) effects on elastic properties. The plotted data reveal that a strong size dependence of the elastic constants exists when the wire diameter is smaller than ~ 100 nm. For larger diameters (Dwire > 100 nm), these ones approach their corresponding bulk values. To enrich this study, we make use of the scanning acoustic microscopy simulation technique. The calculation methodology consists of several steps: determination of longitudinal and transverse wave velocities, calculation of refection coefficients, calculation of acoustic signatures and Rayleigh velocity determination. Quantitatively, it was found that changes in ZnO diameters over the ranges 1 nm ≤ Dwire ≤ 100 nm lead to similar exponential variations, for all elastic parameters, of the from: A = a + b exp(-Dwire/c) where a, b, and c are characteristic constants of a given parameter. The developed relation can be used to predict elastic properties of such NW by just knowing its diameter and vice versa.

A Study of Hamilton-Jacobi-Bellman Equation Systems Arising in Differential Game Models of Changing Society

This paper is concerned with a system of Hamilton-Jacobi-Bellman equations coupled with an autonomous dynamical system. The mathematical system arises in the differential game formulation of political economy models as an infinite-horizon continuous-time differential game with discounted instantaneous payoff rates and continuously and discretely varying state variables. The existence of a weak solution of the PDE system is proven and a computational scheme of approximate solution is developed for a class of such systems. A model of democratization is mathematically analyzed as an illustration of application.

Adsorption of Phenolic Compounds on Activated Carbon DSAC36-24

Activated carbon DSAC36-24 iy is adsorbent materials, characterized by a specific surface area of 548.13 m²g⁻¹. Their manufacture uses the natural raw materials like the nucleus of dates. In this study the treatment is done in two stages: A chemical treatment by H3PO4 followed by a physical treatment under nitrogen for 1 hour then under stream of CO2 for 24 hours. A characterization of the various parameters was determined such as the measurement of the specific surface area, determination of pHPZC, bulk density, iodine value. The study of the adsorption of organic molecules (hydroquinone, paranitrophenol, 2,4-dinitrophenol, 2,4,6-trinitrophenol) indicates that the adsorption phenomena are essentially due to the van der Waals interaction. In the case of organic molecules carrying the polar substituents, the existence of hydrogen bonds is also proved by the donor-acceptor forces. The study of the pH effect was done with modeling by different models (Langmuir, Freundlich, Langmuir-Freundlich, Redlich-Peterson), a kinetic treatment is also followed by the application of Lagergren, Weber, Macky.

Taguchi-Based Six Sigma Approach to Optimize Surface Roughness for Milling Processes

This paper focuses on using Six Sigma methodologies to improve the surface roughness of a manufactured part produced by the CNC milling machine. It presents a case study where the surface roughness of milled aluminum is required to reduce or eliminate defects and to improve the process capability index Cp and Cpk for a CNC milling process. The six sigma methodology, DMAIC (design, measure, analyze, improve, and control) approach, was applied in this study to improve the process, reduce defects, and ultimately reduce costs. The Taguchi-based six sigma approach was applied to identify the optimized processing parameters that led to the targeted surface roughness specified by our customer. A L9 orthogonal array was applied in the Taguchi experimental design, with four controllable factors and one non-controllable/noise factor. The four controllable factors identified consist of feed rate, depth of cut, spindle speed, and surface roughness. The noise factor is the difference between the old cutting tool and the new cutting tool. The confirmation run with the optimal parameters confirmed that the new parameter settings are correct. The new settings also improved the process capability index. The purpose of this study is that the Taguchi–based six sigma approach can be efficiently used to phase out defects and improve the process capability index of the CNC milling process.

Comparative Study of Experimental and Theoretical Convective, Evaporative for Two Model Distiller

The purification of brackish seawater becomes a necessity and not a choice against demographic and industrial growth especially in third world countries. Two models can be used in this work: simple solar still and simple solar still coupled with a heat pump. In this research, the productivity of water by Simple Solar Distiller (SSD) and Simple Solar Distiller Hybrid Heat Pump (SSDHP) was determined by the orientation, the use of heat pump, the simple or double glass cover. The productivity can exceed 1.2 L/m²h for the SSDHP and 0.5 L/m²h for SSD model. The result of the global efficiency is determined for two models SSD and SSDHP give respectively 30%, 50%. The internal efficiency attained 35% for SSD and 60% of the SSDHP models. Convective heat coefficient can be determined by attained 2.5 W/m²°C and 0.5 W/m²°C respectively for SSDHP and SSD models.

A Low-Area Fully-Reconfigurable Hardware Design of Fast Fourier Transform System for 3GPP-LTE Standard

This paper presents a low-area and fully-reconfigurable Fast Fourier Transform (FFT) hardware design for 3GPP-LTE communication standard. It can fully support 32 different FFT sizes, up to 2048 FFT points. Besides, a special processing element is developed for making reconfigurable computing characteristics possible, while first-in first-out (FIFO) scheduling scheme design technique is proposed for hardware-friendly FIFO resource arranging. In a synthesis chip realization via TSMC 40 nm CMOS technology, the hardware circuit only occupies core area of 0.2325 mm2 and dissipates 233.5 mW at maximal operating frequency of 250 MHz.

Exploring Influence Range of Tainan City Using Electronic Toll Collection Big Data

Big Data has been attracted a lot of attentions in many fields for analyzing research issues based on a large number of maternal data. Electronic Toll Collection (ETC) is one of Intelligent Transportation System (ITS) applications in Taiwan, used to record starting point, end point, distance and travel time of vehicle on the national freeway. This study, taking advantage of ETC big data, combined with urban planning theory, attempts to explore various phenomena of inter-city transportation activities. ETC, one of government's open data, is numerous, complete and quick-update. One may recall that living area has been delimited with location, population, area and subjective consciousness. However, these factors cannot appropriately reflect what people’s movement path is in daily life. In this study, the concept of "Living Area" is replaced by "Influence Range" to show dynamic and variation with time and purposes of activities. This study uses data mining with Python and Excel, and visualizes the number of trips with GIS to explore influence range of Tainan city and the purpose of trips, and discuss living area delimited in current. It dialogues between the concepts of "Central Place Theory" and "Living Area", presents the new point of view, integrates the application of big data, urban planning and transportation. The finding will be valuable for resource allocation and land apportionment of spatial planning.

Design and Optimization for a Compliant Gripper with Force Regulation Mechanism

This paper presents a design and optimization for a compliant gripper. The gripper is constructed based on the concept of compliant mechanism with flexure hinge. A passive force regulation mechanism is presented to control the grasping force a micro-sized object instead of using a sensor force. The force regulation mechanism is designed using the planar springs. The gripper is expected to obtain a large range of displacement to handle various sized objects. First of all, the statics and dynamics of the gripper are investigated by using the finite element analysis in ANSYS software. And then, the design parameters of the gripper are optimized via Taguchi method. An orthogonal array L9 is used to establish an experimental matrix. Subsequently, the signal to noise ratio is analyzed to find the optimal solution. Finally, the response surface methodology is employed to model the relationship between the design parameters and the output displacement of the gripper. The design of experiment method is then used to analyze the sensitivity so as to determine the effect of each parameter on the displacement. The results showed that the compliant gripper can move with a large displacement of 213.51 mm and the force regulation mechanism is expected to be used for high precision positioning systems.

Image Rotation Using an Augmented 2-Step Shear Transform

Image rotation is one of main pre-processing steps for image processing or image pattern recognition. It is implemented with a rotation matrix multiplication. It requires a lot of floating point arithmetic operations and trigonometric calculations, so it takes a long time to execute. Therefore, there has been a need for a high speed image rotation algorithm without two major time-consuming operations. However, the rotated image has a drawback, i.e. distortions. We solved the problem using an augmented two-step shear transform. We compare the presented algorithm with the conventional rotation with images of various sizes. Experimental results show that the presented algorithm is superior to the conventional rotation one.

Fabrication, Testing and Machinability Evaluation of Glass Fiber Reinforced Epoxy Composites

The present paper deals with designing and fabricating an apparatus for the speedy and accurate manufacturing of fiber reinforced composite lamina of different orientation, thickness and stacking sequences for testing. Properties derived through an analytical approach are verified through measuring the elastic modulus, ultimate tensile strength, flexural modulus and flexural strength of the samples. The 00 orientation ply looks stiffer compared to the 900 ply. Similarly, the flexural strength of 00 ply is higher than to the 900 ply. Sample machinability has been studied by conducting numbers of drilling based on Taguchi Design experiments. Multi Responses (Delamination and Damage grading) is obtained using the desirability approach and optimum cutting condition (spindle speed, feed and drill diameter), at which responses are minimized is obtained thereafter. Delamination increases nonlinearly with the increase in spindle speed. Similarly, the influence of the drill diameter on delamination is higher than the spindle speed and feed rate.

A Compact Quasi-Zero Stiffness Vibration Isolator Using Flexure-Based Spring Mechanisms Capable of Tunable Stiffness

This study presents a quasi-zero stiffness (QZS) vibration isolator using flexure-based spring mechanisms which afford both negative and positive stiffness elements, which enable self-adjustment. The QZS property of the isolator is achieved at the equilibrium position. A nonlinear mathematical model is then developed, based on the pre-compression of the flexure-based spring mechanisms. The dynamics are further analyzed using the Harmonic Balance method. The vibration attention efficiency is illustrated using displacement transmissibility, which is then compared with the corresponding linear isolator. The effects of parameters on performance are also investigated by numerical solutions. The flexure-based spring mechanisms are subsequently designed using the concept of compliant mechanisms, with evaluation by ANSYS software, and simulations of the QZS isolator.

Secure Proxy Signature Based on Factoring and Discrete Logarithm

A digital signature is an electronic signature form used by an original signer to sign a specific document. When the original signer is not in his office or when he/she travels outside, he/she delegates his signing capability to a proxy signer and then the proxy signer generates a signing message on behalf of the original signer. The two parties must be able to authenticate one another and agree on a secret encryption key, in order to communicate securely over an unreliable public network. Authenticated key agreement protocols have an important role in building a secure communications network between the two parties. In this paper, we present a secure proxy signature scheme over an efficient and secure authenticated key agreement protocol based on factoring and discrete logarithm problem.

An Efficient Proxy Signature Scheme Over a Secure Communications Network

Proxy signature scheme permits an original signer to delegate his/her signing capability to a proxy signer, and then the proxy signer generates a signing message on behalf of the original signer. The two parties must be able to authenticate one another and agree on a secret encryption key, in order to communicate securely over an unreliable public network. Authenticated key agreement protocols have an important role in building secure communications network between the two parties. In this paper, we present a secure proxy signature scheme over an efficient and secure authenticated key agreement protocol based on the discrete logarithm problem.