A Study of the Costs and Benefits of Smart City Projects Including the Scenario of Public-Private Partnerships

A smart city project embraces benefits and costs which can be classified under direct and indirect categories. Externalities come into the picture, but they are often difficult to quantify. Despite this barrier, policy makers need to carry out cost-benefit analysis to justify the huge investments needed to make a city smart. The recent trend is towards the engagement of the private sector to utilize their resources and expertise, especially in the Information and Communication Technology (ICT) areas, where innovations blossom. This study focuses on the identification of costs (on a life cycle basis) and benefits associated with smart city project developments based on a comprehensive literature review and case studies, where public-private partnerships would warrant consideration, the related costs and benefits are highlighted. The findings will be useful for policy makers of cities.

An Alternative Approach for Assessing the Impact of Cutting Conditions on Surface Roughness Using Single Decision Tree

In this study, an approach to identify factors affecting on surface roughness in a machining process is presented. This study is based on 81 data about surface roughness over a wide range of cutting tools (conventional, cutting tool with holes, cutting tool with composite material), workpiece materials (AISI 1045 Steel, AA2024 aluminum alloy, A48-class30 gray cast iron), spindle speed (630-1000 rpm), feed rate (0.05-0.075 mm/rev), depth of cut (0.05-0.15 mm) and tool overhang (41-65 mm). A single decision tree (SDT) analysis was done to identify factors for predicting a model of surface roughness, and the CART algorithm was employed for building and evaluating regression tree. Results show that a single decision tree is better than traditional regression models with higher rate and forecast accuracy and strong value.

Minimum-Fuel Optimal Trajectory for Reusable First-Stage Rocket Landing Using Particle Swarm Optimization

Reusable launch vehicles (RLVs) present a more environmentally-friendly approach to accessing space when compared to traditional launch vehicles that are discarded after each flight. This paper studies the recyclable nature of RLVs by presenting a solution method for determining minimum-fuel optimal trajectories using principles from optimal control theory and particle swarm optimization (PSO). This problem is formulated as a minimum-landing error powered descent problem where it is desired to move the RLV from a fixed set of initial conditions to three different sets of terminal conditions. However, unlike other powered descent studies, this paper considers the highly nonlinear effects caused by atmospheric drag, which are often ignored for studies on the Moon or on Mars. Rather than optimizing the controls directly, the throttle control is assumed to be bang-off-bang with a predetermined thrust direction for each phase of flight. The PSO method is verified in a one-dimensional comparison study, and it is then applied to the two-dimensional cases, the results of which are illustrated.

Performance Analysis of Artificial Neural Network Based Land Cover Classification

Landcover classification using automated classification techniques, while employing remotely sensed multi-spectral imagery, is one of the promising areas of research. Different land conditions at different time are captured through satellite and monitored by applying different classification algorithms in specific environment. In this paper, a SPOT-5 image provided by SUPARCO has been studied and classified in Environment for Visual Interpretation (ENVI), a tool widely used in remote sensing. Then, Artificial Neural Network (ANN) classification technique is used to detect the land cover changes in Abbottabad district. Obtained results are compared with a pixel based Distance classifier. The results show that ANN gives the better overall accuracy of 99.20% and Kappa coefficient value of 0.98 over the Mahalanobis Distance Classifier.

Modelling a Hospital as a Queueing Network: Analysis for Improving Performance

In this paper, the flow of different classes of patients into a hospital is modelled and analyzed by using the queueing network analyzer (QNA) algorithm and discrete event simulation. Input data for QNA are the rate and variability parameters of the arrival and service times in addition to the number of servers in each facility. Patient flows mostly match real flow for a hospital in Egypt. Based on the analysis of the waiting times, two approaches are suggested for improving performance: Separating patients into service groups, and adopting different service policies for sequencing patients through hospital units. The separation of a specific group of patients, with higher performance target, to be served separately from the rest of patients requiring lower performance target, requires the same capacity while improves performance for the selected group of patients with higher target. Besides, it is shown that adopting the shortest processing time and shortest remaining processing time service policies among other tested policies would results in, respectively, 11.47% and 13.75% reduction in average waiting time relative to first come first served policy.

Forensic Speaker Verification in Noisy Environmental by Enhancing the Speech Signal Using ICA Approach

We propose a system to real environmental noise and channel mismatch for forensic speaker verification systems. This method is based on suppressing various types of real environmental noise by using independent component analysis (ICA) algorithm. The enhanced speech signal is applied to mel frequency cepstral coefficients (MFCC) or MFCC feature warping to extract the essential characteristics of the speech signal. Channel effects are reduced using an intermediate vector (i-vector) and probabilistic linear discriminant analysis (PLDA) approach for classification. The proposed algorithm is evaluated by using an Australian forensic voice comparison database, combined with car, street and home noises from QUT-NOISE at a signal to noise ratio (SNR) ranging from -10 dB to 10 dB. Experimental results indicate that the MFCC feature warping-ICA achieves a reduction in equal error rate about (48.22%, 44.66%, and 50.07%) over using MFCC feature warping when the test speech signals are corrupted with random sessions of street, car, and home noises at -10 dB SNR.

Electromyography Pattern Classification with Laplacian Eigenmaps in Human Running

Electromyography (EMG) is one of the most important interfaces between humans and robots for rehabilitation. Decoding this signal helps to recognize muscle activation and converts it into smooth motion for the robots. Detecting each muscle’s pattern during walking and running is vital for improving the quality of a patient’s life. In this study, EMG data from 10 muscles in 10 subjects at 4 different speeds were analyzed. EMG signals are nonlinear with high dimensionality. To deal with this challenge, we extracted some features in time-frequency domain and used manifold learning and Laplacian Eigenmaps algorithm to find the intrinsic features that represent data in low-dimensional space. We then used the Bayesian classifier to identify various patterns of EMG signals for different muscles across a range of running speeds. The best result for vastus medialis muscle corresponds to 97.87±0.69 for sensitivity and 88.37±0.79 for specificity with 97.07±0.29 accuracy using Bayesian classifier. The results of this study provide important insight into human movement and its application for robotics research.

Improvement of Ventilation and Thermal Comfort Using the Atrium Design for Traditional Folk Houses-Fujian Earthen Building

Fujian earthen building which was known as a classic for ecological buildings was listed on the world heritage in 2008 (UNESCO) in China. Its design strategy can be applied to modern architecture planning and design. This study chose two different cases (Round Atrium: Er-Yi Building, Double Round Atrium: Zhen-Chen Building) of earthen building in Fu-Jian to compare the ventilation effects of different atrium forms. We adopt field measurements and computational fluid dynamics (CFD) simulation of temperature, humidity, and wind environment to identify the relationship between external environment and atrium about comfort and to confirm the relationship about atrium H/W (height/width). Results indicate that, through the atrium convection effect, it makes the natural wind guides to each space surrounded and keeps indoor comfort. It illustrates that the smaller the ratio of the H/W which is the relationship between the height and the width of an atrium is, the greater the wind speed generated within the street valley. Moreover, the wind speed is very close to the reference wind speed. This field measurement verifies that the value of H/W has great influence of solar radiation heat and sunshine shadows. The ventilation efficiency is: Er-Yi Building (H/W =0.2778) > Zhen-Chen Building (H/W=0.3670). Comparing the cases with the same shape but with different H/W, through the different size patios, airflow revolves in the atriums and can be brought into each interior space. The atrium settings meet the need of building ventilation, and can adjust the humidity and temperature within the buildings. It also creates good ventilation effect.

Measuring E-Learning Effectiveness Using a Three-Way Comparison

The way e-learning effectiveness has been notoriously measured within an academic setting is by comparing the e-learning medium to the traditional face-to-face teaching methodology. In this paper, a simple yet innovative comparison methodology is introduced, whereby the effectiveness of next generation e-learning systems are assessed in contrast not only to the face-to-face mode, but also to the classical e-learning modality. Ethical and logistical issues are also discussed, as this three-way approach to compare teaching methodologies was applied and documented in a real empirical study within a higher education institution.

Genetic Algorithms for Feature Generation in the Context of Audio Classification

Choosing good features is an essential part of machine learning. Recent techniques aim to automate this process. For instance, feature learning intends to learn the transformation of raw data into a useful representation to machine learning tasks. In automatic audio classification tasks, this is interesting since the audio, usually complex information, needs to be transformed into a computationally convenient input to process. Another technique tries to generate features by searching a feature space. Genetic algorithms, for instance, have being used to generate audio features by combining or modifying them. We find this approach particularly interesting and, despite the undeniable advances of feature learning approaches, we wanted to take a step forward in the use of genetic algorithms to find audio features, combining them with more conventional methods, like PCA, and inserting search control mechanisms, such as constraints over a confusion matrix. This work presents the results obtained on particular audio classification problems.

Data Projects for “Social Good”: Challenges and Opportunities

One of the application fields for data analysis techniques and technologies gaining momentum is the area of social good or “common good”, covering cases related to humanitarian crises, global health care, or ecology and environmental issues, among others. The promotion of data-driven projects in this field aims at increasing the efficacy and efficiency of social initiatives, improving the way these actions help humanity in general and people in need in particular. This application field, however, poses its own barriers and challenges when developing data-driven projects, lagging behind in comparison with other scenarios. These challenges derive from aspects such as the scope and scale of the social issue to solve, cultural and political barriers, the skills of main stakeholders and the technological resources available, the motivation to be engaged in such projects, or the ethical and legal issues related to sensitive data. This paper analyzes the application of data projects in the field of social good, reviewing its current state and noteworthy initiatives, and presenting a framework covering the key aspects to analyze in such projects. The goal is to provide guidelines to understand the main challenges and opportunities for this type of data project, as well as identifying the main differential issues compared to “classical” data projects in general. A case study is presented on the initial steps and stakeholder analysis of a data project for the inclusion of refugees in the city of Frankfurt, Germany, in order to empirically confront the framework with a real example.

Movies and Dynamic Mathematical Objects on Trigonometry for Mobile Phones

This paper is about movies and dynamic objects for mobile phones. Dynamic objects are the software programmed by JavaScript. They consist of geometric figures and work on HTML5-compliant browsers. Mobile phones are very popular among teenagers. They like watching movies and playing games on them. So, mathematics movies and dynamic objects would enhance teaching and learning processes. In the movies, manga characters speak with artificially synchronized voices. They teach trigonometry together with dynamic mathematical objects. Many movies are created. They are Windows Media files or MP4 movies. These movies and dynamic objects are not only used in the classroom but also distributed to students. By watching movies, students can study trigonometry before or after class.

Lean Models Classification: Towards a Holistic View

The purpose of this paper is to present a classification of Lean models which aims to capture all the concepts related to this approach and thus facilitate its implementation. This classification allows the identification of the most relevant models according to several dimensions. From this perspective, we present a review and an analysis of Lean models literature and we propose dimensions for the classification of the current proposals while respecting among others the axes of the Lean approach, the maturity of the models as well as their application domains. This classification allowed us to conclude that researchers essentially consider the Lean approach as a toolbox also they design their models to solve problems related to a specific environment. Since Lean approach is no longer intended only for the automotive sector where it was invented, but to all fields (IT, Hospital, ...), we consider that this approach requires a generic model that is capable of being implemented in all areas.

Biaxial Buckling of Single Layer Graphene Sheet Based on Nonlocal Plate Model and Molecular Dynamics Simulation

The biaxial buckling behavior of single-layered graphene sheets (SLGSs) is studied in the present work. To consider the size-effects in the analysis, Eringen’s nonlocal elasticity equations are incorporated into classical plate theory (CLPT). A Generalized Differential Quadrature Method (GDQM) approach is utilized and numerical solutions for the critical buckling loads are obtained. Then, molecular dynamics (MD) simulations are performed for a series of zigzag SLGSs with different side-lengths and with various boundary conditions, the results of which are matched with those obtained by the nonlocal plate model to numerical the appropriate values of nonlocal parameter relevant to each type of boundary conditions.

The Incidence of Obesity among Adult Women in Pekanbaru City, Indonesia, Related to High Fat Consumption, Stress Level, and Physical Activity

Background: Obesity has been recognized as a global health problem. Individuals classified as overweight and obese are increasing at an alarming rate. This condition is associated with psychological and physiological problems. as a person reaches adulthood, somatic growth ceases. At this stage, the human body has developed fully, to a stable state. As the capital of Riau Province in Indonesia, Pekanbaru is dominated by Malay ethnic population habitually consuming cholesterol-rich fatty foods as a daily menu, a trigger to the onset of obesity resulting in high prevalence of degenerative diseases. Research objectives: The aim of this study is elaborating the relationship between high-fat consumption pattern, stress level, physical activity and the incidence of obesity in adult women in Pekanbaru city. Research Methods: Among the combined research methods applied in this study, the first stage is quantitative observational, analytical cross-sectional research design with adult women aged 20-40 living in Pekanbaru city. The sample consists of 200 women with BMI≥25. Sample data is processed with univariate, bivariate (correlation and simple linear regression) and multivariate (multiple linear regression) analysis. The second phase is qualitative descriptive study purposive sampling by in-depth interviews. six participants withdrew from the study. Results: According to the results of the bivariate analysis, there are relationships between the incidence of obesity and the pattern of high fat foods consumption (energy intake (p≤0.000; r = 0.536), protein intake (p≤0.000; r=0.307), fat intake (p≤0.000; r=0.416), carbohydrate intake (p≤0.000; r=0.430), frequency of fatty food consumption (p≤0.000; r=0.506) and frequency of viscera foods consumption (p≤0.000; r=0.535). There is a relationship between physical activity and incidence of obesity (p≤0.000; r=-0.631). However, there is no relationship between the level of stress (p=0.741; r=0.019-) and the incidence of obesity. Physical activity is a predominant factor in the incidence of obesity in adult women in Pekanbaru city. Conclusion: There are relationships between high-fat food consumption pattern, physical activity and the incidence of obesity in Pekanbaru city whereas physical activity is a predominant factor in the occurrence of obesity, supported by the unchangeable pattern of high-fat foods consumption.

Effect of Low Plastic Clay Quantity on Behavioral Characteristics of Loose Sand

After the Nigatta earthquake in Japan, in 1960, the liquefaction and its related hazards, moved to the thick of matter. Most of the research have been carried out on clean sands and silty sands so far, in order to study the effect of fine particles, confinement pressures, density and so on. However, because of this delusion that adhesiveness of clay prevents the liquefaction in sand, studies on clayey sands have not been taken seriously. However, several liquefactions happened in clayey sands in recent years, and lead to the necessity of more studies in this field. The studies which were carried out so far focused on high plastic clays. In this paper, the effect of low plasticity clays on the behavioral characteristics of sands is discussed. Thus, some triaxial tests were carried out on clean sands and clayey sands with different percentages of added clay. Specimens were compacted in various densities to study the effect of quantity of clay on various densities, too. Based on the findings, the amount of clay affects the behavior of sand greatly and leads to substantial changes in peak bearing capacity and steady state values.

The Role of Paper in the Copy Identification of Safavid Era Shahnamehs of Tabriz Doctrine

To investigate and explain the history of each copy, we must refer to its past because it highlights parts of the civilization of people among which this copy has been codified. In this paper, eight Ferdowsi’s Shahnameh of Safavid era of Tabriz doctrine available in Iranian libraries and museums are studied. Undoubtedly, it can be said that Ferdowsi’s Shahnameh is one of the most important books that has been transcribed many times in different eras because it explains the Iranian champions’ prowess and it includes the history of Iran from Pishdadian to Sasanian dynasty. In addition, it has been attractive for governors and artists. The research methodology of this article is based on the analytical-descriptive arguments. The research hypothesis is based on papers used in Shahnameh writing in Safavid era of Tabriz doctrine were mostly Isfahanian papers existed. At that time, Isfahanian paper was unique in terms of quality, clarity, flatness of the sheets, volume, shape, softness and elegance, strength, and smoothness. This paper was mostly used to prepare the courtier and exquisite copies. This shows that the prepared copies in Safavid era of Tabriz doctrine were very important because the artists and people who ordered and were out of the court have ordered Isfahanian paper for writing their books.

Secure Multiparty Computations for Privacy Preserving Classifiers

Secure computations are essential while performing privacy preserving data mining. Distributed privacy preserving data mining involve two to more sites that cannot pool in their data to a third party due to the violation of law regarding the individual. Hence in order to model the private data without compromising privacy and information loss, secure multiparty computations are used. Secure computations of product, mean, variance, dot product, sigmoid function using the additive and multiplicative homomorphic property is discussed. The computations are performed on vertically partitioned data with a single site holding the class value.

H∞ Takagi-Sugeno Fuzzy State-Derivative Feedback Control Design for Nonlinear Dynamic Systems

This paper considers an H∞ TS fuzzy state-derivative feedback controller for a class of nonlinear dynamical systems. A Takagi-Sugeno (TS) fuzzy model is used to approximate a class of nonlinear dynamical systems. Then, based on a linear matrix inequality (LMI) approach, we design an H∞ TS fuzzy state-derivative feedback control law which guarantees L2-gain of the mapping from the exogenous input noise to the regulated output to be less or equal to a prescribed value. We derive a sufficient condition such that the system with the fuzzy controller is asymptotically stable and H∞ performance is satisfied. Finally, we provide and simulate a numerical example is provided to illustrate the stability and the effectiveness of the proposed controller.

The Thermochemical Conversion of Lactic Acid in Subcritical and Supercritical Water

One way to utilize biomass is to thermochemically convert it into gases and chemicals. For conversion of biomass, glucose is a particularly popular model compound for cellulose, or more generally for biomass. The present study takes a different approach by employing lactic acid as the model compound for cellulose. Since lactic acid and glucose have identical elemental composition, they are expected to produce similar results as they go through the conversion process. In the current study, lactic acid was thermochemically converted to assess its reactivity and reaction mechanism in subcritical and supercritical water, by using a 16-ml autoclave reactor. The major operating parameters investigated include: The reaction temperature, from 673 to 873 K, the reaction pressure, 10 and 25 MPa, the dosage of oxidizing agent, 0 and 0.5 chemical oxygen demand, and the concentration of lactic acid in the feed, 0.5 and 1.0 M. Gaseous products from the conversion were generally found to be comparable to those derived from the conversion of glucose.