Normal and Peaberry Coffee Beans Classification from Green Coffee Bean Images Using Convolutional Neural Networks and Support Vector Machine

The aim of this study is to develop a system which can identify and sort peaberries automatically at low cost for coffee producers in developing countries. In this paper, the focus is on the classification of peaberries and normal coffee beans using image processing and machine learning techniques. The peaberry is not bad and not a normal bean. The peaberry is born in an only single seed, relatively round seed from a coffee cherry instead of the usual flat-sided pair of beans. It has another value and flavor. To make the taste of the coffee better, it is necessary to separate the peaberry and normal bean before green coffee beans roasting. Otherwise, the taste of total beans will be mixed, and it will be bad. In roaster procedure time, all the beans shape, size, and weight must be unique; otherwise, the larger bean will take more time for roasting inside. The peaberry has a different size and different shape even though they have the same weight as normal beans. The peaberry roasts slower than other normal beans. Therefore, neither technique provides a good option to select the peaberries. Defect beans, e.g., sour, broken, black, and fade bean, are easy to check and pick up manually by hand. On the other hand, the peaberry pick up is very difficult even for trained specialists because the shape and color of the peaberry are similar to normal beans. In this study, we use image processing and machine learning techniques to discriminate the normal and peaberry bean as a part of the sorting system. As the first step, we applied Deep Convolutional Neural Networks (CNN) and Support Vector Machine (SVM) as machine learning techniques to discriminate the peaberry and normal bean. As a result, better performance was obtained with CNN than with SVM for the discrimination of the peaberry. The trained artificial neural network with high performance CPU and GPU in this work will be simply installed into the inexpensive and low in calculation Raspberry Pi system. We assume that this system will be used in under developed countries. The study evaluates and compares the feasibility of the methods in terms of accuracy of classification and processing speed.

Research on Reservoir Lithology Prediction Based on Residual Neural Network and Squeeze-and- Excitation Neural Network

Conventional reservoir prediction methods ar not sufficient to explore the implicit relation between seismic attributes, and thus data utilization is low. In order to improve the predictive classification accuracy of reservoir lithology, this paper proposes a deep learning lithology prediction method based on ResNet (Residual Neural Network) and SENet (Squeeze-and-Excitation Neural Network). The neural network model is built and trained by using seismic attribute data and lithology data of Shengli oilfield, and the nonlinear mapping relationship between seismic attribute and lithology marker is established. The experimental results show that this method can significantly improve the classification effect of reservoir lithology, and the classification accuracy is close to 70%. This study can effectively predict the lithology of undrilled area and provide support for exploration and development.

Convergence Analysis of Training Two-Hidden-Layer Partially Over-Parameterized ReLU Networks via Gradient Descent

Over-parameterized neural networks have attracted a great deal of attention in recent deep learning theory research, as they challenge the classic perspective of over-fitting when the model has excessive parameters and have gained empirical success in various settings. While a number of theoretical works have been presented to demystify properties of such models, the convergence properties of such models are still far from being thoroughly understood. In this work, we study the convergence properties of training two-hidden-layer partially over-parameterized fully connected networks with the Rectified Linear Unit activation via gradient descent. To our knowledge, this is the first theoretical work to understand convergence properties of deep over-parameterized networks without the equally-wide-hidden-layer assumption and other unrealistic assumptions. We provide a probabilistic lower bound of the widths of hidden layers and proved linear convergence rate of gradient descent. We also conducted experiments on synthetic and real-world datasets to validate our theory.

Multi-Objective Optimal Design of a Cascade Control System for a Class of Underactuated Mechanical Systems

This paper presents a multi-objective optimal design of a cascade control system for an underactuated mechanical system. Cascade control structures usually include two control algorithms (inner and outer). To design such a control system properly, the following conflicting objectives should be considered at the same time: 1) the inner closed-loop control must be faster than the outer one, 2) the inner loop should fast reject any disturbance and prevent it from propagating to the outer loop, 3) the controlled system should be insensitive to measurement noise, and 4) the controlled system should be driven by optimal energy. Such a control problem can be formulated as a multi-objective optimization problem such that the optimal trade-offs among these design goals are found. To authors best knowledge, such a problem has not been studied in multi-objective settings so far. In this work, an underactuated mechanical system consisting of a rotary servo motor and a ball and beam is used for the computer simulations, the setup parameters of the inner and outer control systems are tuned by NSGA-II (Non-dominated Sorting Genetic Algorithm), and the dominancy concept is used to find the optimal design points. The solution of this problem is not a single optimal cascade control, but rather a set of optimal cascade controllers (called Pareto set) which represent the optimal trade-offs among the selected design criteria. The function evaluation of the Pareto set is called the Pareto front. The solution set is introduced to the decision-maker who can choose any point to implement. The simulation results in terms of Pareto front and time responses to external signals show the competing nature among the design objectives. The presented study may become the basis for multi-objective optimal design of multi-loop control systems.

Lexicon-Based Sentiment Analysis for Stock Movement Prediction

Sentiment analysis is a broad and expanding field that aims to extract and classify opinions from textual data. Lexicon-based approaches are based on the use of a sentiment lexicon, i.e., a list of words each mapped to a sentiment score, to rate the sentiment of a text chunk. Our work focuses on predicting stock price change using a sentiment lexicon built from financial conference call logs. We introduce a method to generate a sentiment lexicon based upon an existing probabilistic approach. By using a domain-specific lexicon, we outperform traditional techniques and demonstrate that domain-specific sentiment lexicons provide higher accuracy than generic sentiment lexicons when predicting stock price change.

Age, Body Composition, Body Mass Index and Chronic Venous Diseases in Postmenopausal Women

Chronic venous diseases (CVD) are one of the common, though controversial problems in medicine. It is generally accepted that this pathology predominantly occurs in women. The issue of excessive weight as a risk factor for CVD is still considered debatable. To the author's best knowledge, today in Ukraine, there are barely any studies that describe the relationship between CVD and obesity. Our study aims to determine the association between age, body composition, obesity and CVD in postmenopausal women. The study was conducted in D. F. Chebotarev Institute of Gerontology, National Academy of Medical Sciences of Ukraine. We have examined 96 postmenopausal women aged 46-85 years (mean age – 66.19 ± 0.96 years), who were divided into two groups depending on the presence of CVD. The women were examined by vascular surgeons. For the diagnosis of CVD, we used clinical, anatomic and pathophysiologic classifications. We also performed clinical, ultrasound and densitometry examinations. We found that the CVD frequency in postmenopausal women increased with age (from 72% in those aged 45-59 years to 84% in those aged 75-89 years). A significant correlation between the total fat mass and age was determined in postmenopausal women with CVD. We also observed a significant correlation between the lower extremities’ fat mass and age in both examined groups. A significant correlation between body mass index and age was determined only in postmenopausal women without CVD.

Rank-Based Chain-Mode Ensemble for Binary Classification

In the field of machine learning, the ensemble has been employed as a common methodology to improve the performance upon multiple base classifiers. However, the true predictions are often canceled out by the false ones during consensus due to a phenomenon called “curse of correlation” which is represented as the strong interferences among the predictions produced by the base classifiers. In addition, the existing practices are still not able to effectively mitigate the problem of imbalanced classification. Based on the analysis on our experiment results, we conclude that the two problems are caused by some inherent deficiencies in the approach of consensus. Therefore, we create an enhanced ensemble algorithm which adopts a designed rank-based chain-mode consensus to overcome the two problems. In order to evaluate the proposed ensemble algorithm, we employ a well-known benchmark data set NSL-KDD (the improved version of dataset KDDCup99 produced by University of New Brunswick) to make comparisons between the proposed and 8 common ensemble algorithms. Particularly, each compared ensemble classifier uses the same 22 base classifiers, so that the differences in terms of the improvements toward the accuracy and reliability upon the base classifiers can be truly revealed. As a result, the proposed rank-based chain-mode consensus is proved to be a more effective ensemble solution than the traditional consensus approach, which outperforms the 8 ensemble algorithms by 20% on almost all compared metrices which include accuracy, precision, recall, F1-score and area under receiver operating characteristic curve.

Design for Classroom Units: A Collaborative Multicultural Studio Development with Chinese Students

In this paper, we present the main results achieved during a five-week international workshop on Interactive Furniture for the Classroom, with 22 Chinese design students, in Jiangmen city (Guangdong province, China), and five teachers from Portugal, France, Iran, Macao SAR, and China. The main goal was to engage design students from China with new skills and practice methodologies towards interactive design research for furniture and product design for the classroom. The final results demonstrate students' concerns on improving Chinese furniture design for the classrooms, including solutions related to collaborative learning and human-interaction design for interactive furniture products. The findings of the research led students to the fabrication of five original prototypes: two for kindergartens ('Candy' and 'Tilt-tilt'), two for primary schools ('Closer' and 'Eks(x)'), and one for art/creative schools ('Wave'). From the findings, it was also clear that collaboration, personalization, and project-based teaching are still neglected when designing furniture products for the classroom in China. Students focused on these issues and came up with creative solutions that could transform this educational field in China.

Finite Element Analysis of Thermally-Induced Bistable Plate Using Four Plate Elements

The present study deals with the finite element (FE) analysis of thermally-induced bistable plate using various plate elements. The quadrilateral plate elements include the 4-node conforming plate element based on the classical laminate plate theory (CLPT), the 4-node and 9-node Mindlin plate element based on the first-order shear deformation laminated plate theory (FSDT), and a displacement-based 4-node quadrilateral element (RDKQ-NL20). Using the von-Karman’s large deflection theory and the total Lagrangian (TL) approach, the nonlinear FE governing equations for plate under thermal load are derived. Convergence analysis for four elements is first conducted. These elements are then used to predict the stable shapes of thermally-induced bistable plate. Numerical test shows that the plate element based on FSDT, namely the 4-node and 9-node Mindlin, and the RDKQ-NL20 plate element can predict two stable cylindrical shapes while the 4-node conforming plate predicts a saddles shape. Comparing the simulation results with ABAQUS, the RDKQ-NL20 element shows the best accuracy among all the elements.

Thermoplastic Composites with Reduced Discoloration and Enhanced Fire-Retardant Property

This paper discusses a light-weight reinforced thermoplastic (LWRT) composite with superior fire retardancy. This porous LWRT composite is manufactured using polyolefin, fiberglass, and fire retardant additives via a wet-lay process. However, discoloration of the LWRT can be induced by various mechanisms, which may be a concern in the building and construction industry. It is commonly understood that discoloration is strongly associated with the presence of phenolic antioxidant(s) and NOx. The over-oxidation of phenolic antioxidant(s) is probably the root-cause of the discoloration (pinking/yellowing). Hanwha Azdel, Inc. developed a LWRT with fire-retardant property of ASTM E84-Class A specification, as well as negligible discoloration even under harsh conditions. In addition, this thermoplastic material is suitable for secondary processing (e.g. compression molding) if necessary.

A Corpus-Based Study on the Styles of Three Translators

The present paper is preoccupied with the different styles of three translators in their translating a Chinese classical novel Shuihu Zhuan. Based on a parallel corpus, it adopts a target-oriented approach to look into whether and what stylistic differences and shifts the three translations have revealed. The findings show that the three translators demonstrate different styles concerning their word choices and sentence preferences, which implies that identification of recurrent textual patterns may be a basic step for investigating the style of a translator.

Evaluating the Effectiveness of Electronic Response Systems in Technology-Oriented Classes

Electronic Response Systems such as Kahoot, Poll Everywhere, and Google Classroom are gaining a lot of popularity when surveying audiences in events, meetings, and classroom. The reason is mainly because of the ease of use and the convenience these tools bring since they provide mobile applications with a simple user interface. In this paper, we present a case study on the effectiveness of using Electronic Response Systems on student participation and learning experience in a classroom. We use a polling application for class exercises in two different technology-oriented classes. We evaluate the effectiveness of the usage of the polling applications through statistical analysis of the students performance in these two classes and compare them to the performances of students who took the same classes without using the polling application for class participation. Our results show an increase in the performances of the students who used the Electronic Response System when compared to those who did not by an average of 11%.

Government (Big) Data Ecosystem: Definition, Classification of Actors, and Their Roles

Organizations, including governments, generate (big) data that are high in volume, velocity, veracity, and come from a variety of sources. Public Administrations are using (big) data, implementing base registries, and enforcing data sharing within the entire government to deliver (big) data related integrated services, provision of insights to users, and for good governance. Government (Big) data ecosystem actors represent distinct entities that provide data, consume data, manipulate data to offer paid services, and extend data services like data storage, hosting services to other actors. In this research work, we perform a systematic literature review. The key objectives of this paper are to propose a robust definition of government (big) data ecosystem and a classification of government (big) data ecosystem actors and their roles. We showcase a graphical view of actors, roles, and their relationship in the government (big) data ecosystem. We also discuss our research findings. We did not find too much published research articles about the government (big) data ecosystem, including its definition and classification of actors and their roles. Therefore, we lent ideas for the government (big) data ecosystem from numerous areas that include scientific research data, humanitarian data, open government data, industry data, in the literature.

Blended Learning through Google Classroom

This paper discusses that good learning involves all academic groups in the school. Blended learning is learning outside the classroom. Google Classroom is a free service learning app for schools, non-profit organizations and anyone with a personal Google account. Facilities accessed through computers and mobile phones are very useful for school teachers and students. Blended learning classrooms using both traditional and technology-based methods for teaching have become the norm for many educators. Using Google Classroom gives students access to online learning. Even if the teacher is not in the classroom, the teacher can provide learning. This is the supervision of the form of the teacher when the student is outside the school.

Analytical Authentication of Butter Using Fourier Transform Infrared Spectroscopy Coupled with Chemometrics

Fourier Transform Infrared (FT-IR) spectroscopy coupled with chemometrics was used to distinguish between butter samples and non-butter samples. Further, quantification of the content of margarine in adulterated butter samples was investigated. Fingerprinting region (1400-800 cm–1) was used to develop unsupervised pattern recognition (Principal Component Analysis, PCA), supervised modeling (Soft Independent Modelling by Class Analogy, SIMCA), classification (Partial Least Squares Discriminant Analysis, PLS-DA) and regression (Partial Least Squares Regression, PLS-R) models. PCA of the fingerprinting region shows a clustering of the two sample types. All samples were classified in their rightful class by SIMCA approach; however, nine adulterated samples (between 1% and 30% w/w of margarine) were classified as belonging both at the butter class and at the non-butter one. In the two-class PLS-DA model’s (R2 = 0.73, RMSEP, Root Mean Square Error of Prediction = 0.26% w/w) sensitivity was 71.4% and Positive Predictive Value (PPV) 100%. Its threshold was calculated at 7% w/w of margarine in adulterated butter samples. Finally, PLS-R model (R2 = 0.84, RMSEP = 16.54%) was developed. PLS-DA was a suitable classification tool and PLS-R a proper quantification approach. Results demonstrate that FT-IR spectroscopy combined with PLS-R can be used as a rapid, simple and safe method to identify pure butter samples from adulterated ones and to determine the grade of adulteration of margarine in butter samples.

Upgraded Cuckoo Search Algorithm to Solve Optimisation Problems Using Gaussian Selection Operator and Neighbour Strategy Approach

An Upgraded Cuckoo Search Algorithm is proposed here to solve optimization problems based on the improvements made in the earlier versions of Cuckoo Search Algorithm. Short comings of the earlier versions like slow convergence, trap in local optima improved in the proposed version by random initialization of solution by suggesting an Improved Lambda Iteration Relaxation method, Random Gaussian Distribution Walk to improve local search and further proposing Greedy Selection to accelerate to optimized solution quickly and by “Study Nearby Strategy” to improve global search performance by avoiding trapping to local optima. It is further proposed to generate better solution by Crossover Operation. The proposed strategy used in algorithm shows superiority in terms of high convergence speed over several classical algorithms. Three standard algorithms were tested on a 6-generator standard test system and the results are presented which clearly demonstrate its superiority over other established algorithms. The algorithm is also capable of handling higher unit systems.

Building a Transformative Continuing Professional Development Experience for Educators through a Principle-Based, Technological-Driven Knowledge Building Approach: A Case Study of a Professional Learning Team in Secondary Education

There has been a growing emphasis in elevating the teachers’ proficiency and competencies through continuing professional development (CPD) opportunities. In this era of a Volatile, Uncertain, Complex, Ambiguous (VUCA) world, teachers are expected to be collaborative designers, critical thinkers and creative builders. However, many of the CPD structures are still revolving in the model of transmission, which stands in contradiction to the cultivation of future-ready teachers for the innovative world of emerging technologies. This article puts forward the framing of CPD through a Principle-Based, Technological-Driven Knowledge Building Approach grounded in the essence of andragogy and progressive learning theories where growth is best exemplified through an authentic immersion in a social/community experience-based setting. Putting this Knowledge Building Professional Development Model (KBPDM) in operation via a Professional Learning Team (PLT) situated in a Secondary School in Singapore, research findings reveal that the intervention has led to a fundamental change in the learning paradigm of the teachers, henceforth equipping and empowering them successfully in their pedagogical design and practices for a 21st century classroom experience. This article concludes with the possibility in leveraging the Learning Analytics to deepen the CPD experiences for educators.

A Hybrid Feature Selection and Deep Learning Algorithm for Cancer Disease Classification

Learning from very big datasets is a significant problem for most present data mining and machine learning algorithms. MicroRNA (miRNA) is one of the important big genomic and non-coding datasets presenting the genome sequences. In this paper, a hybrid method for the classification of the miRNA data is proposed. Due to the variety of cancers and high number of genes, analyzing the miRNA dataset has been a challenging problem for researchers. The number of features corresponding to the number of samples is high and the data suffer from being imbalanced. The feature selection method has been used to select features having more ability to distinguish classes and eliminating obscures features. Afterward, a Convolutional Neural Network (CNN) classifier for classification of cancer types is utilized, which employs a Genetic Algorithm to highlight optimized hyper-parameters of CNN. In order to make the process of classification by CNN faster, Graphics Processing Unit (GPU) is recommended for calculating the mathematic equation in a parallel way. The proposed method is tested on a real-world dataset with 8,129 patients, 29 different types of tumors, and 1,046 miRNA biomarkers, taken from The Cancer Genome Atlas (TCGA) database.

Determination of Post-Failure Characteristic Behaviour of Rocks under Conventional Method Based on the Mechanism of Rock Deformation Process

This work is intended to study the post-failure characteristic behaviour of rocks and the techniques of controlling the post-failure regime based on the mechanism of rocks deformation process. It is impossible to determine the post-failure regime of rocks using conventional laboratory testing equipment. This is because most testing machines are soft and therefore no information can be obtained after the peak load. Stress-strain deformation tests were conducted using both conventional and unconventional method (i.e. the closed loop servo-controlled testing machine) in accordance to ISRM standard. Normalised pre-failure curves were constructed to show the stages in the deformation process. The first type contains the Class I and progress to Class II with low strength soft brittle rocks. The second type shows entirely Class II characteristic behaviour. The third type is extremely brittle under axial loading, resulted in explosive failure, so its class could not be determined. The difficulty in obtaining the post-failure curves increases as the total volumetric strain approaches a positive value. The author’s use of normalised pre-failure curves enables identification of additional type of deformation process with very brittle response under axial loading. Testing the third type without confinement could cause equipment damage. Identification of the deformation process with the rock classes using conventional test could guide the personnel conducting tests using closed-loop servo-controlled system, to avoid equipment damage when testing rocks with third type deformation process so that testing is performed safely. It has also improved our understanding on total specimen failure and brittleness of rocks (e.g. brittle for Class II and less brittle or ductile for Class I).

Associations between Surrogate Insulin Resistance Indices and the Risk of Metabolic Syndrome in Children

A well-defined insulin resistance (IR) is one of the requirements for the good understanding and evaluation of metabolic syndrome (MetS). However, underlying causes for the development of IR are not clear. Endothelial dysfunction also participates in the pathogenesis of this disease. IR indices are being determined in various obesity groups and also in diagnosing MetS. Components of MetS have been well established and used in adult studies. However, there are some ambiguities particularly in the field of pediatrics. The aims of this study were to compare the performance of fasting blood glucose (FBG), one of MetS components, with some other IR indices and check whether FBG may be replaced by some other parameter or ratio for a better evaluation of pediatric MetS. Five-hundred and forty-nine children were involved in the study. Five groups were constituted. Groups 109, 40, 100, 166, 110, 24 children were included in normal-body mass index (N-BMI), overweight (OW), obese (OB), morbid obese (MO), MetS with two components (MetS2) and MetS with three components (MetS3) groups, respectively. Age and sex-adjusted BMI percentiles tabulated by World Health Organization were used for the classification of obesity groups. MetS components were determined. Aside from one of the MetS components-FBG, eight measures of IR [homeostatic model assessment of IR (HOMA-IR), homeostatic model assessment of beta cell function (HOMA-%β), alanine transaminase-to-aspartate transaminase ratio (ALT/AST), alanine transaminase (ALT), insulin (INS), insulin-to-FBG ratio (INS/FBG), the product of fasting triglyceride and glucose (TyG) index, McAuley index] were evaluated. Statistical analyses were performed. A p value less than 0.05 was accepted as the statistically significance degree. Mean values for BMI of the groups were 15.7 kg/m2, 21.0 kg/m2, 24.7 kg/m2, 27.1 kg/m2, 28.7 kg/m2, 30.4 kg/m2 for N-BMI, OW, OB, MO, MetS2, MetS3, respectively. Differences between the groups were significant (p < 0.001). The only exception was MetS2-MetS3 couple, in spite of an increase detected in MetS3 group. Waist-to-hip circumference ratios significantly differed only for N-BMI vs, OB, MO, MetS2; OW vs MO; OB vs MO, MetS2 couples. ALT and ALT/AST did not differ significantly among MO-MetS2-MetS3. HOMA-%β differed only between MO and MetS2. INS/FBG, McAuley index and TyG were not significant between MetS2 and MetS3. HOMA-IR and FBG were not significant between MO and MetS2. INS was the only parameter, which showed statistically significant differences between MO-MetS2, MO-MetS3, and MetS2-MetS3. In conclusion, these findings have suggested that FBG presently considered as one of the five MetS components, may be replaced by INS during the evaluation of pediatric morbid obesity and MetS.