Comparative Forensic Analysis of Lipsticks Using Thin Layer Chromatography and Gas Chromatography

Lipsticks constitute a significant source of transfer evidence, and can, therefore, provide corroborative or inclusionary evidence in criminal investigation. This study aimed to determine the uniqueness and persistence of different lipstick smears using Thin Layer Chromatography (TLC), and Gas Chromatography with a Flame Ionisation Detector (GC-FID). In this study, we analysed lipstick smears retrieved from tea cups exposed to the environment for up to four weeks. The n-alkane content of each sample was determined using GC-FID, while TLC was used to determine the number of bands, and retention factor of each band per smear. This study shows that TLC gives more consistent results over a 4-week period than GC-FID. It also proposes a maximum exposure time of two weeks for the analysis of lipsticks left in the open using GC-FID. Finally, we conclude that neither TLC nor GC-FID can distinguish lipstick evidence recovered from hypothetical crime scenes.

Feasibility Study of Friction Stir Welding Application for Kevlar Material

Friction stir welding (FSW) is a joining process in the solid state, which eliminates problems associated with the material melting and solidification, such as cracks, residual stresses and distortions generated during conventional welding. Among the most important advantages of FSW are; easy automation, less distortion, lower residual stress and good mechanical properties in the joining region. FSW is a recent approach to metal joining and although originally intended for aluminum alloys, it is investigated in a variety of metallic materials. The basic concept of FSW is a rotating tool, made of non-consumable material, specially designed with a geometry consisting of a pin and a recess (shoulder). This tool is inserted as spinning on its axis at the adjoining edges of two sheets or plates to be joined and then it travels along the joining path line. The tool rotation axis defines an angle of inclination with which the components to be welded. This angle is used for receiving the material to be processed at the tool base and to promote the gradual forge effect imposed by the shoulder during the passage of the tool. This prevents the material plastic flow at the tool lateral, ensuring weld closure on the back of the pin. In this study, two 4 mm Kevlar® plates which were produced with the Kevlar® fabrics, are analyzed with COMSOL Multiphysics in order to investigate the weldability via FSW. Thereafter, some experimental investigation is done with an appropriate workbench in order to compare them with the analysis results.

Changes in Student Definition of De-Escalation in Professional Peace Officer Education

Since the release of the 21st century policing report in the United States, the techniques of de-escalation have received a lot of attention and focus in political systems, policy changes, and the media. The challenge in professional peace officer education is that there is a vast range of defining de-escalation and understanding the various techniques involved, many of which are based on popular media. This research surveyed professional peace officer education university students on their definition of de-escalation and the techniques associated with de-escalation before specific communications coursework was completed. The students were then surveyed after the communication coursework was completed to determine the changes in defining and understanding de-escalation techniques. This research has found that clearly defining de-escalation and emphasizing the broad range of techniques available enhances the students’ understanding and application of proper de-escalation. This research demonstrates the need for professional peace officer education to move students from media concepts of law enforcement to theoretical concepts.

Measuring Text-Based Semantics Relatedness Using WordNet

Measuring semantic similarity between texts is calculating semantic relatedness between texts using various techniques. Our web application (Measuring Relatedness of Concepts-MRC) allows user to input two text corpuses and get semantic similarity percentage between both using WordNet. Our application goes through five stages for the computation of semantic relatedness. Those stages are: Preprocessing (extracts keywords from content), Feature Extraction (classification of words into Parts-of-Speech), Synonyms Extraction (retrieves synonyms against each keyword), Measuring Similarity (using keywords and synonyms, similarity is measured) and Visualization (graphical representation of similarity measure). Hence the user can measure similarity on basis of features as well. The end result is a percentage score and the word(s) which form the basis of similarity between both texts with use of different tools on same platform. In future work we look forward for a Web as a live corpus application that provides a simpler and user friendly tool to compare documents and extract useful information.

Measuring Text-Based Semantics Relatedness Using WordNet

Measuring semantic similarity between texts is calculating semantic relatedness between texts using various techniques. Our web application (Measuring Relatedness of Concepts-MRC) allows user to input two text corpuses and get semantic similarity percentage between both using WordNet. Our application goes through five stages for the computation of semantic relatedness. Those stages are: Preprocessing (extracts keywords from content), Feature Extraction (classification of words into Parts-of-Speech), Synonyms Extraction (retrieves synonyms against each keyword), Measuring Similarity (using keywords and synonyms, similarity is measured) and Visualization (graphical representation of similarity measure). Hence the user can measure similarity on basis of features as well. The end result is a percentage score and the word(s) which form the basis of similarity between both texts with use of different tools on same platform. In future work we look forward for a Web as a live corpus application that provides a simpler and user friendly tool to compare documents and extract useful information.

Technology, Organizational and Environmental Determinants of Business Intelligence Systems Adoption in Croatian SME: A Case Study of Medium-Sized Enterprise

In the last few years, examples from scientific literature and business practices show that the adoption of technological innovations increases enterprises' performance. Recently, when it comes to the field of information technology innovation, business intelligence systems (BISs) have drawn a significant amount of attention of the scientific circles. BISs can be understood as a form of technological innovation which can bring certain benefits to the organizations that are adopting it. Therefore, the aim of this paper is twofold: (1) to define determinants of successful BISs adoption in small and medium enterprises and thus contribute to this neglected research area and (2) to present the current state of BISs adoption in small and medium-sized companies. In order to do so, determinants are defined and classified into three dimensions, according to the Technology – Organization – Environment (TOE) theoretical framework that describes the impact of each dimension on technological innovations adoption. Moreover, paper brings a case study presenting the adoption of BISs in practice within an organization from tertiary (service) industry sector. Based on the results of the study, guidelines for more efficient, faster and easier BISs adoption are presented.

Managing Business Processes in the Age of Digital Transformation: A Literature Review

Today, digital transformation is one of the leading topics that occupy the attention of scientific circles and business experts. Organizational success is most often reflected through the successful managing of business processes. Given the growing market for digital innovations and its ever-increasing impact on business, organizations need to be prepared for organizational changes that come with the digital era. In order to maintain their competitive advantage in the global market, organizations must adapt their processes to new digitalization conditions. The main goal of this study is to point out the link between the digital transformation and the business process management concept. Therefore, in order to contribute to the scientific field that explores the potential relation between business process management concept and digital transformation, a literature review has been conducted. Papers have been searched within the Business Process Management Journal by keywords related to the term digital transformation. Selected papers have been analyzed according to the topic, type of publication, year of publication, keywords, etc. The results reveal a growing number of papers published on the topic of digital transformation to the Business Process Management Journal, but the lack of case studies. This paper contributes to the extension of academic literature in this important, yet insufficiently researched, scientific field that creates the bond between two strong concepts of digital transformation and business process management.

On the Efficiency of Five Step Approximation Method for the Solution of General Third Order Ordinary Differential Equations

In this work, a five step continuous method for the solution of third order ordinary differential equations was developed in block form using collocation and interpolation techniques of the shifted Legendre polynomial basis function. The method was found to be zero-stable, consistent and convergent. The application of the method in solving third order initial value problem of ordinary differential equations revealed that the method compared favorably with existing methods.

Perceived Determinants of Obesity among Primary School Pupils in Eti Osa Local Government Area of Lagos State, Nigeria

Children in today’s world need attention and care even with their physique as obesity is also at the increased. Several factors can be responsible for obesity in children and adequate attention is paramount in other not to accommodate it into adolescent period. This study investigated perceived determinants of obesity among primary school pupils in Eti Osa Local Government area of Lagos State. Descriptive survey research design was used and population was all obese pupils in Eti Osa Local Government Area of Lagos State. 92 pupils were selected from randomly picked 12 primary schools while purposive sampling technique was used to pick primary 4-6 pupils. With the aid of body mass index (BMI) and age percentile chart the obese pupils were selected. The instrument for the study was a self-developed and structured questionnaire on perceived determinant of obesity. The questionnaire was divided into three sections. The Cronbach’s Alpha reliability coefficient of 0.74 was obtained. The hypotheses were tested at 0.05 significant levels. The completed questionnaire was collated coded and analyzed using descriptive statistics of frequency counts and percentage and inferential statistics of chi-square (X2). Findings of this study revealed that physical activities and parental influences were determinant of obesity. Physical activity is essential in reducing the rate of obesity in Eti Osa Local Government Area both at home and within the school environment. Primary schools need to create more playing ground for pupils to exercise themselves. Parents need to cater for their children diet ensuring not just the quantity but the quality as well.

Use of Hierarchical Temporal Memory Algorithm in Heart Attack Detection

In order to reduce the number of deaths due to heart problems, we propose the use of Hierarchical Temporal Memory Algorithm (HTM) which is a real time anomaly detection algorithm. HTM is a cortical learning algorithm based on neocortex used for anomaly detection. In other words, it is based on a conceptual theory of how the human brain can work. It is powerful in predicting unusual patterns, anomaly detection and classification. In this paper, HTM have been implemented and tested on ECG datasets in order to detect cardiac anomalies. Experiments showed good performance in terms of specificity, sensitivity and execution time.

An Approach towards Designing an Energy Efficient Building through Embodied Energy Assessment: A Case of Apartment Building in Composite Climate

In today’s world, the growing demand for urban built forms has resulted in the production and consumption of building materials i.e. embodied energy in building construction, leading to pollution and greenhouse gas (GHG) emissions. Therefore, new buildings will offer a unique opportunity to implement more energy efficient building without compromising on building performance of the building. Embodied energy of building materials forms major contribution to embodied energy in buildings. The paper results in an approach towards designing an energy efficient apartment building through embodied energy assessment. This paper discusses the trend of residential development in Rourkela, which includes three case studies of the contemporary houses, followed by architectural elements, number of storeys, predominant material use and plot sizes using primary data. It results in identification of predominant material used and other characteristics in urban area. Further, the embodied energy coefficients of various dominant building materials and alternative materials manufactured in Indian Industry is taken in consideration from secondary source i.e. literature study. The paper analyses the embodied energy by estimating materials and operational energy of proposed building followed by altering the specifications of the materials based on the building components i.e. walls, flooring, windows, insulation and roof through res build India software and comparison of different options is assessed with consideration of sustainable parameters. This paper results that autoclaved aerated concrete block only reaches the energy performance Index benchmark i.e. 69.35 kWh/m2 yr i.e. by saving 4% of operational energy and as embodied energy has no particular index, out of all materials it has the highest EE 23206202.43  MJ.

Evaluation of Storage Stability and Quality Parameters in Biscuit Made from Blends of Wheat, Cassava (Manihot esculenta) and Carrot (Daucus carota) Flour

Biscuit is one of the most consumed cereal foods in Nigeria and research has shown that locally available tropical crops like cassava, sweet potato can be made into flour and used in the production of biscuits and other pastries. This study investigates some quality parameters in biscuits made from blends of wheat, cassava and carrot flour. The values of result of samples increased with increasing percentage substitution of cassava and carrot flour in some quality parameter like fiber, ash, gluten content, and carbohydrate. The protein content reduced significantly (P < 0.05) with increasing percentage substitution of cassava and carrot flour which ranged from 14.80% to 11.80% compared with the control sample which had 15.60%. There was a recorded significant increase (P < 0.05) in some mineral composition such as calcium, magnesium, sodium, iron, phosphorus, and vitamin A and C composition as the percentage substitution of cassava and carrot flour increased. During storage stability test, samples stored in the fridge and freezer were found to be the best storage location to preserve the sensory attributes and inhibit microbial growth when compared with storage under the sun and on the shelf. Biscuit made with blends of wheat, cassava and carrot flour can therefore serve as an alternative to biscuits made from 100% wheat flour, as they are richer in vitamin A, vitamin C, carbohydrate, dietary fiber and some essential minerals.

Mixed Mode Fracture Analyses Using Finite Element Method of Edge Cracked Heavy Spinning Annulus Pulley

Rotating disk is one of the most indispensable parts of a rotating machine. Rotating disk has found many applications in the diverging field of science and technology. In this paper, we have taken into consideration the problem of a heavy spinning disk mounted on a rotor system acted upon by boundary traction. Finite element modelling is used at various loading condition to determine the mixed mode stress intensity factors. The effect of combined shear and normal traction on the boundary is incorporated in the analysis under the action of gravity. The variation near the crack tip is characterized in terms of the stress intensity factor (SIF) with an aim to find the SIF for a wide range of parameters. The results of the finite element analyses carried out on the compressed disk of a belt pulley arrangement using fracture mechanics concepts are shown. A total of hundred cases of the problem are solved for each of the variations in loading arc parameter and crack orientation using finite element models of the disc under compression. All models were prepared and analyzed for the uncracked disk, disk with a single crack at different orientation emanating from shaft hole as well as for a disc with pair of cracks emerging from the same center hole. Curves are plotted for various loading conditions. Finally, crack propagation paths are determined using kink angle concepts.

Waste Management in a Hot Laboratory of Japan Atomic Energy Agency – 1: Overview and Activities in Chemical Processing Facility

Chemical Processing Facility of Japan Atomic Energy Agency is a basic research field for advanced back-end technology developments with using actual high-level radioactive materials such as irradiated fuels from the fast reactor, high-level liquid waste from reprocessing plant. In the nature of a research facility, various kinds of chemical reagents have been offered for fundamental tests. Most of them were treated properly and stored in the liquid waste vessel equipped in the facility, but some were not treated and remained at the experimental space as a kind of legacy waste. It is required to treat the waste in safety. On the other hand, we formulated the Medium- and Long-Term Management Plan of Japan Atomic Energy Agency Facilities. This comprehensive plan considers Chemical Processing Facility as one of the facilities to be decommissioned. Even if the plan is executed, treatment of the “legacy” waste beforehand must be a necessary step for decommissioning operation. Under this circumstance, we launched a collaborative research project called the STRAD project, which stands for Systematic Treatment of Radioactive liquid waste for Decommissioning, in order to develop the treatment processes for wastes of the nuclear research facility. In this project, decomposition methods of chemicals causing a troublesome phenomenon such as corrosion and explosion have been developed and there is a prospect of their decomposition in the facility by simple method. And solidification of aqueous or organic liquid wastes after the decomposition has been studied by adding cement or coagulants. Furthermore, we treated experimental tools of various materials with making an effort to stabilize and to compact them before the package into the waste container. It is expected to decrease the number of transportation of the solid waste and widen the operation space. Some achievements of these studies will be shown in this paper. The project is expected to contribute beneficial waste management outcome that can be shared world widely.

Choice Experiment Approach on Evaluation of Non-Market Farming System Outputs: First Results from Lithuanian Case Study

Market and non-market outputs are produced jointly in agriculture. Their supply depends on the intensity and type of production. The role of agriculture as an economic activity and its effects are important for the Lithuanian case study, as agricultural land covers more than a half of country. Positive and negative externalities, created in agriculture are not considered in the market. Therefore, specific techniques such as stated preferences methods, in particular choice experiments (CE) are used for evaluation of non-market outputs in agriculture. The main aim of this paper is to present construction of the research path for evaluation of non-market farming system outputs in Lithuania. The conventional and organic farming, covering crops (including both cereal and industrial crops) and livestock (including dairy and cattle) production has been selected. The CE method and nested logit (NL) model were selected as appropriate for evaluation of non-market outputs of different farming systems in Lithuania. A pilot survey was implemented between October–November 2018, in order to test and improve the CE questionnaire. The results of the survey showed that the questionnaire is accepted and well understood by the respondents. The econometric modelling showed that the selected NL model could be used for the main survey. The understanding of the differences between organic and conventional farming by residents was identified. It was revealed that they are more willing to choose organic farming in comparison to conventional farming.

Semantic Preference across Research Articles: A Corpus-Based Study of Adjectives in English

The goal of the present study is to investigate the semantic preference of the most frequent adjectives in research articles through a corpus-based analysis of texts published in journals in Applied Linguistics (AL). The corpus used in this study contains texts published in the period from 2014 to 2018 in the three journals: Language Learning and Technology; English for Academic Purposes, and TESOL Quaterly, totaling more than one million words. A corpus-based analysis was carried out on the corpus to identify the most frequent adjectives that co-occurred in the three journals. By observing the concordance lines of the adjectives and analyzing the words they associated with, the semantic preferences of each adjective were determined. Later, the AL corpus analysis was compared to the investigation of the same adjectives in a corpus of Chemistry. This second part of the study aimed to identify possible differences and similarities between the two corpora in relation to the use of the adjectives in research articles from both areas. The results show that there are some preferences which seem to be closely related not only to the academic genre of the texts but also to the specific domain of the discipline and, to a lesser extent, to the context of research in each journal. This research illustrates a possible contribution of Corpus Linguistics to explore the concept of semantic preference in more detail, considering the complex nature of the phenomenon.

Optimization of a Bioremediation Strategy for an Urban Stream of Matanza-Riachuelo Basin

In the present work, a remediation bioprocess based on the use of a local isolate of the microalgae Chlorella vulgaris immobilized in alginate beads is proposed. This process was shown to be effective for the reduction of several chemical and microbial contaminants present in Cildáñez stream, a water course that is part of the Matanza-Riachuelo Basin (Buenos Aires, Argentina). The bioprocess, involving the culture of the microalga in autotrophic conditions in a stirred-tank bioreactor supplied with a marine propeller for 6 days, allowed a significant reduction of Escherichia coli and total coliform numbers (over 95%), as well as of ammoniacal nitrogen (96%), nitrates (86%), nitrites (98%), and total phosphorus (53%) contents. Pb content was also significantly diminished after the bioprocess (95%). Standardized cytotoxicity tests using Allium cepa seeds and Cildáñez water pre- and post-remediation were also performed. Germination rate and mitotic index of onion seeds imbibed in Cildáñez water subjected to the bioprocess was similar to that observed in seeds imbibed in distilled water and significantly superior to that registered when untreated Cildáñez water was used for imbibition. Our results demonstrate the potential of this simple and cost-effective technology to remove urban-water contaminants, offering as an additional advantage the possibility of an easy biomass recovery, which may become a source of alternative energy.

Predictive Semi-Empirical NOx Model for Diesel Engine

Accurate prediction of NOx emission is a continuous challenge in the field of diesel engine-out emission modeling. Performing experiments for each conditions and scenario cost significant amount of money and man hours, therefore model-based development strategy has been implemented in order to solve that issue. NOx formation is highly dependent on the burn gas temperature and the O2 concentration inside the cylinder. The current empirical models are developed by calibrating the parameters representing the engine operating conditions with respect to the measured NOx. This makes the prediction of purely empirical models limited to the region where it has been calibrated. An alternative solution to that is presented in this paper, which focus on the utilization of in-cylinder combustion parameters to form a predictive semi-empirical NOx model. The result of this work is shown by developing a fast and predictive NOx model by using the physical parameters and empirical correlation. The model is developed based on the steady state data collected at entire operating region of the engine and the predictive combustion model, which is developed in Gamma Technology (GT)-Power by using Direct Injected (DI)-Pulse combustion object. In this approach, temperature in both burned and unburnt zone is considered during the combustion period i.e. from Intake Valve Closing (IVC) to Exhaust Valve Opening (EVO). Also, the oxygen concentration consumed in burnt zone and trapped fuel mass is also considered while developing the reported model.  Several statistical methods are used to construct the model, including individual machine learning methods and ensemble machine learning methods. A detailed validation of the model on multiple diesel engines is reported in this work. Substantial numbers of cases are tested for different engine configurations over a large span of speed and load points. Different sweeps of operating conditions such as Exhaust Gas Recirculation (EGR), injection timing and Variable Valve Timing (VVT) are also considered for the validation. Model shows a very good predictability and robustness at both sea level and altitude condition with different ambient conditions. The various advantages such as high accuracy and robustness at different operating conditions, low computational time and lower number of data points requires for the calibration establishes the platform where the model-based approach can be used for the engine calibration and development process. Moreover, the focus of this work is towards establishing a framework for the future model development for other various targets such as soot, Combustion Noise Level (CNL), NO2/NOx ratio etc.

A Unification and Relativistic Correction for Boltzmann’s Law

The distribution of velocities of particles in plasma is a well understood discipline of plasma physics. Boltzmann’s law and the Maxwell-Boltzmann distribution describe the distribution of velocity of a particle in plasma as a function of mass and temperature. Particles with the same mass tend to have the same velocity. By expressing the same law in terms of energy alone, the author obtains a distribution independent of mass. In summary, for particles in plasma, the energies tend to equalize, independent of the masses of the individual particles. For high-energy plasma, the original law predicts velocities greater than the speed of light. If one uses Einstein’s formula for energy (E=mc2), then a relativistic correction is not required.

Estimation of Geotechnical Parameters by Comparing Monitoring Data with Numerical Results: Case Study of Arash–Esfandiar-Niayesh Under-Passing Tunnel, Africa Tunnel, Tehran, Iran

The under passing tunnels are strongly influenced by the soils around. There are some complexities in the specification of real soil behavior, owing to the fact that lots of uncertainties exist in soil properties, and additionally, inappropriate soil constitutive models. Such mentioned factors may cause incompatible settlements in numerical analysis with the obtained values in actual construction. This paper aims to report a case study on a specific tunnel constructed by NATM. The tunnel has a depth of 11.4 m, height of 12.2 m, and width of 14.4 m with 2.5 lanes. The numerical modeling was based on a 2D finite element program. The soil material behavior was modeled by hardening soil model. According to the field observations, the numerical estimated settlement at the ground surface was approximately four times more than the measured one, after the entire installation of the initial lining, indicating that some unknown factors affect the values. Consequently, the geotechnical parameters are accurately revised by a numerical back-analysis using laboratory and field test data and based on the obtained monitoring data. The obtained result confirms that typically, the soil parameters are conservatively low-estimated. And additionally, the constitutive models cannot be applied properly for all soil conditions.