A Note on MHD Flow and Heat Transfer over a Curved Stretching Sheet by Considering Variable Thermal Conductivity

The mixed convective flow of MHD incompressible, steady boundary layer in heat transfer over a curved stretching sheet due to temperature dependent thermal conductivity is studied. We use curvilinear coordinate system in order to describe the governing flow equations. Finite difference solutions with central differencing have been used to solve the transform governing equations. Numerical results for the flow velocity and temperature profiles are presented as a function of the non-dimensional curvature radius. Skin friction coefficient and local Nusselt number at the surface of the curved sheet are discussed as well.

Reliability Factors Based Fuzzy Logic Scheme for Spectrum Sensing

The accurate spectrum sensing is a fundamental requirement of dynamic spectrum access for deployment of Cognitive Radio Network (CRN). To acheive this requirement a Reliability factors based Fuzzy Logic (RFL) Scheme for Spectrum Sensing has been proposed in this paper. Cognitive Radio User (CRU) predicts the presence or absence of Primary User (PU) using energy detector and calculates the Reliability factors which are SNR of sensing node, threshold of energy detector and decision difference of each node with other nodes in a cooperative spectrum sensing environment. Then the decision of energy detector is combined with Reliability factors of sensing node using Fuzzy Logic. These Reliability Factors used in RFL Scheme describes the reliability of decision made by a CRU to improve the local spectrum sensing. This Fuzzy combining scheme provides the accuracy of decision made by sensornode. The simulation results have shown that the proposed technique provide better PU detection probability than existing Spectrum Sensing Techniques.

Teacher Culture Inquiry of Classroom Observation at an Elementary School in Taiwan

Three dimensions of teacher culture hinder educational improvement: individualism, conservatism and presentism. To promote the professional development of teachers, these three aspects in teacher culture should be eliminated. Classroom observation may be a useful method of eliminating individualism. The Ministry of Education in Taiwan has attempted to reduce the isolation of teachers to promote their professional growth. Because classroom observation discourse varies, teachers are generally unwilling to allow their teaching to be observed. However, classroom observations take place in the country in the form of school evaluations. The main purpose of this study was to explore the differences in teachers’ conservatism, individualism and presentism after classroom observations had been conducted at an elementary school in Taiwan. The research method was a qualitative case study involving interviews with the school principal, the director of academic affairs, and two classroom teachers. The following conclusions were drawn: (1) Educators in different positions viewed classroom observations differently; (2) The classroom teachers did not highly value classroom observation; (3) There was little change in the teachers’ conservatism, individualism and presentism after classroom observation.

Centrifuge Modeling of Monopiles Subjected to Lateral Monotonic Loading

The type of foundation commonly used today for berthing dolphins is a set of tubular steel piles with large diameters, which are known as monopiles. The design of these monopiles is based on the theories related with laterally loaded piles. One of the most common methods to analyze and design the piles subjected to lateral loads is the p-y curves. In the present study, centrifuge tests are conducted in order to obtain the p-y curves. Series of tests were designed in order to investigate the scaling laws in the centrifuge for monotonic loading. Also, two important parameters, the embedded depth L of the pile in the soil and free length e of the pile, as well as their ratios were studied via five experimental tests. Finally, the p-y curves of API are presented to be compared with the curves obtained from the tests so that the differences could be demonstrated. The results show that the p-y curves proposed by API highly overestimate the lateral load bearing capacity. It suggests that these curves need correction and modification for each site as the soil conditions change.

Analysis of Fertilizer Effect in the Tilapia Growth of Mozambique (Oreochromis mossambicus)

This paper analyses the effect of fertilizer (organic and inorganic) in the growth of tilapia. An experiment was implemented in the Aquapesca Company of Mozambique; there were considered four different treatments. Each type of fertilizer was applied in two of these treatments; a feed was supplied to the third treatment, and the fourth was taken as control. The weight and length of the tilapia were used as the growth parameters, and to measure the water quality, the physical-chemical parameters were registered. The results show that the weight and length were different for tilapias cultivated in different treatments. These differences were evidenced mainly by organic and feed treatments, where there was the largest and smallest value of these parameters, respectively. In order to prove that these differences were caused only by applied treatment without interference for the aquatic environment, a Fisher discriminant analysis was applied, which confirmed that the treatments were exposed to the same environment condition.

Conventional and Islamic Perspective in Accounting: Potential for Alternative Reporting Framework

This paper provides an overview of fundamental philosophical and functional differences in conventional and Islamic accounting. The aim of this research is to undertake a detailed analysis focus on specific illustrations drawn from both these systems and highlight how these differences implicate in recording financial transactions and preparation of financial reports for a range of stakeholders. Accounting as being universally considered as a platform for providing a ‘true and fair’ view of corporate entities can be challenged in the current world view, as the business environment has evolved and transformed significantly. Growth of the non-traditional corporate entity such as Islamic financial institutions, fundamentally questions the applicability of conventional accounting standards in preparation of Shariah-compliant financial reporting. Coupled with this, there are significant concerns about the wider applicability of Islamic accounting standards and framework in order to achieve reporting practices satisfying the information needs generally. Against the backdrop of such a context, this paper raises fundamental question as to how potential convergence could be achieved between these two systems in order to provide users’ a transparent and comparable state of financial information resulting in an alternative framework of financial reporting.

Determination of Safety Distance Around Gas Pipelines Using Numerical Methods

Energy transmission pipelines are one of the most vital parts of each country which several strict laws have been conducted to enhance the safety of these lines and their vicinity. One of these laws is the safety distance around high pressure gas pipelines. Safety distance refers to the minimum distance from the pipeline where people and equipment do not confront with serious damages. In the present study, safety distance around high pressure gas transmission pipelines were determined by using numerical methods. For this purpose, gas leakages from cracked pipeline and created jet fires were simulated as continuous ignition, three dimensional, unsteady and turbulent cases. Numerical simulations were based on finite volume method and turbulence of flow was considered using k-ω SST model. Also, the combustion of natural gas and air mixture was applied using the eddy dissipation method. The results show that, due to the high pressure difference between pipeline and environment, flow chocks in the cracked area and velocity of the exhausted gas reaches to sound speed. Also, analysis of the incident radiation results shows that safety distances around 42 inches high pressure natural gas pipeline based on 5 and 15 kW/m2 criteria are 205 and 272 meters, respectively.

Evaluating Machine Learning Techniques for Activity Classification in Smart Home Environments

With the widespread adoption of the Internet-connected devices, and with the prevalence of the Internet of Things (IoT) applications, there is an increased interest in machine learning techniques that can provide useful and interesting services in the smart home domain. The areas that machine learning techniques can help advance are varied and ever-evolving. Classifying smart home inhabitants’ Activities of Daily Living (ADLs), is one prominent example. The ability of machine learning technique to find meaningful spatio-temporal relations of high-dimensional data is an important requirement as well. This paper presents a comparative evaluation of state-of-the-art machine learning techniques to classify ADLs in the smart home domain. Forty-two synthetic datasets and two real-world datasets with multiple inhabitants are used to evaluate and compare the performance of the identified machine learning techniques. Our results show significant performance differences between the evaluated techniques. Such as AdaBoost, Cortical Learning Algorithm (CLA), Decision Trees, Hidden Markov Model (HMM), Multi-layer Perceptron (MLP), Structured Perceptron and Support Vector Machines (SVM). Overall, neural network based techniques have shown superiority over the other tested techniques.

An Improved Method to Compute Sparse Graphs for Traveling Salesman Problem

The Traveling salesman problem (TSP) is NP-hard in combinatorial optimization. The research shows the algorithms for TSP on the sparse graphs have the shorter computation time than those for TSP according to the complete graphs. We present an improved iterative algorithm to compute the sparse graphs for TSP by frequency graphs computed with frequency quadrilaterals. The iterative algorithm is enhanced by adjusting two parameters of the algorithm. The computation time of the algorithm is O(CNmaxn2) where C is the iterations, Nmax is the maximum number of frequency quadrilaterals containing each edge and n is the scale of TSP. The experimental results showed the computed sparse graphs generally have less than 5n edges for most of these Euclidean instances. Moreover, the maximum degree and minimum degree of the vertices in the sparse graphs do not have much difference. Thus, the computation time of the methods to resolve the TSP on these sparse graphs will be greatly reduced.

Similarity Solutions of Nonlinear Stretched Biomagnetic Flow and Heat Transfer with Signum Function and Temperature Power Law Geometries

Biomagnetic fluid dynamics is an interdisciplinary field comprising engineering, medicine, and biology. Bio fluid dynamics is directed towards finding and developing the solutions to some of the human body related diseases and disorders. This article describes the flow and heat transfer of two dimensional, steady, laminar, viscous and incompressible biomagnetic fluid over a non-linear stretching sheet in the presence of magnetic dipole. Our model is consistent with blood fluid namely biomagnetic fluid dynamics (BFD). This model based on the principles of ferrohydrodynamic (FHD). The temperature at the stretching surface is assumed to follow a power law variation, and stretching velocity is assumed to have a nonlinear form with signum function or sign function. The governing boundary layer equations with boundary conditions are simplified to couple higher order equations using usual transformations. Numerical solutions for the governing momentum and energy equations are obtained by efficient numerical techniques based on the common finite difference method with central differencing, on a tridiagonal matrix manipulation and on an iterative procedure. Computations are performed for a wide range of the governing parameters such as magnetic field parameter, power law exponent temperature parameter, and other involved parameters and the effect of these parameters on the velocity and temperature field is presented. It is observed that for different values of the magnetic parameter, the velocity distribution decreases while temperature distribution increases. Besides, the finite difference solutions results for skin-friction coefficient and rate of heat transfer are discussed. This study will have an important bearing on a high targeting efficiency, a high magnetic field is required in the targeted body compartment.

A Proposal on the Educational Transactional Analysis as a Dialogical Vision of Culture: Conceptual Signposts and Practical Tools for Educators

The multicultural composition of today's societies poses new challenges to educational contexts. Schools are therefore called first to develop dialogic aptitudes and communicative skills adapted to the complex reality of post-modern societies. It is indispensable for educators and for young people to learn theoretical and practical tools during their scholastic path, in order to allow the knowledge of themselves and of the others with the aim of recognizing the value of the others regardless of their culture. Dialogic Skills help to understand and manage individual differences by allowing the solution of problems and preventing conflicts. The Educational Sector of Eric Berne’s Transactional Analysis offers a range of methods and techniques for this purpose. Educational Transactional Analysis is firmly anchored in the Personalist Philosophy and deserves to be promoted as a theoretical frame suitable to face the challenges of contemporary education. The goal of this paper is therefore to outline some conceptual and methodological signposts for the education to dialogue by drawing concepts and methodologies from educational transactional analysis.

Development of a Multi-Factorial Instrument for Accident Analysis Based on Systemic Methods

The present research is built on three major pillars, commencing by making some considerations on accident investigation methods and pointing out both defining aspects and differences between linear and non-linear analysis. The traditional linear focus on accident analysis describes accidents as a sequence of events, while the latest systemic models outline interdependencies between different factors and define the processes evolution related to a specific (normal) situation. Linear and non-linear accident analysis methods have specific limitations, so the second point of interest is mirrored by the aim to discover the drawbacks of systemic models which becomes a starting point for developing new directions to identify risks or data closer to the cause of incidents/accidents. Since communication represents a critical issue in the interaction of human factor and has been proved to be the answer of the problems made by possible breakdowns in different communication procedures, from this focus point, on the third pylon a new error-modeling instrument suitable for risk assessment/accident analysis will be elaborated.

Dialect and Gender Variations in the Place and Manner of Articulation of the Korean Fricatives

This study examines dialect and gender variations in the place and manner of articulation between the two Korean fricatives, /s/ and /s’/, as produced by speakers of the Daegu and Jeju dialects. The acoustic parameters of center of gravity and skewness for the place of articulation, and the rise time and the amplitude rise slope for the manner of articulation were measured. The study results revealed a gender effect, but no dialect effect, for the center of gravity and the skewness. No main effect for either the gender or dialect was found for the rise time and the amplitude rise slope. These findings indicated that, with regard to the place of articulation, Korean fricative sound differences are a gender distinction, not a dialectal one.

Generative Syntaxes: Macro-Heterophony and the Form of ‘Synchrony’

One of the most powerful language innovation in the twentieth century music was the heterophony–hypostasis of the vertical syntax entered into the sphere of interest of many composers, such as George Enescu, Pierre Boulez, Mauricio Kagel, György Ligeti and others. The heterophonic syntax has a history of its growth, which means a succession of different concepts and writing techniques. The trajectory of settling this phenomenon does not necessarily take into account the chronology: there are highly complex primary stages and advanced stages of returning to the simple forms of writing. In folklore, the plurimelodic simultaneities are free or random and originate from the (unintentional) differences/‘deviations’ from the state of unison, through a variety of ornaments, melismas, imitations, elongations and abbreviations, all in a flexible rhythmic and non-periodic/immeasurable framework, proper to the parlando-rubato rhythmics. Within the general framework of the multivocal organization, the heterophonic syntax in elaborate (academic) version has imposed itself relatively late compared with polyphony and homophony. Of course, the explanation is simple, if we consider the causal relationship between the sound vocabulary elements – in this case, the modalism – and the typologies of vertical organization appropriate for it. Therefore, adding up the ‘classic’ pathway of the writing typologies (monody – polyphony – homophony), heterophony - applied equally to the structures of modal, serial or synthesis vocabulary – reclaims necessarily an own macrotemporal form, in the sense of the analogies enshrined by the evolution of the musical styles and languages: polyphony→fugue, homophony→sonata. Concerned about the prospect of edifying a new musical ontology, the composer Ştefan Niculescu experienced – along with the mathematical organization of heterophony according to his own original methods – the possibility of extrapolation of this phenomenon in macrostructural plan, reaching this way to the unique form of ‘synchrony’. Founded on coincidentia oppositorum principle (involving the ‘one-multiple’ binom), the sound architecture imagined by Ştefan Niculescu consists in one (temporal) model / algorithm of articulation of two sound states: 1. monovocality state (principle of identity) and 2. multivocality state (principle of difference). In this context, the heterophony becomes an (auto)generative mechanism, with macrotemporal amplitude, strategy that will be grown by the composer, practically throughout his creation (see the works: Ison I, Ison II, Unisonos I, Unisonos II, Duplum, Triplum, Psalmus, Héterophonies pour Montreux (Homages to Enescu and Bartók etc.). For the present demonstration, we selected one of the most edifying works of Ştefan Niculescu – Simphony II, Opus dacicum – where the form of (heterophony-)synchrony acquires monumental-symphonic features, representing an emblematic case for the complexity level achieved by this type of vertical syntax in the twentieth century music.

Effect of Plant Biostimulants on Fruit Set, Yield, and Quality Attributes of “Farbaly” Apricot Cultivar

Apulia region (southern Italy) is excellent for heavy production of apricot (Prunus armeniaca L.). Fruit quality is a combination of physical, chemical and nutritional characteristics. The present experiment was laid in the commercial orchard in Cerignola (Foggia district, Apulia region, 41°15’49’’N; 15°53’59’’E; 126 a.s.l.) during the 2014-2015 season. The experiment consisted of the use of three biostimulant treatments (Hendophyt®, Ergostim® and Radicon®) compared with untreated control on ‘Farbaly’ apricot cultivar, in order to evaluate the vegeto-productive and fruit qualitative attributes. Foliar spray of biostimulants was applied at different times during the growth season (at red ball, fruit setting and fruit development stages). Experimental data showed some specific differences among the biostimulant treatments, which fruit set, growth and productivity were affected. Moderate influences were found regarding the qualitative attributes of fruits. The soluble solid content was positively affected by Hendophyt® treatment. Antioxidant capacity was significantly higher in Hendophyt® and Radicon® treatments respect to the untreated control.

Preliminary Evaluation of Decommissioning Wastes for the First Commercial Nuclear Power Reactor in South Korea

The commercial nuclear power reactor in South Korea, Kori Unit 1, which was a 587 MWe pressurized water reactor that started operation since 1978, was permanently shut down in June 2017 without an additional operating license extension. The Kori 1 Unit is scheduled to become the nuclear power unit to enter the decommissioning phase. In this study, the preliminary evaluation of the decommissioning wastes for the Kori Unit 1 was performed based on the following series of process: firstly, the plant inventory is investigated based on various documents (i.e., equipment/ component list, construction records, general arrangement drawings). Secondly, the radiological conditions of systems, structures and components (SSCs) are established to estimate the amount of radioactive waste by waste classification. Third, the waste management strategies for Kori Unit 1 including waste packaging are established. Forth, selection of the proper decontamination and dismantling (D&D) technologies is made considering the various factors. Finally, the amount of decommissioning waste by classification for Kori 1 is estimated using the DeCAT program, which was developed by KEPCO-E&C for a decommissioning cost estimation. The preliminary evaluation results have shown that the expected amounts of decommissioning wastes were less than about 2% and 8% of the total wastes generated (i.e., sum of clean wastes and radwastes) before/after waste processing, respectively, and it was found that the majority of contaminated material was carbon or alloy steel and stainless steel. In addition, within the range of availability of information, the results of the evaluation were compared with the results from the various decommissioning experiences data or international/national decommissioning study. The comparison results have shown that the radioactive waste amount from Kori Unit 1 decommissioning were much less than those from the plants decommissioned in U.S. and were comparable to those from the plants in Europe. This result comes from the difference of disposal cost and clearance criteria (i.e., free release level) between U.S. and non-U.S. The preliminary evaluation performed using the methodology established in this study will be useful as a important information in establishing the decommissioning planning for the decommissioning schedule and waste management strategy establishment including the transportation, packaging, handling, and disposal of radioactive wastes.

Comparison of the H-Index of Researchers of Google Scholar and Scopus

H-index has been widely used as a performance indicator of researchers around the world especially in Indonesia. The Government uses Scopus and Google scholar as indexing references in providing recognition and appreciation. However, those two indexing services yield to different H-index values. For that purpose, this paper evaluates the difference of the H-index from those services. Researchers indexed by Webometrics, are used as reference’s data in this paper. Currently, Webometrics only uses H-index from Google Scholar. This paper observed and compared corresponding researchers’ data from Scopus to get their H-index score. Subsequently, some researchers with huge differences in score are observed in more detail on their paper’s publisher. This paper shows that the H-index of researchers in Google Scholar is approximately 2.45 times of their Scopus H-Index. Most difference exists due to the existence of uncertified publishers, which is considered in Google Scholar but not in Scopus.

Hardware Implementation of Local Binary Pattern Based Two-Bit Transform Motion Estimation

Nowadays, demand for using real-time video transmission capable devices is ever-increasing. So, high resolution videos have made efficient video compression techniques an essential component for capturing and transmitting video data. Motion estimation has a critical role in encoding raw video. Hence, various motion estimation methods are introduced to efficiently compress the video. Low bit‑depth representation based motion estimation methods facilitate computation of matching criteria and thus, provide small hardware footprint. In this paper, a hardware implementation of a two-bit transformation based low-complexity motion estimation method using local binary pattern approach is proposed. Image frames are represented in two-bit depth instead of full-depth by making use of the local binary pattern as a binarization approach and the binarization part of the hardware architecture is explained in detail. Experimental results demonstrate the difference between the proposed hardware architecture and the architectures of well-known low-complexity motion estimation methods in terms of important aspects such as resource utilization, energy and power consumption.

3D Numerical Analysis of Stone Columns Reinforced with Horizontal and Vertical Geosynthetic Materials

Improvement and reinforcement of soils with poor strength and engineering properties for constructing low height structures or structures such as liquid storage tanks, bridge columns, and heavy structures have necessitated applying particular techniques. Stone columns are among the well-known methods applied in such soils. This method provides an economically justified way for improving engineering properties of soft clay and loose sandy soils. Stone column implementation in these soils increases their bearing capacity and reduces the settlement of foundation build on them. In the present study, the finite difference based FLAC3D software was used to investigate the performance and effect of soil reinforcement through stone columns without lining and those with geosynthetic lining with different levels of stiffness in horizontal and vertical modes in clayey soils. The results showed that soil improvement using stone columns with lining in vertical and horizontal modes results in improvement of bearing capacity and foundation settlement.

Nutritional Value of Rabbit Meat after Contamination with 1,1-Dimethylhydrazine

In this article reduced nutritional value of the rabbits’ meat at 1, 1 dimethylhydrazine experimental toxicosis is shown. The assay was performed on liquid chromatograph SHIMADZU LC-20 Prominence (Japan) with fluorometric and spectrophotometric detector. This research has revealed that samples of rabbit meat of the experimental group had significant differences from the control group:in amino acids concentration from 1.2% to 9.1%; vitamin concentration from 11.2% to 60.5%, macro – minerals concentration from 17.4% to 78.1% and saturated fatty acids concentration from 17,1% to 34.5%, respectively. The decrease in the chemical composition of rabbits’ meat at 1,1 dimethylhydrazine toxicosis may be due to changes in the internal processes associated with impaired metabolic homeostasis of animals.