Abstract: Economics is not an exact science. It cannot be from the moment it is a social science that concerns society organization, a human science that depends on the behavior of the men and women who make a part of this society. Therefore, it cannot ignore morality, the instinctive sense of good and evil, the natural order which place us between certain values, and which religion often sheds light on. In terms of finance, the reference to ethics is becoming more popular than ever. This is naturally due to the growing financial crises. Finance is less and less ethical, but some financial practices have continued to do so. This is the case of ethical finance and Islamic finance. After attempting to define the concepts of ethical finance and Islamic finance, in a period when financial innovation seeks to encourage differentiation in order to create more profit margins, this article attempts to expose the particularities, the convergences and the potentialities of development of these two sensibilities.
Abstract: In this study, 177Lu-DOTATOC was prepared under optimized conditions (radiochemical purity: > 99%, radionuclidic purity: > 99%). The percentage of injected dose per gram (%ID/g) was calculated for organs up to 168 h post injection. Compartmental model was applied to mathematical description of the drug behaviour in tissue at different times. The biodistribution data showed the significant excretion of the radioactivity from the kidneys. The adrenal and pancreas, as major expression sites for somatostatin receptor (SSTR), had significant uptake. A pharmacokinetic model of 177Lu-DOTATOC was presented by compartmental analysis which demonstrates the behavior of the complex.
Abstract: Any industrial company needs to determine the amount of variation that exists within its measurement process and guarantee the reliability of their data, studying the performance of their measurement system, in terms of linearity, bias, repeatability and reproducibility and stability. This issue is critical for automotive industry suppliers, who are required to be certified by the 16949:2016 standard (replaces the ISO/TS 16949) of International Automotive Task Force, defining the requirements of a quality management system for companies in the automotive industry. Measurement System Analysis (MSA) is one of the mandatory tools. Frequently, the measurement system in companies is not connected to the equipment and do not incorporate the methods proposed by the Automotive Industry Action Group (AIAG). To address these constraints, an R&D project is in progress, whose objective is to develop a web and cloud-based MSA tool. This MSA tool incorporates Industry 4.0 concepts, such as, Internet of Things (IoT) protocols to assure the connection with the measuring equipment, cloud computing, artificial intelligence, statistical tools, and advanced mathematical algorithms. This paper presents the preliminary findings of the project. The web and cloud-based MSA tool is innovative because it implements all statistical tests proposed in the MSA-4 reference manual from AIAG as well as other emerging methods and techniques. As it is integrated with the measuring devices, it reduces the manual input of data and therefore the errors. The tool ensures traceability of all performed tests and can be used in quality laboratories and in the production lines. Besides, it monitors MSAs over time, allowing both the analysis of deviations from the variation of the measurements performed and the management of measurement equipment and calibrations. To develop the MSA tool a ten-step approach was implemented. Firstly, it was performed a benchmarking analysis of the current competitors and commercial solutions linked to MSA, concerning Industry 4.0 paradigm. Next, an analysis of the size of the target market for the MSA tool was done. Afterwards, data flow and traceability requirements were analysed in order to implement an IoT data network that interconnects with the equipment, preferably via wireless. The MSA web solution was designed under UI/UX principles and an API in python language was developed to perform the algorithms and the statistical analysis. Continuous validation of the tool by companies is being performed to assure real time management of the ‘big data’. The main results of this R&D project are: MSA Tool, web and cloud-based; Python API; New Algorithms to the market; and Style Guide of UI/UX of the tool. The MSA tool proposed adds value to the state of the art as it ensures an effective response to the new challenges of measurement systems, which are increasingly critical in production processes. Although the automotive industry has triggered the development of this innovative MSA tool, other industries would also benefit from it. Currently, companies from molds and plastics, chemical and food industry are already validating it.
Abstract: Embodied Cognition (EC) as a learning paradigm is based on the idea of an inseparable link between body, mind, and environment. In recent years, the advent of theoretical learning approaches around EC theory has resulted in a number of empirical studies exploring the implementation of the theory in education. This systematic literature overview identifies the mainstream of EC research and emphasizes on the implementation of the theory across learning environments. Based on a corpus of 43 manuscripts, published between 2013 and 2017, it sets out to describe the range of topics covered under the umbrella of EC and provides a holistic view of the field. The aim of the present review is to investigate the main issues in EC research related to the various learning contexts. Particularly, the study addresses the research methods and technologies that are utilized, and it also explores the integration of body into the learning context. An important finding from the overview is the potential of the theory in different educational environments and disciplines. However, there is a lack of an explicit pedagogical framework from an educational perspective for a successful implementation in various learning contexts.
Abstract: Cyclists often run through a crosswind and sometimes we experience the adverse pressure. We came to an idea that Wells turbine can be used as power augmentation device in the crosswind something like sails of a yacht. Wells turbine always rotates in the same direction irrespective of the incoming flow direction, and we use it in the small-scale power generation in the ocean where waves create an oscillating flow. We incorporate the turbine to the wheel of a bike. A commercial device integrates strain gauges in the crank of a bike and transmitted force and torque applied to the pedal of the bike as an e-mail to the driver’s mobile phone. We can analyze the unsteady data in a spreadsheet sent from the crank sensor. We run the bike with the crank sensor on the rollers at the exit of a low-speed wind tunnel and analyze the effect of the crosswind to the wheel with a Wells turbine. We also test the aerodynamic characteristics of the turbine separately. Although power gain depends on the flow direction, several Watts increase might be possible by the Wells turbine incorporated to a bike wheel.
Abstract: The cardiopulmonary signal monitoring, without the
usage of contact electrodes or any type of in-body sensors, has
several applications such as sleeping monitoring and continuous
monitoring of vital signals in bedridden patients. This system has
also applications in the vehicular environment to monitor the driver,
in order to avoid any possible accident in case of cardiac failure.
Thus, the bio-radar system proposed in this paper, can measure vital
signals accurately by using the Doppler effect principle that relates
the received signal properties with the distance change between the
radar antennas and the person’s chest-wall. Once the bio-radar aim
is to monitor subjects in real-time and during long periods of time,
it is impossible to guarantee the patient immobilization, hence their
random motion will interfere in the acquired signals. In this paper,
a mathematical model of the bio-radar is presented, as well as its
simulation in MATLAB. The used algorithm for breath rate extraction
is explained and a method for DC offsets removal based in a motion
detection system is proposed. Furthermore, experimental tests were
conducted with a view to prove that the unavoidable random motion
can be used to estimate the DC offsets accurately and thus remove
them successfully.
Abstract: Nowadays, the progressive decrease of primary natural resources and ongoing upward trend in terms of energy demand, have resulted in development of new generation technological processes which are focused on step-wise production and residues utilization. Thus, microalgae-based 3rd generation bioeconomy is considered one of the most promising approaches that allow production of value-added products and sophisticated utilization of residues biomass. In comparison to conventional biomass, microalgae can be cultivated in wide range of conditions without compromising food and feed production, and thus, addressing issues associated with negative social and environmental impacts. However, one of the most challenging tasks is to undergo seasonal variations and to achieve optimal growing conditions for indoor closed systems that can cover further demand for material and energetic utilization of microalgae. For instance, outdoor cultivation in St. Petersburg (Russia) is only suitable within rather narrow time frame (from mid-May to mid-September). At earlier and later periods, insufficient sunlight and heat for the growth of microalgae were detected. On the other hand, without additional physical effects, the biomass increment in summer is 3-5 times per week, depending on the solar radiation and the ambient temperature. In order to increase biomass production, scientists from all over the world have proposed various technical solutions for cultivators and have been studying the influence of various physical factors affecting biomass growth namely: magnetic field, radiation impact, and electric field, etc. In this paper, the influence of infrared radiation (IR) and fluorescent light on the growth rate of microalgae Chlorella sorokiniana has been studied. The cultivation of Chlorella sorokiniana was carried out in 500 ml cylindrical glass vessels, which were constantly aerated. To accelerate the cultivation process, the mixture was stirred for 15 minutes at 500 rpm following 120 minutes of rest time. At the same time, the metabolic needs in nutrients were provided by the addition of micro- and macro-nutrients in the microalgae growing medium. Lighting was provided by fluorescent lamps with the intensity of 2500 ± 300 lx. The influence of IR was determined using IR lamps with a voltage of 220 V, power of 250 W, in order to achieve the intensity of 13 600 ± 500 lx. The obtained results show that under the influence of fluorescent lamps along with the combined effect of active aeration and variable mixing, the biomass increment on the 2nd day was three times, and on the 7th day, it was eight-fold. The growth rate of microalgae under the influence of IR radiation was lower and has reached 22.6·106 cells·mL-1. However, application of IR lamps for the biomass growth allows maintaining the optimal temperature of microalgae suspension at approximately 25-28°C, which might especially be beneficial during the cold season in extreme climate zones.
Abstract: The use of waste rubber chips not only can be of great importance in terms of the environment, but also can be used to increase the shear strength of soils. The purpose of this study was to evaluate the variation of the internal friction angle of liquefiable sandy soil using waste rubber chips. For this purpose, the geotechnical properties of unmodified and modified soil samples by waste lining rubber chips have been evaluated and analyzed by performing the triaxial consolidated drained test. In order to prepare the laboratory specimens, the sandy soil in part of Rudsar shores in Gilan province, north of Iran with high liquefaction potential has been replaced by two percent of waste rubber chips. Samples have been compressed until reaching the two levels of density of 15.5 and 16.7 kN/m3. Also, in order to find the optimal length of chips in sandy soil, the rectangular rubber chips with the widths of 0.5 and 1 cm and the lengths of 0.5, 1, and 2 cm were used. The results showed that the addition of rubber chips to liquefiable sandy soil greatly increases the shear resistance of these soils. Also, it can be seen that decreasing the width and increasing the length-to-width ratio of rubber chips has a direct impact on the shear strength of the modified soil samples with rubber chips.
Abstract: Over the last decade due to climate change and a strategy of natural resources preservation, the interest for the aquatic biomass has dramatically increased. Along with mitigation of the environmental pressure and connection of waste streams (including CO2 and heat emissions), microalgae bioeconomy can supply food, feed, as well as the pharmaceutical and power industry with number of value-added products. Furthermore, in comparison to conventional biomass, microalgae can be cultivated in wide range of conditions without compromising food and feed production, thus addressing issues associated with negative social and the environmental impacts. This paper presents the state-of-the art technology for microalgae bioeconomy from cultivation process to production of valuable components and by-streams. Microalgae Chlorella sorokiniana were cultivated in the pilot-scale innovation concept in Hamburg (Germany) using different systems such as race way pond (5000 L) and flat panel reactors (8 x 180 L). In order to achieve the optimum growth conditions along with suitable cellular composition for the further extraction of the value-added components, process parameters such as light intensity, temperature and pH are continuously being monitored. On the other hand, metabolic needs in nutrients were provided by addition of micro- and macro-nutrients into a medium to ensure autotrophic growth conditions of microalgae. The cultivation was further followed by downstream process and extraction of lipids, proteins and saccharides. Lipids extraction is conducted in repeated-batch semi-automatic mode using hot extraction method according to Randall. As solvents hexane and ethanol are used at different ratio of 9:1 and 1:9, respectively. Depending on cell disruption method along with solvents ratio, the total lipids content showed significant variations between 8.1% and 13.9 %. The highest percentage of extracted biomass was reached with a sample pretreated with microwave digestion using 90% of hexane and 10% of ethanol as solvents. Proteins content in microalgae was determined by two different methods, namely: Total Kejadahl Nitrogen (TKN), which further was converted to protein content, as well as Bradford method using Brilliant Blue G-250 dye. Obtained results, showed a good correlation between both methods with protein content being in the range of 39.8–47.1%. Characterization of neutral and acid saccharides from microalgae was conducted by phenol-sulfuric acid method at two wavelengths of 480 nm and 490 nm. The average concentration of neutral and acid saccharides under the optimal cultivation conditions was 19.5% and 26.1%, respectively. Subsequently, biomass residues are used as substrate for anaerobic digestion on the laboratory-scale. The methane concentration, which was measured on the daily bases, showed some variations for different samples after extraction steps but was in the range between 48% and 55%. CO2 which is formed during the fermentation process and after the combustion in the Combined Heat and Power unit can potentially be used within the cultivation process as a carbon source for the photoautotrophic synthesis of biomass.
Abstract: The current study is a comparative analysis of the use of shell nouns as a cohesive device (CD) in an English for Second Language (ESL) setting in order to identify their use and relationship in the quality of second language (L2) writing. As these nouns were established to anticipate the meaning within, across or outside the text, their use has fascinated writing researchers. The corpus of the study included published articles from reputable journals and graduate students’ papers in order to analyze the frequency of shell nouns using “highly prevalent” nouns in the academic community, to identify the different lexicogrammatical patterns where these nouns occur and to the functions connected with these patterns. The result of the study implies that published authors used more shell nouns in their paper than graduate students. However, the functions of the different lexicogrammatical patterns for the frequently occurring shell nouns are somewhat similar. These results could help students in enhancing the cohesion of their text and in comprehending it.
Abstract: Chitosan is a natural polysaccharide prepared by the N-deacetylation of chitin. In this study, the physicochemical and antibacterial properties of chitosan nanoparticles, produced by ultrasound irradiation, were evaluated. The physicochemical properties of the nanoparticles were determined by dynamic light scattering and zeta potential analysis. Chitosan nanoparticles inhibited the growth of E. coli. The minimum inhibitory concentration (MIC) values were lower than 0.5 mg/mL, and the minimum bactericidal concentration (MBC) values were similar or higher than MIC values. Confocal laser scanning micrographs (CLSM) were used to observe the interaction between E. coli suspensions mixed with FITC-labeled chitosan polymers and nanoparticles.
Abstract: Source apportionment using Dispersion model depends primarily on the quality of Emission Inventory. In the present study, a CMB receptor model has been used to identify the sources of PM2.5, while the AERMOD dispersion model has been used to account for missing sources of PM2.5 in the Emission Inventory. A statistical approach has been developed to quantify the missing sources not considered in the Emission Inventory. The inventory of each grid was improved by adjusting emissions based on road lengths and deficit in measured and modelled concentrations. The results showed that in CMB analyses, fugitive sources - soil and road dust - contribute significantly to ambient PM2.5 pollution. As a result, AERMOD significantly underestimated the ambient air concentration at most locations. The revised Emission Inventory showed a significant improvement in AERMOD performance which is evident through statistical tests.
Abstract: Acidification is a technique used in oil reservoirs
to improve annual production, reduce the skin and increase the
pressure of an oil well while eliminating the formation damage that
occurs during the drilling process, completion and, amongst others,
to create new channels allowing the easy circulation of oil around
a producing well. This is achieved by injecting an acidizing fluid
at a relatively low pressure to prevent fracturing formation. The
treatment fluid used depends on the type and nature of the reservoir
rock traversed as well as its petrophysical properties. In order to
understand the interaction mechanisms between the treatment fluids
used for the reservoir rock acidizing, several candidate wells for
stimulation were selected in the large Hassi Messaoud deposit in
southern Algeria. The stimulation of these wells is completed using
different fluids composed mainly of HCl acid with other additives
such as corrosion inhibitors, clay stabilizers and iron controllers.
These treatment fluids are injected over two phases, namely with
clean tube (7.5% HCl) and matrix aidizing with HCl (15%). The
stimulation results obtained are variable according to the type of
rock traversed and its mineralogical composition. These results show
that there has been an increase in production flow and head pressure
respectively from 1.99 m3 / h to 3.56 m3 / h and from 13 Kgf / cm2
to 20 kgf / cm2 in the sands formation having good petrophysical
properties of (porosity = 16%) and low amount of clay (Vsh = 6%).
Abstract: The intelligent management and optimisation of radio resource technologies will lead to a considerable improvement in the overall performance in Next Generation Networks (NGNs). Carrier Aggregation (CA) technology, also known as Spectrum Aggregation, enables more efficient use of the available spectrum by combining multiple Component Carriers (CCs) in a virtual wideband channel. LTE-A (Long Term Evolution–Advanced) CA technology can combine multiple adjacent or separate CCs in the same band or in different bands. In this way, increased data rates and dynamic load balancing can be achieved, resulting in a more reliable and efficient operation of mobile networks and the enabling of high bandwidth mobile services. In this paper, several distinct CA deployment strategies for the utilisation of spectrum bands are compared in indoor-outdoor scenarios, simulated via the recently-developed Realistic Indoor Environment Generator (RIEG). We analyse the performance of the User Equipment (UE) by integrating the average throughput, the level of fairness of radio resource allocation, and other parameters, into one summative assessment termed a Comparative Factor (CF). In addition, comparison of non-CA and CA indoor mobile networks is carried out under different load conditions: varying numbers and positions of UEs. The experimental results demonstrate that the CA technology can improve network performance, especially in the case of indoor scenarios. Additionally, we show that an increase of carrier frequency does not necessarily lead to improved CF values, due to high wall-penetration losses. The performance of users under bad-channel conditions, often located in the periphery of the cells, can be improved by intelligent CA location. Furthermore, a combination of such a deployment and effective radio resource allocation management with respect to user-fairness plays a crucial role in improving the performance of LTE-A networks.
Abstract: Missing values in real-world datasets are a common
problem. Many algorithms were developed to deal with this
problem, most of them replace the missing values with a fixed
value that was computed based on the observed values. In
our work, we used a distance function based on Bhattacharyya
distance to measure the distance between objects with missing
values. Bhattacharyya distance, which measures the similarity of
two probability distributions. The proposed distance distinguishes
between known and unknown values. Where the distance between
two known values is the Mahalanobis distance. When, on the other
hand, one of them is missing the distance is computed based on the
distribution of the known values, for the coordinate that contains
the missing value. This method was integrated with Wikaya, a
digital health company developing a platform that helps to improve
prevention of chronic diseases such as diabetes and cancer. In order
for Wikaya’s recommendation system to work distance between users
need to be measured. Since there are missing values in the collected
data, there is a need to develop a distance function distances between
incomplete users profiles. To evaluate the accuracy of the proposed
distance function in reflecting the actual similarity between different
objects, when some of them contain missing values, we integrated it
within the framework of k nearest neighbors (kNN) classifier, since
its computation is based only on the similarity between objects. To
validate this, we ran the algorithm over diabetes and breast cancer
datasets, standard benchmark datasets from the UCI repository. Our
experiments show that kNN classifier using our proposed distance
function outperforms the kNN using other existing methods.
Abstract: Cardiologists perform cardiac auscultation to detect
abnormalities in heart sounds. Since accurate auscultation is
a crucial first step in screening patients with heart diseases,
there is a need to develop computer-aided detection/diagnosis
(CAD) systems to assist cardiologists in interpreting heart sounds
and provide second opinions. In this paper different algorithms
are implemented for automated heart sound classification using
unsegmented phonocardiogram (PCG) signals. Support vector
machine (SVM), artificial neural network (ANN) and cartesian
genetic programming evolved artificial neural network (CGPANN)
without the application of any segmentation algorithm has been
explored in this study. The signals are first pre-processed to remove
any unwanted frequencies. Both time and frequency domain features
are then extracted for training the different models. The different
algorithms are tested in multiple scenarios and their strengths and
weaknesses are discussed. Results indicate that SVM outperforms
the rest with an accuracy of 73.64%.
Abstract: The influence of a pulsatile electroosmotic flow (PEOF)
at the rate of spread, or dispersivity, for a non-reactive solute released
in a microcapillary with slippage at the boundary wall (modeled by
the Navier-slip condition) is theoretically analyzed. Based on the flow
velocity field developed under such conditions, the present study
implements an analytical scheme of scaling known as the Theory
of Homogenization, in order to obtain a mathematical expression for
the dispersivity, valid at a large time scale where the initial transients
have vanished and the solute spreads under the Taylor dispersion
influence. Our results show the dispersivity is a function of a slip
coefficient, the amplitude of the imposed electric field, the Debye
length and the angular Reynolds number, highlighting the importance
of the latter as an enhancement/detrimental factor on the dispersivity,
which allows to promote the PEOF as a strong candidate for chemical
species separation at lab-on-a-chip devices.
Abstract: The Oscillatory electroosmotic flow (OEOF) in power
law fluids through a microchannel is studied numerically. A
time-dependent external electric field (AC) is suddenly imposed
at the ends of the microchannel which induces the fluid motion.
The continuity and momentum equations in the x and y direction
for the flow field were simplified in the limit of the lubrication
approximation theory (LAT), and then solved using a numerical
scheme. The solution of the electric potential is based on the
Debye-H¨uckel approximation which suggest that the surface potential
is small,say, smaller than 0.025V and for a symmetric (z : z)
electrolyte. Our results suggest that the velocity profiles across
the channel-width are controlled by the following dimensionless
parameters: the angular Reynolds number, Reω, the electrokinetic
parameter, ¯κ, defined as the ratio of the characteristic length scale
to the Debye length, the parameter λ which represents the ratio
of the Helmholtz-Smoluchowski velocity to the characteristic length
scale and the flow behavior index, n. Also, the results reveal that
the velocity profiles become more and more non-uniform across the
channel-width as the Reω and ¯κ are increased, so oscillatory OEOF
can be really useful in micro-fluidic devices such as micro-mixers.
Abstract: Due to the economic and environmental benefits of producing less waste, the US Environmental Protection Agency (EPA) introduces source reduction as one of the most important means to deal with the problems caused by increased landfills and pollution. Waste reduction involves all waste management methods, including source reduction, recycling, and composting, which reduce waste flow to landfills or other disposal facilities. Source reduction of waste can be studied from two perspectives: avoiding waste production, or reducing per capita waste production, and waste deviation that indicates the reduction of waste transfer to landfills. The present paper has investigated home composting as a managerial solution for reduction of waste transfer to landfills. Home composting has many benefits. The use of household waste for the production of compost will result in a much smaller amount of waste being sent to landfills, which in turn will reduce the costs of waste collection, transportation and burial. Reducing the volume of waste for disposal and using them for the production of compost and plant fertilizer might help to recycle the material in a shorter time and to use them effectively in order to preserve the environment and reduce contamination. Producing compost in a home-based manner requires very small piece of land for preparation and recycling compared with other methods. The final product of home-made compost is valuable and helps to grow crops and garden plants. It is also used for modifying the soil structure and maintaining its moisture. The food that is transferred to landfills will spoil and produce leachate after a while. It will also release methane and greenhouse gases. But, composting these materials at home is the best way to manage degradable materials, use them efficiently and reduce environmental pollution. Studies have shown that the benefits of the sale of produced compost and the reduced costs of collecting, transporting, and burying waste can well be responsive to the costs of purchasing home compost machine and the cost of related trainings. Moreover, the process of producing home compost may be profitable within 4 to 5 years and as a result, it will have a major role in reducing waste management.
Abstract: From August to December of 2016, a diagnostic and strategic planning study was carried out on the supply chain of the company Agropecuaria GABO S.A. de C.V. The final product of the study was the development of the strategic plan and a project portfolio to meet the demands of the three links in the supply chain of the Roma tomato exported annually to the United States of America. In this project, the strategic objective of ensuring the proper handling of the product was selected and one of the goals associated with this was the employment of quantitative methods to support decision making. Considering the antecedents, the objective of this case study was to develop a model to analyze the behavioral dynamics in the distribution chain, from the logistics of storage and shipment of Roma tomato in 81-case pallets (11.5 kg per case), to the two pre-cooling rooms and eventual loading onto transports, seeking to reduce the bottleneck and the associated costs by means of a dynamic interface. The methodology used was that of system dynamics, considering four phases that were adapted to the purpose of the study: 1) the conceptualization phase; 2) the formulation phase; 3) the evaluation phase; and 4) the communication phase. The main practical conclusions lead to the possibility of reducing both the bottlenecks in the cooling rooms and the costs by simulating scenarios and modifying certain policies. Furthermore, the creation of the dynamic interface between the model and the stakeholders was achieved by generating interaction with buttons and simple instructions that allow making modifications and observing diverse behaviors.