Laser Beam Micro-Drilling Effect on Ti-6Al-4V Titanium Alloy Sheet Properties

Laser beam micro-drilling (LBMD) is one of the most important non-contact machining processes of materials that are difficult to machine by means oeqf conventional machining methods used in various industries. The paper is focused on LBMD knock-down effect on Ti-6Al-4V (Grade 5) titanium alloy sheets properties. Two various process configurations were verified with a focus on laser damages in back-structure parts affected by the process. The effects of the LBMD on the material properties were assessed by means of tensile and fatigue tests and fracture surface analyses. Fatigue limit of LBMD configurations reached a significantly lower value between 15% and 30% of the static strength as compared to the reference raw material with 58% value. The farther back-structure configuration gives a two-fold fatigue life as compared to the closer LBMD configuration at a given stress applied.

Influence of a Pulsatile Electroosmotic Flow on the Dispersivity of a Non-Reactive Solute through a Microcapillary

The influence of a pulsatile electroosmotic flow (PEOF) at the rate of spread, or dispersivity, for a non-reactive solute released in a microcapillary with slippage at the boundary wall (modeled by the Navier-slip condition) is theoretically analyzed. Based on the flow velocity field developed under such conditions, the present study implements an analytical scheme of scaling known as the Theory of Homogenization, in order to obtain a mathematical expression for the dispersivity, valid at a large time scale where the initial transients have vanished and the solute spreads under the Taylor dispersion influence. Our results show the dispersivity is a function of a slip coefficient, the amplitude of the imposed electric field, the Debye length and the angular Reynolds number, highlighting the importance of the latter as an enhancement/detrimental factor on the dispersivity, which allows to promote the PEOF as a strong candidate for chemical species separation at lab-on-a-chip devices.

Free Vibration of Axially Functionally Graded Simply Supported Beams Using Differential Transformation Method

Free vibration analysis of homogenous and axially functionally graded simply supported beams within the context of Euler-Bernoulli beam theory is presented in this paper. The material properties of the beams are assumed to obey the linear law distribution. The effective elastic modulus of the composite was predicted by using the rule of mixture. Here, the complexities which appear in solving differential equation of transverse vibration of composite beams which limit the analytical solution to some special cases are overcome using a relatively new approach called the Differential Transformation Method. This technique is applied for solving differential equation of transverse vibration of axially functionally graded beams. Natural frequencies and corresponding normalized mode shapes are calculated for different Young’s modulus ratios. MATLAB code is designed to solve the transformed differential equation of the beam. Comparison of the present results with the exact solutions proves the effectiveness, the accuracy, the simplicity, and computational stability of the differential transformation method. The effect of the Young’s modulus ratio on the normalized natural frequencies and mode shapes is found to be very important.

Groundwater Potential Zone Identification in Unconsolidated Aquifer Using Geophysical Techniques around Tarbela Ghazi, District Haripur, Pakistan

Electrical resistivity investigation was conducted in vicinity of Tarbela Ghazi, in order to study the subsurface layer with a view of determining the depth to the aquifer and thickness of groundwater potential zones. Vertical Electrical Sounding (VES) using Schlumberger array was carried out at 16 VES stations. Well logging data at four tube wells have been used to mark the super saturated zones with great discharge rate. The present paper shows a geoelectrical identification of the lithology and an estimate of the relationship between the resistivity and Dar Zarrouk parameters (transverse unit resistance and longitudinal unit conductance). The VES results revealed both homogeneous and heterogeneous nature of the subsurface strata. Aquifer is unconfined to confine in nature, and at few locations though perched aquifer has been identified, groundwater potential zones are developed in unconsolidated deposits layers and more than seven geo-electric layers are observed at some VES locations. Saturated zones thickness ranges from 5 m to 150 m, whereas at few area aquifer is beyond 150 m thick. The average anisotropy, transvers resistance and longitudinal conductance values are 0.86 %, 35750.9821 Ω.m2, 0.729 Siemens, respectively. The transverse unit resistance values fluctuate all over the aquifer system, whereas below at particular depth high values are observed, that significantly associated with the high transmissivity zones. The groundwater quality in all analyzed samples is below permissible limit according to World Health Standard (WHO).

Kinetics and Thermodynamics Adsorption of Phenolic Compounds on Organic-Inorganic Hybrid Mesoporous Material

Mesoporous materials are very commonly used as adsorbent materials for removing phenolic compounds. However, the adsorption mechanism of these compounds is still poorly controlled. However, understanding the interactions mesoporous materials/adsorbed molecules is very important in order to optimize the processes of liquid phase adsorption. The difficulty of synthesis is to keep an orderly and cubic pore structure and achieve a homogeneous surface modification. The grafting of Si(CH3)3 was chosen, to transform hydrophilic surfaces hydrophobic surfaces. The aim of this work is to study the kinetics and thermodynamics of two volatile organic compounds VOC phenol (PhOH) and P hydroxy benzoic acid (4AHB) on a mesoporous material of type MCM-48 grafted with an organosilane of the Trimethylchlorosilane (TMCS) type, the material thus grafted or functionalized (hereinafter referred to as MCM-48-G). In a first step, the kinetic and thermodynamic study of the adsorption isotherms of each of the VOCs in mono-solution was carried out. In a second step, a similar study was carried out on a mixture of these two compounds. Kinetic models (pseudo-first order, pseudo-second order) were used to determine kinetic adsorption parameters. The thermodynamic parameters of the adsorption isotherms were determined by the adsorption models (Langmuir, Freundlich). The comparative study of adsorption of PhOH and 4AHB proved that MCM-48-G had a high adsorption capacity for PhOH and 4AHB; this may be related to the hydrophobicity created by the organic function of TMCS in MCM-48-G. The adsorption results for the two compounds using the Freundlich and Langmuir models show that the adsorption of 4AHB was higher than PhOH. The values ​​obtained by the adsorption thermodynamics show that the adsorption interactions for our sample with the phenol and 4AHB are of a physical nature. The adsorption of our VOCs on the MCM-48 (G) is a spontaneous and exothermic process.

Dosimetric Analysis of Intensity Modulated Radiotherapy versus 3D Conformal Radiotherapy in Adult Primary Brain Tumors: Regional Cancer Centre, India

Radiation therapy has undergone many advancements and evloved from 2D to 3D. Recently, with rapid pace of drug discoveries, cutting edge technology, and clinical trials has made innovative advancements in computer technology and treatment planning and upgraded to intensity modulated radiotherapy (IMRT) which delivers in homogenous dose to tumor and normal tissues. The present study was a hospital-based experience comparing two different conformal radiotherapy techniques for brain tumors. This analytical study design has been conducted at Regional Cancer Centre, India from January 2014 to January 2015. Ten patients have been selected after inclusion and exclusion criteria. All the patients were treated on Artiste Siemens Linac Accelerator. The tolerance level for maximum dose was 6.0 Gyfor lenses and 54.0 Gy for brain stem, optic chiasm and optical nerves as per RTOG criteria. Mean and standard deviation values of PTV98%, PTV 95% and PTV 2% in IMRT were 93.16±2.9, 95.01±3.4 and 103.1±1.1 respectively; for 3DCRT were 91.4±4.7, 94.17±2.6 and 102.7±0.39 respectively. PTV max dose (%) in IMRT and 3D-CRT were 104.7±0.96 and 103.9±1.0 respectively. Maximum dose to the tumor can be delivered with IMRT with acceptable toxicity limits. Variables such as expertise, location of tumor, patient condition, and TPS influence the outcome of the treatment.

Comparison of Automated Zone Design Census Output Areas with Existing Output Areas in South Africa

South Africa is one of the few countries that have stopped using the same Enumeration Areas (EAs) for census enumeration and dissemination. The advantage of this change is that confidentiality issue could be addressed for census dissemination as the design of geographic unit for collection is mainly to ensure that this unit is covered by one enumerator. The objective of this paper was to evaluate the performance of automated zone design output areas against non-zone design developed geographies using the 2001 census data, and 2011 census to some extent, as the main input. The comparison of the Automated Zone-design Tool (AZTool) census output areas with the Small Area Layers (SALs) and SubPlaces based on confidentiality limit, population distribution, and degree of homogeneity, as well as shape compactness, was undertaken. Further, SPSS was employed for validation of the AZTool output results. The results showed that AZTool developed output areas out-perform the existing official SAL and SubPlaces with regard to minimum population threshold, population distribution and to some extent to homogeneity. Therefore, it was concluded that AZTool program provides a new alternative to the creation of optimised census output areas for dissemination of population census data in South Africa.

A Condition-Based Maintenance Policy for Multi-Unit Systems Subject to Deterioration

In this paper, we propose a condition-based maintenance policy for multi-unit systems considering the existence of economic dependency among units. We consider a system composed of N identical units, where each unit deteriorates independently. Deterioration process of each unit is modeled as a three-state continuous time homogeneous Markov chain with two working states and a failure state. The average production rate of units varies in different working states and demand rate of the system is constant. Units are inspected at equidistant time epochs, and decision regarding performing maintenance is determined by the number of units in the failure state. If the total number of units in the failure state exceeds a critical level, maintenance is initiated, where units in failed state are replaced correctively and deteriorated state units are maintained preventively. Our objective is to determine the optimal number of failed units to initiate maintenance minimizing the long run expected average cost per unit time. The problem is formulated and solved in the semi-Markov decision process (SMDP) framework. A numerical example is developed to demonstrate the proposed policy and the comparison with the corrective maintenance policy is presented.

A World Map of Seabed Sediment Based on 50 Years of Knowledge

Production of a global sedimentological seabed map has been initiated in 1995 to provide the necessary tool for searches of aircraft and boats lost at sea, to give sedimentary information for nautical charts, and to provide input data for acoustic propagation modelling. This original approach had already been initiated one century ago when the French hydrographic service and the University of Nancy had produced maps of the distribution of marine sediments of the French coasts and then sediment maps of the continental shelves of Europe and North America. The current map of the sediment of oceans presented was initiated with a UNESCO's general map of the deep ocean floor. This map was adapted using a unique sediment classification to present all types of sediments: from beaches to the deep seabed and from glacial deposits to tropical sediments. In order to allow good visualization and to be adapted to the different applications, only the granularity of sediments is represented. The published seabed maps are studied, if they present an interest, the nature of the seabed is extracted from them, the sediment classification is transcribed and the resulted map is integrated in the world map. Data come also from interpretations of Multibeam Echo Sounder (MES) imagery of large hydrographic surveys of deep-ocean. These allow a very high-quality mapping of areas that until then were represented as homogeneous. The third and principal source of data comes from the integration of regional maps produced specifically for this project. These regional maps are carried out using all the bathymetric and sedimentary data of a region. This step makes it possible to produce a regional synthesis map, with the realization of generalizations in the case of over-precise data. 86 regional maps of the Atlantic Ocean, the Mediterranean Sea, and the Indian Ocean have been produced and integrated into the world sedimentary map. This work is permanent and permits a digital version every two years, with the integration of some new maps. This article describes the choices made in terms of sediment classification, the scale of source data and the zonation of the variability of the quality. This map is the final step in a system comprising the Shom Sedimentary Database, enriched by more than one million punctual and surface items of data, and four series of coastal seabed maps at 1:10,000, 1:50,000, 1:200,000 and 1:1,000,000. This step by step approach makes it possible to take into account the progresses in knowledge made in the field of seabed characterization during the last decades. Thus, the arrival of new classification systems for seafloor has improved the recent seabed maps, and the compilation of these new maps with those previously published allows a gradual enrichment of the world sedimentary map. But there is still a lot of work to enhance some regions, which are still based on data acquired more than half a century ago.

Cellulose Nanocrystals Suspensions as Water-Based Lubricants for Slurry Pump Gland Seals

The tribological tests were performed on a new tribometer, in order to measure the coefficient of friction of a gland seal packing material on stainless steel shafts in presence of Cellulose Nanocrystal (CNC) suspension as a sustainable, environmentally friendly, water-based lubricant. To simulate the real situation from the slurry pumps, silica sands were used as slurry particles. The surface profiles after tests were measured by interferometer microscope to characterize the surface wear. Moreover, the coefficient of friction and surface wear were measured between stainless steel shaft and chrome steel ball to investigate the tribological effects of CNC in boundary lubrication region. Alignment of nanoparticles in the CNC suspensions are the main reason for friction and wear reduction. The homogeneous concentrated suspensions showed fingerprint patterns of a chiral nematic liquid crystal. These properties made CNC a very good lubricant additive in water.

Optimization of Biodiesel Production from Sunflower Oil Using Central Composite Design

The current study investigated the effect of catalyst ratio and methanol to oil ratio on biodiesel production by using central composite design. Biodiesel was produced by transesterification using sodium hydroxide as a homogeneous catalyst, a laboratory scale reactor consisting of flat bottom flask mounts with a reflux condenser and a heating plate was used to produce biodiesel. Key parameters, including, time, temperature and mixing rate were kept constant at 60 minutes, 60 oC and 600 RPM, respectively. From the results obtained, it was observed that the biodiesel yield depends on catalyst ratio and methanol to oil ratio. The highest yield of 50.65% was obtained at catalyst ratio of 0.5 wt.% and methanol to oil mole ratio 10.5. The analysis of variances of biodiesel yield showed the R Squared value of 0.8387. A quadratic mathematical model was developed to predict the biodiesel yield in the specified parameters ranges.

Accuracy of Autonomy Navigation of Unmanned Aircraft Systems through Imagery

The Unmanned Aircraft Systems (UAS) usually navigate through the Global Navigation Satellite System (GNSS) associated with an Inertial Navigation System (INS). However, GNSS can have its accuracy degraded at any time or even turn off the signal of GNSS. In addition, there is the possibility of malicious interferences, known as jamming. Therefore, the image navigation system can solve the autonomy problem, because if the GNSS is disabled or degraded, the image navigation system would continue to provide coordinate information for the INS, allowing the autonomy of the system. This work aims to evaluate the accuracy of the positioning though photogrammetry concepts. The methodology uses orthophotos and Digital Surface Models (DSM) as a reference to represent the object space and photograph obtained during the flight to represent the image space. For the calculation of the coordinates of the perspective center and camera attitudes, it is necessary to know the coordinates of homologous points in the object space (orthophoto coordinates and DSM altitude) and image space (column and line of the photograph). So if it is possible to automatically identify in real time the homologous points the coordinates and attitudes can be calculated whit their respective accuracies. With the methodology applied in this work, it is possible to verify maximum errors in the order of 0.5 m in the positioning and 0.6º in the attitude of the camera, so the navigation through the image can reach values equal to or higher than the GNSS receivers without differential correction. Therefore, navigating through the image is a good alternative to enable autonomous navigation.

miR-200c as a Biomarker for 5-FU Chemosensitivity in Colorectal Cancer

5-FU is a chemotherapeutic agent that has been used in colorectal cancer (CRC) treatment. However, it is usually associated with the acquired resistance, which decreases the therapeutic effects of 5-FU. miR-200c is involved in chemotherapeutic drug resistance, but its mechanism is not fully understood. In this study, the effect of inhibition of miR-200c in sensitivity of HCT-116 CRC cells to 5-FU was evaluated. HCT-116 cells were transfected with LNA-anti- miR-200c for 48 h. mRNA expression of miR-200c was evaluated using quantitative real- time PCR. The protein expression of phosphatase and tensin homolog (PTEN) and E-cadherin were analyzed by western blotting. Annexin V and propidium iodide staining assay were applied for apoptosis detection. The caspase-3 activation was evaluated by an enzymatic assay. The results showed LNA-anti-miR-200c inhibited the expression of PTEN and E-cadherin protein, apoptosis and activation of caspase 3 compared with control cells. In conclusion, these results suggest that miR-200c as a prognostic marker can overcome to 5-FU chemoresistance in CRC.

The Effect of Information vs. Reasoning Gap Tasks on the Frequency of Conversational Strategies and Accuracy in Speaking among Iranian Intermediate EFL Learners

Speaking skills merit meticulous attention both on the side of the learners and the teachers. In particular, accuracy is a critical component to guarantee the messages to be conveyed through conversation because a wrongful change may adversely alter the content and purpose of the talk. Different types of tasks have served teachers to meet numerous educational objectives. Besides, negotiation of meaning and the use of different strategies have been areas of concern in socio-cultural theories of SLA. Negotiation of meaning is among the conversational processes which have a crucial role in facilitating the understanding and expression of meaning in a given second language. Conversational strategies are used during interaction when there is a breakdown in communication that leads to the interlocutor attempting to remedy the gap through talk. Therefore, this study was an attempt to investigate if there was any significant difference between the effect of reasoning gap tasks and information gap tasks on the frequency of conversational strategies used in negotiation of meaning in classrooms on one hand, and on the accuracy in speaking of Iranian intermediate EFL learners on the other. After a pilot study to check the practicality of the treatments, at the outset of the main study, the Preliminary English Test was administered to ensure the homogeneity of 87 out of 107 participants who attended the intact classes of a 15 session term in one control and two experimental groups. Also, speaking sections of PET were used as pretest and posttest to examine their speaking accuracy. The tests were recorded and transcribed to estimate the percentage of the number of the clauses with no grammatical errors in the total produced clauses to measure the speaking accuracy. In all groups, the grammatical points of accuracy were instructed and the use of conversational strategies was practiced. Then, different kinds of reasoning gap tasks (matchmaking, deciding on the course of action, and working out a time table) and information gap tasks (restoring an incomplete chart, spot the differences, arranging sentences into stories, and guessing game) were manipulated in experimental groups during treatment sessions, and the students were required to practice conversational strategies when doing speaking tasks. The conversations throughout the terms were recorded and transcribed to count the frequency of the conversational strategies used in all groups. The results of statistical analysis demonstrated that applying both the reasoning gap tasks and information gap tasks significantly affected the frequency of conversational strategies through negotiation. In the face of the improvements, the reasoning gap tasks had a more significant impact on encouraging the negotiation of meaning and increasing the number of conversational frequencies every session. The findings also indicated both task types could help learners significantly improve their speaking accuracy. Here, applying the reasoning gap tasks was more effective than the information gap tasks in improving the level of learners’ speaking accuracy.

Scorbot-ER 4U Using Forward Kinematics Modelling and Analysis

Robotic arm manipulators are widely used to accomplish many kinds of tasks. SCORBOT-ER 4u is a 5-degree of freedom (DOF) vertical articulated educational robotic arm, and all joints are revolute. It is specifically designed to perform pick and place task with its gripper. The pick and place task consists of consideration of the end effector coordinate of the robotic arm and the desired position coordinate in its workspace. This paper describes about forward kinematics modeling and analysis of the robotic end effector motion through joint space. The kinematics problems are defined by the transformation from the Cartesian space to the joint space. Denavit-Hartenberg (D-H) model is used in order to model the robotic links and joints with 4x4 homogeneous matrix. The forward kinematics model is also developed and simulated in MATLAB. The mathematical model is validated by using robotic toolbox in MATLAB. By using this method, it may be applicable to get the end effector coordinate of this robotic arm and other similar types to this arm. The software development of SCORBOT-ER 4u is also described here. PC-and EtherCAT based control technology from BECKHOFF is used to control the arm to express the pick and place task.

Considerations for Effectively Using Probability of Failure as a Means of Slope Design Appraisal for Homogeneous and Heterogeneous Rock Masses

Probability of failure (PF) often appears alongside factor of safety (FS) in design acceptance criteria for rock slope, underground excavation and open pit mine designs. However, the design acceptance criteria generally provide no guidance relating to how PF should be calculated for homogeneous and heterogeneous rock masses, or what qualifies a ‘reasonable’ PF assessment for a given slope design. Observational and kinematic methods were widely used in the 1990s until advances in computing permitted the routine use of numerical modelling. In the 2000s and early 2010s, PF in numerical models was generally calculated using the point estimate method. More recently, some limit equilibrium analysis software offer statistical parameter inputs along with Monte-Carlo or Latin-Hypercube sampling methods to automatically calculate PF. Factors including rock type and density, weathering and alteration, intact rock strength, rock mass quality and shear strength, the location and orientation of geologic structure, shear strength of geologic structure and groundwater pore pressure influence the stability of rock slopes. Significant engineering and geological judgment, interpretation and data interpolation is usually applied in determining these factors and amalgamating them into a geotechnical model which can then be analysed. Most factors are estimated ‘approximately’ or with allowances for some variability rather than ‘exactly’. When it comes to numerical modelling, some of these factors are then treated deterministically (i.e. as exact values), while others have probabilistic inputs based on the user’s discretion and understanding of the problem being analysed. This paper discusses the importance of understanding the key aspects of slope design for homogeneous and heterogeneous rock masses and how they can be translated into reasonable PF assessments where the data permits. A case study from a large open pit gold mine in a complex geological setting in Western Australia is presented to illustrate how PF can be calculated using different methods and obtain markedly different results. Ultimately sound engineering judgement and logic is often required to decipher the true meaning and significance (if any) of some PF results.

Impact of Fluid Flow Patterns on Metastable Zone Width of Borax in Dual Radial Impeller Crystallizer at Different Impeller Spacings

Conducting crystallization in an agitated vessel requires a proper selection of mixing parameters that would result in a production of crystals of specific properties. In dual impeller systems, which are characterized by a more complex hydrodynamics due to the possible fluid flow interactions, revealing a clear link between mixing parameters and crystallization kinetics is still an open issue. The aim of this work is to establish this connection by investigating how fluid flow patterns, generated by two impellers mounted on the same shaft, reflect on metastable zone width of borax decahydrate, one of the most important parameters of the crystallization process. Investigation was carried out in a 15-dm3 bench scale batch cooling crystallizer with an aspect ratio (H/T) equal to 1.3. For this reason, two radial straight blade turbines (4-SBT) were used for agitation. Experiments were conducted at different impeller spacings at the state of complete suspension. During the process of an unseeded batch cooling crystallization, solution temperature and supersaturation were continuously monitored what enabled a determination of the metastable zone width. Hydrodynamic conditions in the vessel achieved at different impeller spacings investigated were analyzed in detail. This was done firstly by measuring the mixing time required to attain the desired level of homogeneity. Secondly, fluid flow patterns generated in a described dual impeller system were both photographed and simulated by VisiMix Turbulent software. Also, a comparison of these two visualization methods was performed. Experimentally obtained results showed that metastable zone width is definitely affected by the hydrodynamics in the crystallizer. This means that this crystallization parameter can be controlled not only by adjusting the saturation temperature or cooling rate, as is usually done, but also by choosing a suitable impeller spacing that will result in a formation of crystals of wanted size distribution.

Generative Syntaxes: Macro-Heterophony and the Form of ‘Synchrony’

One of the most powerful language innovation in the twentieth century music was the heterophony–hypostasis of the vertical syntax entered into the sphere of interest of many composers, such as George Enescu, Pierre Boulez, Mauricio Kagel, György Ligeti and others. The heterophonic syntax has a history of its growth, which means a succession of different concepts and writing techniques. The trajectory of settling this phenomenon does not necessarily take into account the chronology: there are highly complex primary stages and advanced stages of returning to the simple forms of writing. In folklore, the plurimelodic simultaneities are free or random and originate from the (unintentional) differences/‘deviations’ from the state of unison, through a variety of ornaments, melismas, imitations, elongations and abbreviations, all in a flexible rhythmic and non-periodic/immeasurable framework, proper to the parlando-rubato rhythmics. Within the general framework of the multivocal organization, the heterophonic syntax in elaborate (academic) version has imposed itself relatively late compared with polyphony and homophony. Of course, the explanation is simple, if we consider the causal relationship between the sound vocabulary elements – in this case, the modalism – and the typologies of vertical organization appropriate for it. Therefore, adding up the ‘classic’ pathway of the writing typologies (monody – polyphony – homophony), heterophony - applied equally to the structures of modal, serial or synthesis vocabulary – reclaims necessarily an own macrotemporal form, in the sense of the analogies enshrined by the evolution of the musical styles and languages: polyphony→fugue, homophony→sonata. Concerned about the prospect of edifying a new musical ontology, the composer Ştefan Niculescu experienced – along with the mathematical organization of heterophony according to his own original methods – the possibility of extrapolation of this phenomenon in macrostructural plan, reaching this way to the unique form of ‘synchrony’. Founded on coincidentia oppositorum principle (involving the ‘one-multiple’ binom), the sound architecture imagined by Ştefan Niculescu consists in one (temporal) model / algorithm of articulation of two sound states: 1. monovocality state (principle of identity) and 2. multivocality state (principle of difference). In this context, the heterophony becomes an (auto)generative mechanism, with macrotemporal amplitude, strategy that will be grown by the composer, practically throughout his creation (see the works: Ison I, Ison II, Unisonos I, Unisonos II, Duplum, Triplum, Psalmus, Héterophonies pour Montreux (Homages to Enescu and Bartók etc.). For the present demonstration, we selected one of the most edifying works of Ştefan Niculescu – Simphony II, Opus dacicum – where the form of (heterophony-)synchrony acquires monumental-symphonic features, representing an emblematic case for the complexity level achieved by this type of vertical syntax in the twentieth century music.

Invariant Characters of Tolerance Class and Reduction under Homomorphism in IIS

Some invariant properties of incomplete information systems homomorphism are studied in this paper. Demand conditions of tolerance class, attribute reduction, indispensable attribute and dispensable attribute being invariant under homomorphism in incomplete information system are revealed and discussed. The existing condition of endohomomorphism on an incomplete information system is also explored. It establishes some theoretical foundations for further investigations on incomplete information systems in rough set theory, like in information systems.

Performance of Bridge Approach Slabs in Bridge Construction: A Case Study

Long-term differential settlement between the bridge structure and the bridge embankment typically results in an abrupt grade change, causing driver discomfort, impairing driver safety, and exerting a potentially excessive impact traffic loading on the abutment. This paper has analysed a case of study showing the effect of an approaching slab realized in a bridge constructed at Tirane-Elbasan Motorway. The layer thickness under the slab is modeled as homogenous, the slab is a reinforced concrete structure and over that the asphaltic layers take place. Analysis indicates that reinforced concrete approaching slab distributes the stresses quite uniformly into the road fill layers and settlements varies in a range less than 2.50 cm in the total slab length of 6.00 m with a maximum slope of 1/240. Results taken from analytical analysis are compared with topographic measurements done on field and they carry great similarities.