FRP Bars Spacing Effect on Numerical Thermal Deformations in Concrete Beams under High Temperatures

5 In order to eradicate the degradation of reinforced concrete structures due to the steel corrosion, professionals in constructions suggest using fiber reinforced polymers (FRP) for their excellent properties. Nevertheless, high temperatures may affect the bond between FRP bar and concrete, and consequently the serviceability of FRP-reinforced concrete structures. This paper presents a nonlinear numerical investigation using ADINA software to investigate the effect of the spacing between glass FRP (GFRP) bars embedded in concrete on circumferential thermal deformations and the distribution of radial thermal cracks in reinforced concrete beams submitted to high temperature variations up to 60 °C for asymmetrical problems. The thermal deformations predicted from nonlinear finite elements model, at the FRP bar/concrete interface and at the external surface of concrete cover, were established as a function of the ratio of concrete cover thickness to FRP bar diameter (c/db) and the ratio of spacing between FRP bars in concrete to FRP bar diameter (e/db). Numerical results show that the circumferential thermal deformations at the external surface of concrete cover are linear until cracking thermal load varied from 32 to 55 °C corresponding to the ratio of e/db varied from 1.3 to 2.3, respectively. However, for ratios e/db >2.3 and c/db >1.6, the thermal deformations at the external surface of concrete cover exhibit linear behavior without any cracks observed on the specified surface. The numerical results are compared to those obtained from analytical models validated by experimental tests.

Providing a Practical Model to Reduce Maintenance Costs: A Case Study in GeG Company

In the past, we could increase profit by increasing product prices. But in the new decade, a competitive market does not let us to increase profit with increased prices. Therefore, the only way to increase profit will be to reduce costs. A significant percentage of production costs are the maintenance costs, and analysis of these costs could achieve more profit. Most maintenance strategies such as RCM (Reliability-Center-Maintenance), TPM (Total Productivity Maintenance), PM (Preventive Maintenance) and etc., are trying to reduce maintenance costs. In this paper, decreasing the maintenance costs of Concentration Plant of Golgohar Iron Ore Mining & Industrial Company (GeG) was examined by using of MTBF (Mean Time Between Failures) and MTTR (Mean Time To Repair) analyses. These analyses showed that instead of buying new machines and increasing costs in order to promote capacity, the improving of MTBF and MTTR indexes would solve capacity problems in the best way and decrease costs.

Biodegradability Evaluation of Polylactic Acid Composite with Natural Fiber (Sisal)

Due to increasing environmental pressure for biodegradable products, especially in polymeric materials, in order to meet the demands of the biological cycles of the circular economy, new materials have been developed as a sustainability strategy. This study proposes a composite material developed from the biodegradable polymer PLA Ecovio® (polylactic acid - PLA) with natural sisal fibers, where the soybean ester was used as a plasticizer, which can aid in adhesion between the materials and fibers, making the most attractive final composite from an environmental point of view. The composites were obtained by extrusion. The materials tests were produced and submitted to biodegradation tests. Through the biodegradation tests, it can be seen that the biodegradable polymer composition with 5% sisal fiber presented about 12.4% more biodegradability compared to the polymer without fiber addition. It has also been found that the plasticizer was not a compatible with fibers and the polymer. Finally, fibers help to anticipate the decomposition process of the material when subjected to conditions of a landfill. Therefore, its intrinsic properties are not affected during its use, only the biodegradation process begins after its exposure to landfill conditions.

Safety Assessment of Traditional Ready-to-Eat Meat Products Vended at Retail Outlets in Kebbi and Sokoto States, Nigeria

Food safety is a significant and growing public health problem in the world and Nigeria as a developing country, since food-borne diseases are important contributors to the huge burden of sickness and death of humans. In Nigeria, traditional ready-to-eat meat products (RTE-MPs) like balangu, tsire, guru and dried meat products like kilishi, dambun nama, banda, were reported to be highly appreciated because of their eating qualities. The consumption of these products was considered as safe due to the treatments that are usually involved during their production process. However, during processing and handling, the products could be contaminated by pathogens that could cause food poisoning. Therefore, a hazard identification for pathogenic bacteria on some traditional RTE-MPs was conducted in Kebbi and Sokoto States, Nigeria. A total of 116 RTE-MPs (balangu-38, kilishi-39 and tsire-39) samples were obtained from retail outlets and analyzed using standard cultural microbiological procedures in general and selective enrichment media to isolate the target pathogens. A six-fold serial dilution was prepared and using the pour plating method, colonies were counted. Serial dilutions were selected based on the prepared pre-labeled Petri dishes for each sample. A volume of 10-12 ml of molten Nutrient agar cooled to 42-45°C was poured into each Petri dish and 1 ml each from dilutions of 102, 104 and 106 for every sample was respectively poured on a pre-labeled Petri plate after which colonies were counted. The isolated pathogens were identified and confirmed after series of biochemical tests. Frequencies and percentages were used to describe the presence of pathogens. The General Linear Model was used to analyze data on pathogen presence according to RTE-MPs and means were separated using the Tukey test at 0.05 confidence level. Of the 116 RTE-MPs samples collected, 35 (30.17%) samples were found to be contaminated with some tested pathogens. Prevalence results showed that Escherichia coli, salmonella and Staphylococcus aureus were present in the samples. Mean total bacterial count was 23.82×106 cfu/g. The frequency of individual pathogens isolated was; Staphylococcus aureus 18 (15.51%), Escherichia coli 12 (10.34%) and Salmonella 5 (4.31%). Also, among the RTE-MPs tested, the total bacterial counts were found to differ significantly (P < 0.05), with 1.81, 2.41 and 2.9×104 cfu/g for tsire, kilishi, and balangu, respectively. The study concluded that the presence of pathogenic bacteria in balangu could pose grave health risks to consumers, and hence, recommended good manufacturing practices in the production of balangu to improve the products’ safety.

Hydrogel Based on Cellulose Acetate Used as Scaffold for Cell Growth

A hydrogel from cellulose acetate cross linked with ethylenediaminetetraacetic dianhydride (HAC-EDTA) was synthesized by our research group, and submitted to characterization and biological tests. Cytocompatibility analysis was performed by confocal microscopy using human adipocyte derived stem cells (ASCs). The FTIR analysis showed characteristic bands of cellulose acetate and hydroxyl groups and the tensile tests evidence that HAC-EDTA present a Young’s modulus of 643.7 MPa. The confocal analysis revealed that there was cell growth at the surface of HAC-EDTA. After one day of culture the cells presented spherical morphology, which may be caused by stress of the sequestration of Ca2+ and Mg2+ ions at the cell medium by HAC-EDTA, as demonstrated by ICP-MS. However, after seven days and 14 days of culture, the cells present fibroblastoid morphology, phenotype expected by this cellular type. The results give efforts to indicate this new material as a potential biomaterial for tissue engineering, in the future in vivo approach.

Estimation of PM2.5 Emissions and Source Apportionment Using Receptor and Dispersion Models

Source apportionment using Dispersion model depends primarily on the quality of Emission Inventory. In the present study, a CMB receptor model has been used to identify the sources of PM2.5, while the AERMOD dispersion model has been used to account for missing sources of PM2.5 in the Emission Inventory. A statistical approach has been developed to quantify the missing sources not considered in the Emission Inventory. The inventory of each grid was improved by adjusting emissions based on road lengths and deficit in measured and modelled concentrations. The results showed that in CMB analyses, fugitive sources - soil and road dust - contribute significantly to ambient PM2.5 pollution. As a result, AERMOD significantly underestimated the ambient air concentration at most locations. The revised Emission Inventory showed a significant improvement in AERMOD performance which is evident through statistical tests.

Experimental Investigation of Visual Comfort Requirement in Garment Factories and Identify the Cost Saving Opportunities

Visual comfort is one of the major parameters that can be taken to measure the human comfort in any environment. If the provided illuminance level in a working environment does not meet the workers visual comfort, it will lead to eye-strain, fatigue, headache, stress, accidents and finally, poor productivity. However, improvements in lighting do not necessarily mean that the workplace requires more light. Unnecessarily higher illuminance levels will also cause poor visual comfort and health risks. In addition, more power consumption on lighting will also result in higher energy costs. So, during this study, visual comfort and the illuminance requirement for the workers in textile/apparel industry were studied to perform different tasks (i.e. cutting, sewing and knitting) at their workplace. Experimental studies were designed to identify the optimum illuminance requirement depending upon the varied fabric colour and type and finally, energy saving potentials due to controlled illuminance level depending on the workforce requirement were analysed. Visual performance of workers during the sewing operation was studied using the ‘landolt ring experiment’. It was revealed that around 36.3% of the workers would like to work if the illuminance level varies from 601 lux to 850 lux illuminance level and 45.9% of the workers are not happy to work if the illuminance level reduces less than 600 lux and greater than 850 lux. Moreover, more than 65% of the workers who do not satisfy with the existing illuminance levels of the production floors suggested that they have headache, eye diseases, or both diseases due to poor visual comfort. In addition, findings of the energy analysis revealed that the energy-saving potential of 5%, 10%, 24%, 8% and 16% can be anticipated for fabric colours, red, blue, yellow, black and white respectively, when the 800 lux is the prevailing illuminance level for sewing operation.

Evaluation of TRIS-DMA-NVP Hydrogels for Making Silicone-Based Contact Lenses

In this study, contact lenses were prepared through the polymerization of tris-(trimethyl-silyl-propyl-methacrylate) (TRIS), N,N-dimethylacrylamide (DMA), N-vinylpyrrolidone (NVP), and cross-linked with ethylene glycol dimethylacrylate (EGDMA). The equilibrium water content (EWC), oxygen permeability (Dk), light transmittance, and in vitro cytotoxicity of TRIS-DMA-NVP with various ratios were measured. The results showed that the EWC increased while the Dk decreased with the increase of NVP content. For the sample with 25 wt% NVP, the EWC attained 53% whereas the Dk decreased to 46 barrers. All these lenses exhibited light transmittance over than 95%. In addition, all these lenses exhibited no inhibition to the growth of L292 fibroblasts. Thus, this study showed that TRIS-DMA-NVP can be applicable for making contact lens.

Valorization of Residues from Forest Industry for the Generation of Energy

The use of biomass to produce renewable energy is one of the forms that can be used to reduce the impact of energy production. Like any other energy resource, there are limitations for biomass use, and it must compete not only with fossil fuels but also with other renewable energy sources such as solar or wind energy. Combustion is currently the most efficient and widely used waste-to-energy process, in the areas where direct use of biomass is possible, without the need to make large transfers of raw material. Many industrial facilities can use agricultural or forestry waste, straw, chips, bagasse, etc. in their thermal systems without making major transformations or adjustments in the feeding to the ovens, making this waste an attractive and cost-effective option in terms of availability, access, and costs. In spite of the facilities and benefits, the environmental reasons (emission of gases and particulate material) are decisive for its use for energy purpose. This paper describes a valorization of residues from forest industry to generate energy, using a case study.

A Hybrid Algorithm for Collaborative Transportation Planning among Carriers

In this paper, there is concentration on collaborative transportation planning (CTP) among multiple carriers with pickup and delivery requests and time windows. This problem is a vehicle routing problem with constraints from standard vehicle routing problems and new constraints from a real-world application. In the problem, each carrier has a finite number of vehicles, and each request is a pickup and delivery request with time window. Moreover, each carrier has reserved requests, which must be served by itself, whereas its exchangeable requests can be outsourced to and served by other carriers. This collaboration among carriers can help them to reduce total transportation costs. A mixed integer programming model is proposed to the problem. To solve the model, a hybrid algorithm that combines Genetic Algorithm and Simulated Annealing (GASA) is proposed. This algorithm takes advantages of GASA at the same time. After tuning the parameters of the algorithm with the Taguchi method, the experiments are conducted and experimental results are provided for the hybrid algorithm. The results are compared with those obtained by a commercial solver. The comparison indicates that the GASA significantly outperforms the commercial solver.

Sensitivity Analysis of the Heat Exchanger Design in Net Power Oxy-Combustion Cycle for Carbon Capture

The global warming and its impact on climate change is one of main challenges for current century. Global warming is mainly due to the emission of greenhouse gases (GHG) and carbon dioxide (CO2) is known to be the major contributor to the GHG emission profile. Whilst the energy sector is the primary source for CO2 emission, Carbon Capture and Storage (CCS) are believed to be the solution for controlling this emission. Oxyfuel combustion (Oxy-combustion) is one of the major technologies for capturing CO2 from power plants. For gas turbines, several Oxy-combustion power cycles (Oxyturbine cycles) have been investigated by means of thermodynamic analysis. NetPower cycle is one of the leading oxyturbine power cycles with almost full carbon capture capability from a natural gas fired power plant. In this manuscript, sensitivity analysis of the heat exchanger design in NetPower cycle is completed by means of process modelling. The heat capacity variation and supercritical CO2 with gaseous admixtures are considered for multi-zone analysis with Aspen Plus software. It is found that the heat exchanger design has a major role to increase the efficiency of NetPower cycle. The pinch-point analysis is done to extract the composite and grand composite curve for the heat exchanger. In this paper, relationship between the cycle efficiency and the minimum approach temperature (∆Tmin) of the heat exchanger has also been evaluated.  Increase in ∆Tmin causes a decrease in the temperature of the recycle flue gases (RFG) and an overall decrease in the required power for the recycled gas compressor. The main challenge in the design of heat exchangers in power plants is a tradeoff between the capital and operational costs. To achieve lower ∆Tmin, larger size of heat exchanger is required. This means a higher capital cost but leading to a better heat recovery and lower operational cost. To achieve this, ∆Tmin is selected from the minimum point in the diagrams of capital and operational costs. This study provides an insight into the NetPower Oxy-combustion cycle’s performance analysis and operational condition based on its heat exchanger design.

Impact of Positive Psychology Education and Interventions on Well-Being: A Study of Students Engaged in Pastoral Care

Positive psychology investigates human strengths and virtues and promotes well-being. Relying on this assumption, positive interventions have been continuously designed to build pleasure and happiness, joy and contentment, engagement and meaning, hope and optimism, satisfaction and gratitude, spirituality, and various other positive measures of well-being. In line with this model of positive psychology and interventions, this study investigated certain measures of well-being in a group of 45 students enrolled in an 18-week positive psychology course and simultaneously engaged in service-oriented interventions that they chose for themselves based on the course content and individual interests. Students’ well-being was measured at the beginning and end of the course. The well-being indicators included positive automatic thoughts, optimism and hope, satisfaction with life, and spirituality. A paired-samples t-test conducted to evaluate the impact of class content and service-oriented interventions on students’ scores of well-being indicators indicated statistically significant increase from pre-class to post-class scores. There were also significant gender differences in post-course well-being scores, with females having higher levels of well-being than males. A two-way between groups analysis of variance indicated a significant interaction effect of age by gender on the post-course well-being scores, with females in the age group of 56-65 having the highest scores of well-being in comparison to the males in the same age group. Regression analyses indicated that positive automatic thought significantly predicted hope and satisfaction with life in the pre-course analysis. In the post-course regression analysis, spiritual transcendence made a significant contribution to optimism, and positive automatic thought made a significant contribution to both hope and satisfaction with life. Finally, a significant test between pre-course and post-course regression coefficients indicated that the regression coefficients at pre-course were significantly different from post-course coefficients, suggesting that the positive psychology course and the interventions were helpful in raising the levels of well-being. The overall results suggest a substantial increase in the participants’ well-being scores after engaging in the positive-oriented interventions, implying a need for designing more positive interventions in education to promote well-being.  

Early Diagnosis of Alzheimer's Disease Using a Combination of Images Processing and Brain Signals

Alzheimer's prevalence is on the rise, and the disease comes with problems like cessation of treatment, high cost of treatment, and the lack of early detection methods. The pathology of this disease causes the formation of protein deposits in the brain of patients called plaque amyloid. Generally, the diagnosis of this disease is done by performing tests such as a cerebrospinal fluid, CT scan, MRI, and spinal cord fluid testing, or mental testing tests and eye tracing tests. In this paper, we tried to use the Medial Temporal Atrophy (MTA) method and the Leave One Out (LOO) cycle to extract the statistical properties of the three Fz, Pz, and Cz channels of ERP signals for early diagnosis of this disease. In the process of CT scan images, the accuracy of the results is 81% for the healthy person and 88% for the severe patient. After the process of ERP signaling, the accuracy of the results for a healthy person in the delta band in the Cz channel is 81% and in the alpha band the Pz channel is 90%. In the results obtained from the signal processing, the results of the severe patient in the delta band of the Cz channel were 89% and in the alpha band Pz channel 92%.

Thermalytix: An Advanced Artificial Intelligence Based Solution for Non-Contact Breast Screening

Diagnosis of breast cancer at early stages has seen better clinical and survival outcomes. Survival rates in developing countries like India are very low due to accessibility and affordability issues of screening tests such as Mammography. In addition, Mammography is not much effective in younger women with dense breasts. This leaves a gap in current screening methods. Thermalytix is a new technique for detecting breast abnormality in a non-contact, non-invasive way. It is an AI-enabled computer-aided diagnosis solution that automates interpretation of high resolution thermal images and identifies potential malignant lesions. The solution is low cost, easy to use, portable and is effective in all age groups.  This paper presents the results of a retrospective comparative analysis of Thermalytix over Mammography and Clinical Breast Examination for breast cancer screening. Thermalytix was found to have better sensitivity than both the tests, with good specificity as well. In addition, Thermalytix identified all malignant patients without palpable lumps.

Risk Based Maintenance Planning for Loading Equipment in Underground Hard Rock Mine: Case Study

Mining industry is known for its appetite to spend sizeable capital on mine equipment. However, in the current scenario, the mining industry is challenged by daunting factors of non-uniform geological conditions, uneven ore grade, uncontrollable and volatile mineral commodity prices and the ever increasing quest to optimize the capital and operational costs. Thus, the role of equipment reliability and maintenance planning inherits a significant role in augmenting the equipment availability for the operation and in turn boosting the mine productivity. This paper presents the Risk Based Maintenance (RBM) planning conducted on mine loading equipment namely Load Haul Dumpers (LHDs) at Vedanta Resources Ltd subsidiary Hindustan Zinc Limited operated Sindesar Khurd Mines, an underground zinc and lead mine situated in Dariba, Rajasthan, India. The mining equipment at the location is maintained by the Original Equipment Manufacturers (OEMs) namely Sandvik and Atlas Copco, who carry out the maintenance and inspection operations for the equipment. Based on the downtime data extracted for the equipment fleet over the period of 6 months spanning from 1st January 2017 until 30th June 2017, it was revealed that significant contribution of three downtime issues related to namely Engine, Hydraulics, and Transmission to be common among all the loading equipment fleet and substantiated by Pareto Analysis. Further scrutiny through Bubble Matrix Analysis of the given factors revealed the major influence of selective factors namely Overheating, No Load Taken (NTL) issues, Gear Changing issues and Hose Puncture and leakage issues. Utilizing the equipment wise analysis of all the downtime factors obtained, spares consumed, and the alarm logs extracted from the machines, technical design changes in the equipment and pre shift critical alarms checklist were proposed for the equipment maintenance. The given analysis is beneficial to allow OEMs or mine management to focus on the critical issues hampering the reliability of mine equipment and design necessary maintenance strategies to mitigate them.

Comparing Test Equating by Item Response Theory and Raw Score Methods with Small Sample Sizes on a Study of the ARTé: Mecenas Learning Game

The purpose of the present research is to equate two test forms as part of a study to evaluate the educational effectiveness of the ARTé: Mecenas art history learning game. The researcher applied Item Response Theory (IRT) procedures to calculate item, test, and mean-sigma equating parameters. With the sample size n=134, test parameters indicated “good” model fit but low Test Information Functions and more acute than expected equating parameters. Therefore, the researcher applied equipercentile equating and linear equating to raw scores and compared the equated form parameters and effect sizes from each method. Item scaling in IRT enables the researcher to select a subset of well-discriminating items. The mean-sigma step produces a mean-slope adjustment from the anchor items, which was used to scale the score on the new form (Form R) to the reference form (Form Q) scale. In equipercentile equating, scores are adjusted to align the proportion of scores in each quintile segment. Linear equating produces a mean-slope adjustment, which was applied to all core items on the new form. The study followed a quasi-experimental design with purposeful sampling of students enrolled in a college level art history course (n=134) and counterbalancing design to distribute both forms on the pre- and posttests. The Experimental Group (n=82) was asked to play ARTé: Mecenas online and complete Level 4 of the game within a two-week period; 37 participants completed Level 4. Over the same period, the Control Group (n=52) did not play the game. The researcher examined between group differences from post-test scores on test Form Q and Form R by full-factorial Two-Way ANOVA. The raw score analysis indicated a 1.29% direct effect of form, which was statistically non-significant but may be practically significant. The researcher repeated the between group differences analysis with all three equating methods. For the IRT mean-sigma adjusted scores, form had a direct effect of 8.39%. Mean-sigma equating with a small sample may have resulted in inaccurate equating parameters. Equipercentile equating aligned test means and standard deviations, but resultant skewness and kurtosis worsened compared to raw score parameters. Form had a 3.18% direct effect. Linear equating produced the lowest Form effect, approaching 0%. Using linearly equated scores, the researcher conducted an ANCOVA to examine the effect size in terms of prior knowledge. The between group effect size for the Control Group versus Experimental Group participants who completed the game was 14.39% with a 4.77% effect size attributed to pre-test score. Playing and completing the game increased art history knowledge, and individuals with low prior knowledge tended to gain more from pre- to post test. Ultimately, researchers should approach test equating based on their theoretical stance on Classical Test Theory and IRT and the respective  assumptions. Regardless of the approach or method, test equating requires a representative sample of sufficient size. With small sample sizes, the application of a range of equating approaches can expose item and test features for review, inform interpretation, and identify paths for improving instruments for future study.

Three-Dimensional, Non-Linear Finite Element Analysis of Bullet Penetration through Thin AISI 4340 Steel Target Plate

Bullet penetration in steel plate is investigated with the help of three-dimensional, non-linear, transient, dynamic, finite elements analysis using explicit time integration code LSDYNA. The effect of large strain, strain-rate and temperature at very high velocity regime was studied from number of simulations of semi-spherical nose shape bullet penetration through single layered circular plate with 2 mm thickness at impact velocities of 500, 1000, and 1500 m/s with the help of Johnson Cook material model. Mie-Gruneisen equation of state is used in conjunction with Johnson Cook material model to determine pressure-volume relationship at various points of interests. Two material models viz. Plastic-Kinematic and Johnson- Cook resulted in different deformation patterns in steel plate. It is observed from the simulation results that the velocity drop and loss of kinetic energy occurred very quickly up to perforation of plate, after that the change in velocity and changes in kinetic energy are negligibly small. The physics behind this kind of behaviour is presented in the paper.

Investigations of Flame Retardant Properties of Beneficiated Huntite and Hydromagnesite Mineral Reinforced Polymer Composites

Huntite and hydromagnesite minerals have been used as additive materials to achieve incombustible material due to their inflammability property. Those fire retardants materials can help to extinguish in the early stages of fire. Thus dispersion of the flame can be prevented even if the fire started. Huntite and hydromagnesite minerals are known to impart fire-proofing of the polymer composites. However, the additives used in the applications led to deterioration in the mechanical properties due to the usage of high amount of the powders in the composites. In this study, by enriching huntite and hydromagnesite, it was aimed to use purer minerals to reinforce the polymer composites. Thus, predictably, using purer mineral will lead to use lower amount of mineral powders. By this manner, the minerals free from impurities by various processes were added to the polymer matrix with different loading level and grades. Different types of samples were manufactured, and subsequently characterized by XRD, SEM-EDS, XRF and flame-retardant tests. Tensile strength and elongation at break values were determined according to loading levels and grades. Besides, a comparison on the properties of the polymer composites produced by using of minerals with and without impurities was performed. As a result of the work, it was concluded that it is required to use beneficiated minerals to provide better fire-proofing behaviors in the polymer composites.

Evaluation of Wind Fragility for Set Anchor Used in Sign Structure in Korea

Recently, damage to domestic facilities by strong winds and typhoons are growing. Therefore, this study focused on sign structure among various vulnerable facilities. The evaluation of the wind fragility was carried out considering the destruction of the anchor, which is one of the various failure modes of the sign structure. The performance evaluation of the anchor was carried out to derive the wind fragility. Two parameters were set and four anchor types were selected to perform the pull-out and shear tests. The resistance capacity was estimated based on the experimental results. Wind loads were estimated using Monte Carlo simulation method. Based on these results, we derived the wind fragility according to anchor type and wind exposure category. Finally, the evaluation of the wind fragility was performed according to the experimental parameters such as anchor length and anchor diameter. This study shows that the depth of anchor was more significant for the safety of structure compare to diameter of anchor.

Complementing Assessment Processes with Standardized Tests: A Work in Progress

ABET accredited programs must assess the development of student learning outcomes (SOs) in engineering programs. Different institutions implement different strategies for this assessment, and they are usually designed “in house.” This paper presents a proposal for including standardized tests to complement the ABET assessment model in an engineering college made up of six distinct engineering programs. The engineering college formulated a model of quality assurance in education to be implemented throughout the six engineering programs to regularly assess and evaluate the achievement of SOs in each program offered. The model uses diverse techniques and sources of data to assess student performance and to implement actions of improvement based on the results of this assessment. The model is called “Assessment Process Model” and it includes SOs A through K, as defined by ABET. SOs can be divided into two categories: “hard skills” and “professional skills” (soft skills). The first includes abilities, such as: applying knowledge of mathematics, science, and engineering and designing and conducting experiments, as well as analyzing and interpreting data. The second category, “professional skills”, includes communicating effectively, and understanding professional and ethnical responsibility. Within the Assessment Process Model, various tools were used to assess SOs, related to both “hard” as well as “soft” skills. The assessment tools designed included: rubrics, surveys, questionnaires, and portfolios. In addition to these instruments, the Engineering College decided to use tools that systematically gather consistent quantitative data. For this reason, an in-house exam was designed and implemented, based on the curriculum of each program. Even though this exam was administered during various academic periods, it is not currently considered standardized. In 2017, the Engineering College included three standardized tests: one to assess mathematical and scientific reasoning and two more to assess reading and writing abilities. With these exams, the college hopes to obtain complementary information that can help better measure the development of both hard and soft skills of students in the different engineering programs. In the first semester of 2017, the three exams were given to three sample groups of students from the six different engineering programs. Students in the sample groups were either from the first, fifth, and tenth semester cohorts. At the time of submission of this paper, the engineering college has descriptive statistical data and is working with various statisticians to have a more in-depth and detailed analysis of the sample group of students’ achievement on the three exams. The overall objective of including standardized exams in the assessment model is to identify more precisely the least developed SOs in order to define and implement educational strategies necessary for students to achieve them in each engineering program.