Evaluation of Corrosion in Steel Reinforced Concrete with Brick Waste

The massive demolition of old buildings in recent years has generated tons of waste, especially brick waste. Thus, a concern of recent research is the use of this waste for the production of environmentally friendly concrete. At the same time, corrosion of the reinforcement steel rebar in classical concrete is a current problem. In this context, in the present paper a study was carried out on the corrosion of metal reinforcement in cement mortars with added brick waste. The corrosion process was analyzed on four compositions of mortars without and with 15%, 25% and 35% brick waste replacing the sand. The brick waste has majority content in SiO2, Al2O3, FeO3 and CaO. The grain size distribution of brick waste was close to that of the sand (dmax = 2 mm). The preparation method of the samples was similar to ordinary mortars. The corrosion action on the rebar in concrete, at different brick waste concentrations, was investigated by electrochemical measurements (polarization curves and electrochemical impedance spectroscopy (EIS)) at 1 month and 26 months. The results obtained at 26 months revealed that the addition of the brick waste in mortar improved the anticorrosion properties in the case of all samples compared with the etalon mortar. The best results were obtained in the case of the sample with 15% brick waste (the efficiency was ≈ 90%). The corrosion intermediary layer formed on the rebar surface was evidenced by SEM-EDX.

Engineering Education for Sustainable Development in China: Perceptions Bias between Experienced Engineers and Engineering Students

Nowadays sustainable development has increasingly become an important research topic of engineering education all over the world. Engineering Education for Sustainable Development (EESD) highlighted the importance of addressing sustainable development in engineering practice. However, whether and how the professional engineering learning and experience affect those perceptions is an interesting research topic especially in Chinese context. Our study fills this gap by investigating perceptions bias of EESD among first-grade engineering students, fourth-grade engineering students and experienced engineers using a triple-dimensional model. Our goal is to find the effect of engineering learning and experience on sustainable development and make these learning and experiences more accessible for students and engineers in school and workplace context. The data (n = 138) came from a Likert questionnaire based on the triple-dimensional model of EESD adopted from literature reviews and the data contain 48 first-grade students, 56 fourth-grade students and 34 engineers with rich working experience from Environmental Engineering, Energy Engineering, Chemical Engineering and Civil Engineering in or graduated from Zhejiang University, China. One-way ANOVA analysis was used to find the difference in different dimensions among the three groups. The statistical results show that both engineering students and engineers have a well understanding of sustainable development in ecology dimension of EESD while there are significant differences among three groups as to the socio-economy and value rationality dimensions of EESD. The findings provide empirical evidence that both engineering learning and professional engineering experience are helpful to cultivate the cognition and perception of sustainable development in engineering education. The results of this work indicate that more practical content should be added to students’ engineering education while more theoretical content should be added to engineers’ training in order to promote the engineering students’ and engineers’ perceptions of sustainable development. In addition, as to the design of engineering courses and professional practice system for sustainable development, we should not only pay attention to the ecological aspects, but also emphasize the coordination of ecological, socio-economic and human-centered sustainable development (e.g., engineer's ethical responsibility).

Methane versus Carbon Dioxide: Mitigation Prospects

Atmospheric carbon dioxide (CO2) has dominated the discussion around the causes of climate change. This is a reflection of a 100-year time horizon for all greenhouse gases that became a norm.  The 100-year time horizon is much too long – and yet, almost all mitigation efforts, including those set in the near-term frame of within 30 years, are still geared toward it. In this paper, we show that for a 30-year time horizon, methane (CH4) is the greenhouse gas whose radiative forcing exceeds that of CO2. In our analysis, we use the radiative forcing of greenhouse gases in the atmosphere, because they directly affect the rise in temperature on Earth. We found that in 2019, the radiative forcing (RF) of methane was ~2.5 W/m2 and that of carbon dioxide was ~2.1 W/m2. Under a business-as-usual (BAU) scenario until 2050, such forcing would be ~2.8 W/m2 and ~3.1 W/m2 respectively. There is a substantial spread in the data for anthropogenic and natural methane (CH4) emissions, along with natural gas, (which is primarily CH4), leakages from industrial production to consumption. For this reason, we estimate the minimum and maximum effects of a reduction of these leakages, and assume an effective immediate reduction by 80%. Such action may serve to reduce the annual radiative forcing of all CH4 emissions by ~15% to ~30%. This translates into a reduction of RF by 2050 from ~2.8 W/m2 to ~2.5 W/m2 in the case of the minimum effect that can be expected, and to ~2.15 W/m2 in the case of the maximum effort to reduce methane leakages. Under the BAU, we find that the RF of CO2 will increase from ~2.1 W/m2 now to ~3.1 W/m2 by 2050. We assume a linear reduction of 50% in anthropogenic emission over the course of the next 30 years, which would reduce the radiative forcing of CO2 from ~3.1 W/m2 to ~2.9 W/m2. In the case of "net zero," the other 50% of only anthropogenic CO2 emissions reduction would be limited to being either from sources of emissions or directly from the atmosphere. In this instance, the total reduction would be from ~3.1 W/m2 to ~2.7 W/m2, or ~0.4 W/m2. To achieve the same radiative forcing as in the scenario of maximum reduction of methane leakages of ~2.15 W/m2, an additional reduction of radiative forcing of CO2 would be approximately 2.7 -2.15 = 0.55 W/m2. In total, one would need to remove ~660 GT of CO2 from the atmosphere in order to match the maximum reduction of current methane leakages, and ~270 GT of CO2 from emitting sources, to reach "negative emissions". This amounts to over 900 GT of CO2.

Simulation and Assessment of Carbon Dioxide Separation by Piperazine Blended Solutions Using E-NRTL and Peng-Robinson Models: A Study of Regeneration Heat Duty

High pressure carbon dioxide (CO2) absorption from a specific off-gas in a conventional column has been evaluated for the environmental concerns by the Aspen HYSYS simulator using a wide range of single absorbents and piperazine (PZ) blended solutions to estimate the outlet CO2 concentration, CO2 loading, reboiler power supply and regeneration heat duty to choose the most efficient solution in terms of CO2 removal and required heat duty. The property package, which is compatible with all applied solutions for the simulation in this study, estimates the properties based on electrolyte non-random two-liquid (E-NRTL) model for electrolyte thermodynamics and Peng-Robinson equation of state for vapor phase and liquid hydrocarbon phase properties. The results of the simulation indicate that PZ in addition to the mixture of PZ and monoethanolamine (MEA) demand the highest regeneration heat duty compared with other studied single and blended amine solutions respectively. The blended amine solutions with the lowest PZ concentrations (5wt% and 10wt%) were considered and compared to reduce the cost of process, among which the blended solution of 10wt%PZ+35wt%MDEA (methyldiethanolamine) was found as the most appropriate solution in terms of CO2 content in the outlet gas, rich-CO2 loading and regeneration heat duty.

6D Posture Estimation of Road Vehicles from Color Images

Currently, in the field of object posture estimation, there is research on estimating the position and angle of an object by storing a 3D model of the object to be estimated in advance in a computer and matching it with the model. However, in this research, we have succeeded in creating a module that is much simpler, smaller in scale, and faster in operation. Our 6D pose estimation model consists of two different networks – a classification network and a regression network. From a single RGB image, the trained model estimates the class of the object in the image, the coordinates of the object, and its rotation angle in 3D space. In addition, we compared the estimation accuracy of each camera position, i.e., the angle from which the object was captured. The highest accuracy was recorded when the camera position was 75°, the accuracy of the classification was about 87.3%, and that of regression was about 98.9%.

Numerical and Experimental Analyses of a Semi-Active Pendulum Tuned Mass Damper

Modern structures such as floor systems, pedestrian bridges and high-rise buildings have become lighter in mass and more flexible with negligible damping and thus prone to vibration. In this paper, a semi-actively controlled pendulum tuned mass dampers (PTMD) is presented that uses air springs as both the restoring (resilient) and energy dissipating (damping) elements; the tuned mass damper (TMD) uses no passive dampers. The proposed PTMD can readily be fine-tuned and re-tuned, via software, without changing any hardware. Almost all existing semi-active systems have the three elements that passive TMDs have, i.e., inertia, resilient, and dissipative elements with some adjustability built into one or two of these elements. The proposed semi-active air suspended TMD, on the other hand, is made up of only inertia and resilience elements. A notable feature of this TMD is the absence of a physical damping element in its make-up. The required viscous damping is introduced into the TMD using a semi-active control scheme residing in a micro-controller which actuates a high-speed proportional valve regulating the flow of air in and out of the air springs. In addition to introducing damping into the TMD, the semi-active control scheme adjusts the stiffness of the TMD. The focus of this work has been the synthesis and analysis of the control algorithms and strategies to vary the tuning accuracy, introduce damping into air suspended PTMD, and enable the PTMD to self-tune itself. The accelerations of the main structure and PTMD as well as the pressure in the air springs are used as the feedback signals in control strategies. Numerical simulation and experimental evaluation of the proposed tuned damping system are presented in this paper.

Designing for Inclusion within the Learning Management System: Social Justice, Identities, and Online Design for Digital Spaces in Higher Education

The aim of this paper is to propose pedagogical design for learning management systems (LMS) that offers greater inclusion for students based on a number of theoretical perspectives and delineated through an example. Considering the impact of COVID-19, including on student mental health, the research suggesting the importance of student sense of belonging on retention, success, and student well-being, the author describes intentional LMS design incorporating theoretically based practices informed by critical theory, feminist theory, indigenous theory and practices, and new materiality. This article considers important aspects of these theories and practices which attend to inclusion, identities, and socially just learning environments. Additionally, increasing student sense of belonging and mental health through LMS design influenced by adult learning theory and the community of inquiry model are described.  The process of thinking through LMS pedagogical design with inclusion intentionally in mind affords the opportunity to allow LMS to go beyond course use as a repository of documents, to an intentional community of practice that facilitates belonging and connection, something much needed in our times. In virtual learning environments it has been harder to discern how students are doing, especially in feeling connected to their courses, their faculty, and their student peers. Increasingly at the forefront of public universities is addressing the needs of students with multiple and intersecting identities and the multiplicity of needs and accommodations. Education in 2020, and moving forward, calls for embedding critical theories and inclusive ideals and pedagogies to the ways instructors design and teach in online platforms. Through utilization of critical theoretical frameworks and instructional practices, students may experience the LMS as a welcoming place with intentional plans for welcoming diversity in identities.

Lamb Wave Wireless Communication in Healthy Plates Using Coherent Demodulation

Guided ultrasonic waves are used in Non-Destructive Testing and Structural Health Monitoring for inspection and damage detection. Recently, wireless data transmission using ultrasonic waves in solid metallic channels has gained popularity in some industrial applications such as nuclear, aerospace and smart vehicles. The idea is to find a good substitute for electromagnetic waves since they are highly attenuated near metallic components due to Faraday shielding. The proposed solution is to use ultrasonic guided waves such as Lamb waves as an information carrier due to their capability of propagation for long distances. In addition to this, valuable information about the health of the structure could be extracted simultaneously. In this work, the reliable frequency bandwidth for communication is extracted experimentally from dispersion curves at first. Then, an experimental platform for wireless communication using Lamb waves is described and built. After this, coherent demodulation algorithm used in telecommunications is tested for Amplitude Shift Keying, On-Off Keying and Binary Phase Shift Keying modulation techniques. Signal processing parameters such as threshold choice, number of cycles per bit and Bit Rate are optimized. Experimental results are compared based on the average bit error percentage. Results has shown high sensitivity to threshold selection for Amplitude Shift Keying and On-Off Keying techniques resulting a Bit Rate decrease. Binary Phase Shift Keying technique shows the highest stability and data rate between all tested modulation techniques.

Experimental Study on the Variation of Young's Modulus of Hollow Clay Brick Obtained from Static and Dynamic Tests

In parallel with the appearance of new materials, brick masonry had and still has an essential part of the construction market today, with new technical challenges in designing bricks to meet additional requirements. Being used in structural applications, predicting the performance of clay brick masonry allows a significant cost reduction, in terms of practical experimentation. The behavior of masonry walls depends on the behavior of their elementary components, such as bricks, joints, and coatings. Therefore, it is necessary to consider it at different scales (from the scale of the intrinsic material to the real scale of the wall) and then to develop appropriate models, using numerical simulations. The work presented in this paper focuses on the mechanical characterization of the terracotta material at ambient temperature. As a result, the static Young’s modulus obtained from the flexural test shows different values in comparison with the compression test, as well as with the dynamic Young’s modulus obtained from the Impulse excitation of vibration test. Moreover, the Young's modulus varies according to the direction in which samples are extracted, where the values in the extrusion direction diverge from the ones in the orthogonal directions. Based on these results, hollow bricks can be considered as transversely isotropic bimodulus material.

The Impact of Alumina Cement on Properties of Portland Cement Slurries and Mortars

The addition of a small amount of alumina cement to Portland cement results in immediate setting, a rapid increase in the compressive strength and a clear increase of the adhesion to concrete substrate. This phenomenon is used, among others, for the production of liquid floor self-levelling compounds. Alumina cement is several times more expensive than Portland cement and is a component having a significant impact on prices of products manufactured with its use. For the production of liquid floor self-levelling compounds, low-alumina cement containing approximately 40% Al2O3 is normally used. The aim of the study was to determine the impact of Portland cement with the addition of alumina cement on the basic physical and mechanical properties of cement slurries and mortars. CEM I 42.5R and three types of alumina cement containing 40%, 50% and 70% of Al2O3 were used for the tests. Mixes containing 4%, 6%, 8%, 10% and 12% of different varieties of alumina cement were prepared; for which, the time of initial and final setting, compressive and flexural strength and adhesion to concrete substrate were determined. The analysis of the obtained test results showed that a similar immediate setting effect and clearly better adhesion strength can be obtained using the addition of 6% of high-alumina cement than 12% of low-alumina cement. As the prices of these cements are similar, this can give significant financial savings in the production of liquid floor self-levelling compounds.

Systematic Examination of Methods Supporting the Social Innovation Process

Innovation is the key element of economic development and a key factor in social processes. Technical innovations can be identified as prerequisites and causes of social change and cannot be created without the renewal of society. The study of social innovation can be characterised as one of the significant research areas of our day. The study’s aim is to identify the process of social innovation, which can be defined by input, transformation, and output factors. This approach divides the social innovation process into three parts: situation analysis, implementation, follow-up. The methods associated with each stage of the process are illustrated by the chronological line of social innovation. In this study, we have sought to present methodologies that support long- and short-term decision-making that is easy to apply, have different complementary content, and are well visualised for different user groups. When applying the methods, the reference objects are different: county, district, settlement, specific organisation. The solution proposed by the study supports the development of a methodological combination adapted to different situations. Having reviewed metric and conceptualisation issues, we wanted to develop a methodological combination along with a change management logic suitable for structured support to the generation of social innovation in the case of a locality or a specific organisation. In addition to a theoretical summary, in the second part of the study, we want to give a non-exhaustive picture of the two counties located in the north-eastern part of Hungary through specific analyses and case descriptions.

A Real-Time Bayesian Decision-Support System for Predicting Suspect Vehicle’s Intended Target Using a Sparse Camera Network

We present a decision-support tool to assist an operator in the detection and tracking of a suspect vehicle traveling to an unknown target destination. Multiple data sources, such as traffic cameras, traffic information, weather, etc., are integrated and processed in real-time to infer a suspect’s intended destination chosen from a list of pre-determined high-value targets. Previously, we presented our work in the detection and tracking of vehicles using traffic and airborne cameras. Here, we focus on the fusion and processing of that information to predict a suspect’s behavior. The network of cameras is represented by a directional graph, where the edges correspond to direct road connections between the nodes and the edge weights are proportional to the average time it takes to travel from one node to another. For our experiments, we construct our graph based on the greater Los Angeles subset of the Caltrans’s “Performance Measurement System” (PeMS) dataset. We propose a Bayesian approach where a posterior probability for each target is continuously updated based on detections of the suspect in the live video feeds. Additionally, we introduce the concept of ‘soft interventions’, inspired by the field of Causal Inference. Soft interventions are herein defined as interventions that do not immediately interfere with the suspect’s movements; rather, a soft intervention may induce the suspect into making a new decision, ultimately making their intent more transparent. For example, a soft intervention could be temporarily closing a road a few blocks from the suspect’s current location, which may require the suspect to change their current course. The objective of these interventions is to gain the maximum amount of information about the suspect’s intent in the shortest possible time. Our system currently operates in a human-on-the-loop mode where at each step, a set of recommendations are presented to the operator to aid in decision-making. In principle, the system could operate autonomously, only prompting the operator for critical decisions, allowing the system to significantly scale up to larger areas and multiple suspects. Once the intended target is identified with sufficient confidence, the vehicle is reported to the authorities to take further action. Other recommendations include a selection of road closures, i.e., soft interventions, or to continue monitoring. We evaluate the performance of the proposed system using simulated scenarios where the suspect, starting at random locations, takes a noisy shortest path to their intended target. In all scenarios, the suspect’s intended target is unknown to our system. The decision thresholds are selected to maximize the chances of determining the suspect’s intended target in the minimum amount of time and with the smallest number of interventions. We conclude by discussing the limitations of our current approach to motivate a machine learning approach, based on reinforcement learning in order to relax some of the current limiting assumptions.

Toward Discovering an Architectural Typology Based on the Theory of Affordance

This paper revolves around the concept of affordance. It aims to discover and develop an architectural typology based on the ecological concept of affordance. In order to achieve this aim, an analytical study is conducted and two sources were taken into account: 1- Gibson's definition of the concept of affordance and 2- The researches that are concerned on the affordance categorisation. As a result, this paper concluded 16 typologies of affordances, including the possibilities of mixing them based on both sources. To clarify these typologies and provide further understanding, a wide range of architectural examples are presented and proposed in the paper. To prove this vocabulary’s capability to diagnose and evaluate the affordance of different environments, an experimental study with two processes have been adapted: 1. Diagnostic process: the interpretation of the environments with regards to its affordance by using the new vocabulary (the developed typologies). 2. Evaluating process: the evaluation of the environments that have been interpreted and classified with regards to their affordances. By using the measures of emotional experience (the positive affect ‘PA’ and the negative affect ‘NA’) and the architectural evaluation criteria (beauty, economy and function). The experimental study proves that the typologies are capable of reading the affordance within different environments. Additionally, it explains how these different typologies reflect different interactions based on the previous processes. The data which are concluded from the evaluation of measures explain how different typologies of affordance that have already reflected different environments had different evaluations. In fact, some of them are recommended while the others are not. In other words, the paper draws a roadmap for designers to diagnose, evaluate and analyse the affordance into different architectural environments. After that, it guides them through adapting the best interaction (affordance category), which they intend to adapt into their proposed designs.

Cantilever Shoring Piles with Prestressing Strands: An Experimental Approach

Underground space is becoming a necessity nowadays, especially in highly congested urban areas. Retaining underground excavations using shoring systems is essential in order to protect adjoining structures from potential damage or collapse. Reinforced Concrete Piles (RCP) supported by multiple rows of tie-back anchors are commonly used type of shoring systems in deep excavations. However, executing anchors can sometimes be challenging because they might illegally trespass neighboring properties or get obstructed by infrastructure and other underground facilities. A technique is proposed in this paper, and it involves the addition of eccentric high-strength steel strands to the RCP section through ducts without providing the pile with lateral supports. The strands are then vertically stressed externally on the pile cap using a hydraulic jack, creating a compressive strengthening force in the concrete section. An experimental study about the behavior of the shoring wall by pre-stressed piles is presented during the execution of an open excavation in an urban area (Beirut city) followed by numerical analysis using finite element software. Based on the experimental results, this technique is proven to be cost-effective and provides flexible and sustainable construction of shoring works.

A Review on Process Parameters of Ti/Al Dissimilar Joint Using Laser Beam Welding

The use of laser beam welding for joining titanium and aluminum offers more advantages compared with conventional joining processes. Dissimilar metal combination is very much needed for aircraft structural industries and research activities. The quality of a weld joint is directly influenced by the welding input parameters. The common problem that is faced by the manufactures is the control of the process parameters to obtain a good weld joint with minimal detrimental. To overcome this issue, various parameters can be preferred to obtain quality of weld joint. In this present study an overall literature review on processing parameters such as offset distance, welding speed, laser power, shielding gas and filler metals are discussed with the effects on quality weldment. Additionally, mechanical properties of welds joint are discussed. The aim of the report is to review the recent progress in the welding of dissimilar titanium (Ti) and aluminum (Al) alloys to provide a basis for follow up research.

Generative Adversarial Network Based Fingerprint Anti-Spoofing Limitations

Fingerprint Anti-Spoofing approaches have been actively developed and applied in real-world applications. One of the main problems for Fingerprint Anti-Spoofing is not robust to unseen samples, especially in real-world scenarios. A possible solution will be to generate artificial, but realistic fingerprint samples and use them for training in order to achieve good generalization. This paper contains experimental and comparative results with currently popular GAN based methods and uses realistic synthesis of fingerprints in training in order to increase the performance. Among various GAN models, the most popular StyleGAN is used for the experiments. The CNN models were first trained with the dataset that did not contain generated fake images and the accuracy along with the mean average error rate were recorded. Then, the fake generated images (fake images of live fingerprints and fake images of spoof fingerprints) were each combined with the original images (real images of live fingerprints and real images of spoof fingerprints), and various CNN models were trained. The best performances for each CNN model, trained with the dataset of generated fake images and each time the accuracy and the mean average error rate, were recorded. We observe that current GAN based approaches need significant improvements for the Anti-Spoofing performance, although the overall quality of the synthesized fingerprints seems to be reasonable. We include the analysis of this performance degradation, especially with a small number of samples. In addition, we suggest several approaches towards improved generalization with a small number of samples, by focusing on what GAN based approaches should learn and should not learn.

Loss Function Optimization for CNN-Based Fingerprint Anti-Spoofing

As biometric systems become widely deployed, the security of identification systems can be easily attacked by various spoof materials. This paper contributes to finding a reliable and practical anti-spoofing method using Convolutional Neural Networks (CNNs) based on the types of loss functions and optimizers. The types of CNNs used in this paper include AlexNet, VGGNet, and ResNet. By using various loss functions including Cross-Entropy, Center Loss, Cosine Proximity, and Hinge Loss, and various loss optimizers which include Adam, SGD, RMSProp, Adadelta, Adagrad, and Nadam, we obtained significant performance changes. We realize that choosing the correct loss function for each model is crucial since different loss functions lead to different errors on the same evaluation. By using a subset of the Livdet 2017 database, we validate our approach to compare the generalization power. It is important to note that we use a subset of LiveDet and the database is the same across all training and testing for each model. This way, we can compare the performance, in terms of generalization, for the unseen data across all different models. The best CNN (AlexNet) with the appropriate loss function and optimizers result in more than 3% of performance gain over the other CNN models with the default loss function and optimizer. In addition to the highest generalization performance, this paper also contains the models with high accuracy associated with parameters and mean average error rates to find the model that consumes the least memory and computation time for training and testing. Although AlexNet has less complexity over other CNN models, it is proven to be very efficient. For practical anti-spoofing systems, the deployed version should use a small amount of memory and should run very fast with high anti-spoofing performance. For our deployed version on smartphones, additional processing steps, such as quantization and pruning algorithms, have been applied in our final model.

Impact of Climate Change on Sea Level Rise along the Coastline of Mumbai City, India

Sea-level rise being one of the most important impacts of anthropogenic induced climate change resulting from global warming and melting of icebergs at Arctic and Antarctic, the investigations done by various researchers both on Indian Coast and elsewhere during the last decade has been reviewed in this paper. The paper aims to ascertain the propensity of consistency of different suggested methods to predict the near-accurate future sea level rise along the coast of Mumbai. Case studies at East Coast, Southern Tip and West and South West coast of India have been reviewed. Coastal Vulnerability Index of several important international places has been compared, which matched with Intergovernmental Panel on Climate Change forecasts. The application of Geographic Information System mapping, use of remote sensing technology, both Multi Spectral Scanner and Thematic Mapping data from Landsat classified through Iterative Self-Organizing Data Analysis Technique for arriving at high, moderate and low Coastal Vulnerability Index at various important coastal cities have been observed. Instead of data driven, hindcast based forecast for Significant Wave Height, additional impact of sea level rise has been suggested. Efficacy and limitations of numerical methods vis-à-vis Artificial Neural Network has been assessed, importance of Root Mean Square error on numerical results is mentioned. Comparing between various computerized methods on forecast results obtained from MIKE 21 has been opined to be more reliable than Delft 3D model.

Transformation of Industrial Policy towards Industry 4.0 and Its Impact on Firms' Competition

Although Europe is on the threshold of a new industrial revolution called Industry 4.0, many believe that this will increase the flexibility of production, the mass adaptation of products to consumers and the speed of their service; it will also improve product quality and dramatically increase productivity. However, as expected, all the benefits of Industry 4.0 face many of the inevitable changes and challenges they pose. One of them is the inevitable transformation of current competition and business models. This article examines the possible results of competitive conversion from the classic Bertrand and Cournot models to qualitatively new competition based on innovation. Ability to deliver a new product quickly and the possibility to produce the individual design (through flexible and quickly configurable factories) by reducing equipment failures and increasing process automation and control is highly important. This study shows that the ongoing transformation of the competition model is changing the game. This, together with the creation of complex value networks, means huge investments that make it particularly difficult for small and medium-sized enterprises. In addition, the ongoing digitalization of data raises new concerns regarding legal obligations, intellectual property, and security.

Modified Genome-Scale Metabolic Model of Escherichia coli by Adding Hyaluronic Acid Biosynthesis-Related Enzymes (GLMU2 and HYAD) from Pasteurella multocida

Hyaluronic acid (HA) consists of linear heteropolysaccharides repeat of D-glucuronic acid and N-acetyl-D-glucosamine. HA has various useful properties to maintain skin elasticity and moisture, reduce inflammation, and lubricate the movement of various body parts without causing immunogenic allergy. HA can be found in several animal tissues as well as in the capsule component of some bacteria including Pasteurella multocida. This study aimed to modify a genome-scale metabolic model of Escherichia coli using computational simulation and flux analysis methods to predict HA productivity under different carbon sources and nitrogen supplement by the addition of two enzymes (GLMU2 and HYAD) from P. multocida to improve the HA production under the specified amount of carbon sources and nitrogen supplements. Result revealed that threonine and aspartate supplement raised the HA production by 12.186%. Our analyses proposed the genome-scale metabolic model is useful for improving the HA production and narrows the number of conditions to be tested further.