Modelling and Simulation of the Freezing Systems and Heat Pumps Using Unisim® Design

The paper describes the modeling and simulation of the heat pumps domain processes. The main objective of the study is the use of the heat pump in propene–propane distillation processes. The modeling and simulation instrument is the Unisim® Design simulator. The paper is structured in three parts: An overview of the compressing gases, the modeling and simulation of the freezing systems, and the modeling and simulation of the heat pumps. For each of these systems, there are presented the Unisim® Design simulation diagrams, the input–output system structure and the numerical results. Future studies will consider modeling and simulation of the propene–propane distillation process with heat pump.

Leukocyte Detection Using Image Stitching and Color Overlapping Windows

Blood cell analysis plays a significant role in the diagnosis of human health. As an alternative to the traditional technique conducted by laboratory technicians, this paper presents an automatic white blood cell (leukocyte) detection system using Image Stitching and Color Overlapping Windows. The advantage of this method is to present a detection technique of white blood cells that are robust to imperfect shapes of blood cells with various image qualities. The input for this application is images from a microscope-slide translation video. The preprocessing stage is performed by stitching the input images. First, the overlapping parts of the images are determined, then stitching and blending processes of two input images are performed. Next, the Color Overlapping Windows is performed for white blood cell detection which consists of color filtering, window candidate checking, window marking, finds window overlaps, and window cropping processes. Experimental results show that this method could achieve an average of 82.12% detection accuracy of the leukocyte images.

A Paradigm Shift towards Personalized and Scalable Product Development and Lifecycle Management Systems in the Aerospace Industry

Integrated systems for product design, manufacturing, and lifecycle management are difficult to implement and customize. Commercial software vendors, including CAD/CAM and third party PDM/PLM developers, create user interfaces and functionality that allow their products to be applied across many industries. The result is that systems become overloaded with functionality, difficult to navigate, and use terminology that is unfamiliar to engineers and production personnel. For example, manufacturers of automotive, aeronautical, electronics, and household products use similar but distinct methods and processes. Furthermore, each company tends to have their own preferred tools and programs for controlling work and information flow and that connect design, planning, and manufacturing processes to business applications. This paper presents a methodology and a case study that addresses these issues and suggests that in the future more companies will develop personalized applications that fit to the natural way that their business operates. A functioning system has been implemented at a highly competitive U.S. aerospace tooling and component supplier that works with many prominent airline manufacturers around the world including The Boeing Company, Airbus, Embraer, and Bombardier Aerospace. During the last three years, the program has produced significant benefits such as the automatic creation and management of component and assembly designs (parametric models and drawings), the extensive use of lightweight 3D data, and changes to the way projects are executed from beginning to end. CATIA (CAD/CAE/CAM) and a variety of programs developed in C#, VB.Net, HTML, and SQL make up the current system. The web-based platform is facilitating collaborative work across multiple sites around the world and improving communications with customers and suppliers. This work demonstrates that the creative use of Application Programming Interface (API) utilities, libraries, and methods is a key to automating many time-consuming tasks and linking applications together.

Role of Organic Wastewater Constituents in Iron Redox Cycling for Ferric Sludge Reuse in the Fenton-Based Treatment

The practical application of the Fenton-based treatment method for organic compounds-contaminated water purification is limited mainly because of the large amount of ferric sludge formed during the treatment, where ferrous iron (Fe(II)) is used as the activator of the hydrogen peroxide oxidation processes. Reuse of ferric sludge collected from clarifiers to substitute Fe(II) salts allows reducing the total cost of Fenton-type treatment technologies and minimizing the accumulation of hazardous ferric waste. Dissolution of ferric iron (Fe(III)) from the sludge to liquid phase at acidic pH and autocatalytic transformation of Fe(III) to Fe(II) by phenolic compounds (tannic acid, lignin, phenol, catechol, pyrogallol and hydroquinone) added or present as water/wastewater constituents were found to be essentially involved in the Fenton-based oxidation mechanism. Observed enhanced formation of highly reactive species, hydroxyl radicals, resulted in a substantial organic contaminant degradation increase. Sludge reuse at acidic pH and in the presence of ferric iron reductants is a novel strategy in the Fenton-based treatment application for organic compounds-contaminated water purification.

Business-Intelligence Mining of Large Decentralized Multimedia Datasets with a Distributed Multi-Agent System

The rapid generation of high volume and a broad variety of data from the application of new technologies pose challenges for the generation of business-intelligence. Most organizations and business owners need to extract data from multiple sources and apply analytical methods for the purposes of developing their business. Therefore, the recently decentralized data management environment is relying on a distributed computing paradigm. While data are stored in highly distributed systems, the implementation of distributed data-mining techniques is a challenge. The aim of this technique is to gather knowledge from every domain and all the datasets stemming from distributed resources. As agent technologies offer significant contributions for managing the complexity of distributed systems, we consider this for next-generation data-mining processes. To demonstrate agent-based business intelligence operations, we use agent-oriented modeling techniques to develop a new artifact for mining massive datasets.

Precision Grinding of Titanium (Ti-6Al-4V) Alloy Using Nanolubrication

In this current era of competitive machinery productions, the industries are designed to place more emphasis on the product quality and reduction of cost whilst abiding by the pollution-preventing policy. In attempting to delve into the concerns, the industries are aware that the effectiveness of existing lubrication systems must be improved to achieve power-efficient and pollution-preventing machining processes. As such, this research is targeted to study on a plausible solution to the issue in grinding titanium alloy (Ti-6Al-4V) by using nanolubrication, as an alternative to flood grinding. The aim of this research is to evaluate the optimum condition of grinding force and surface roughness using MQL lubricating system to deliver nano-oil at different level of weight concentration of Silicon Dioxide (SiO2) mixed normal mineral oil. Taguchi Design of Experiment (DoE) method is carried out using a standard Taguchi orthogonal array of L16(43) to find the optimized combination of weight concentration mixture of SiO2, nozzle orientation and pressure of MQL. Surface roughness and grinding force are also analyzed using signal-to-noise(S/N) ratio to determine the best level of each factor that are tested. Consequently, the best combination of parameters is tested for a period of time and the results are compared with conventional grinding method of dry and flood condition. The results show a positive performance of MQL nanolubrication.  

Importance of Knowledge in the Interdisciplinary Production Processes of Innovative Medical Tools

Processes of production of innovative medical tools have interdisciplinary character. They consist of direct and indirect close cooperation of specialists of different scientific branches. The Knowledge they have seems to be important for undertaken design, construction and manufacturing processes. The Knowledge exchange between participants of these processes is therefore crucial for the final result, which are innovative medical products. The paper draws attention to the necessity of feedback from the end user to the designer / manufacturer of medical tools which will allow for more accurate understanding of user needs. The study describes prerequisites of production processes of innovative medical (surgical) tools including participants and category of knowledge resources occurring in these processes. They are the result of research in selected Polish organizations involved in the production of medical instruments and are the basis for further work on the development of knowledge sharing model in interdisciplinary teams geographically dispersed.

Continuous Fixed Bed Reactor Application for Decolourization of Textile Effluent by Adsorption on NaOH Treated Eggshell

Fixed bed adsorption has become a frequently used industrial application in wastewater treatment processes. Various low cost adsorbents have been studied for their applicability in treatment of different types of effluents. In this work, the intention of the study was to explore the efficacy and feasibility for azo dye, Acid Orange 7 (AO7) adsorption onto fixed bed column of NaOH Treated eggshell (TES). The effect of various parameters like flow rate, initial dye concentration, and bed height were exploited in this study. The studies confirmed that the breakthrough curves were dependent on flow rate, initial dye concentration solution of AO7 and bed depth. The Thomas, Yoon–Nelson, and Adams and Bohart models were analysed to evaluate the column adsorption performance. The adsorption capacity, rate constant and correlation coefficient associated to each model for column adsorption was calculated and mentioned. The column experimental data were fitted well with Thomas model with coefficients of correlation R2 ≥0.93 at different conditions but the Yoon–Nelson, BDST and Bohart–Adams model (R2=0.911), predicted poor performance of fixed-bed column. The (TES) was shown to be suitable adsorbent for adsorption of AO7 using fixed-bed adsorption column.

Knowledge Management Strategies within a Corporate Environment of Papers

Knowledge transfer between personnel could benefit an organization’s improved competitive advantage in the marketplace from a strategic approach to knowledge management. The lack of information sharing between personnel could create knowledge transfer gaps while restricting the decision-making processes. Knowledge transfer between personnel can potentially improve information sharing based on an implemented knowledge management strategy. An organization’s capacity to gain more knowledge is aligned with the organization’s prior or existing captured knowledge. This case study attempted to understand the overall influence of a KMS within the corporate environment and knowledge exchange between personnel. The significance of this study was to help understand how organizations can improve the Return on Investment (ROI) of a knowledge management strategy within a knowledge-centric organization. A qualitative descriptive case study was the research design selected for this study. The lack of information sharing between personnel may create knowledge transfer gaps while restricting the decision-making processes. Developing a knowledge management strategy acceptable at all levels of the organization requires cooperation in support of a common organizational goal. Working with management and executive members to develop a protocol where knowledge transfer becomes a standard practice in multiple tiers of the organization. The knowledge transfer process could be measurable when focusing on specific elements of the organizational process, including personnel transition to help reduce time required understanding the job. The organization studied in this research acknowledged the need for improved knowledge management activities within the organization to help organize, retain, and distribute information throughout the workforce. Data produced from the study indicate three main themes including information management, organizational culture, and knowledge sharing within the workforce by the participants. These themes indicate a possible connection between an organizations KMS, the organizations culture, knowledge sharing, and knowledge transfer.

Questions Categorization in E-Learning Environment Using Data Mining Technique

Nowadays, education cannot be imagined without digital technologies. It broadens the horizons of teaching learning processes. Several universities are offering online courses. For evaluation purpose, e-examination systems are being widely adopted in academic environments. Multiple-choice tests are extremely popular. Moving away from traditional examinations to e-examination, Moodle as Learning Management Systems (LMS) is being used. Moodle logs every click that students make for attempting and navigational purposes in e-examination. Data mining has been applied in various domains including retail sales, bioinformatics. In recent years, there has been increasing interest in the use of data mining in e-learning environment. It has been applied to discover, extract, and evaluate parameters related to student’s learning performance. The combination of data mining and e-learning is still in its babyhood. Log data generated by the students during online examination can be used to discover knowledge with the help of data mining techniques. In web based applications, number of right and wrong answers of the test result is not sufficient to assess and evaluate the student’s performance. So, assessment techniques must be intelligent enough. If student cannot answer the question asked by the instructor then some easier question can be asked. Otherwise, more difficult question can be post on similar topic. To do so, it is necessary to identify difficulty level of the questions. Proposed work concentrate on the same issue. Data mining techniques in specific clustering is used in this work. This method decide difficulty levels of the question and categories them as tough, easy or moderate and later this will be served to the desire students based on their performance. Proposed experiment categories the question set and also group the students based on their performance in examination. This will help the instructor to guide the students more specifically. In short mined knowledge helps to support, guide, facilitate and enhance learning as a whole.

A Quasi-Systematic Review on Effectiveness of Social and Cultural Sustainability Practices in Built Environment

With the advancement of knowledge about the utility and impact of sustainability, its feasibility has been explored into different walks of life. Scientists, however; have established their knowledge in four areas viz environmental, economic, social and cultural, popularly termed as four pillars of sustainability. Aspects of environmental and economic sustainability have been rigorously researched and practiced and huge volume of strong evidence of effectiveness has been founded for these two sub-areas. For the social and cultural aspects of sustainability, dependable evidence of effectiveness is still to be instituted as the researchers and practitioners are developing and experimenting methods across the globe. Therefore, the present research aimed to identify globally used practices of social and cultural sustainability and through evidence synthesis assess their outcomes to determine the effectiveness of those practices. A PICO format steered the methodology which included all populations, popular sustainability practices including walkability/cycle tracks, social/recreational spaces, privacy, health & human services and barrier free built environment, comparators included ‘Before’ and ‘After’, ‘With’ and ‘Without’, ‘More’ and ‘Less’ and outcomes included Social well-being, cultural coexistence, quality of life, ethics and morality, social capital, sense of place, education, health, recreation and leisure, and holistic development. Search of literature included major electronic databases, search websites, organizational resources, directory of open access journals and subscribed journals. Grey literature, however, was not included. Inclusion criteria filtered studies on the basis of research designs such as total randomization, quasirandomization, cluster randomization, observational or single studies and certain types of analysis. Studies with combined outcomes were considered but studies focusing only on environmental and/or economic outcomes were rejected. Data extraction, critical appraisal and evidence synthesis was carried out using customized tabulation, reference manager and CASP tool. Partial meta-analysis was carried out and calculation of pooled effects and forest plotting were done. As many as 13 studies finally included for final synthesis explained the impact of targeted practices on health, behavioural and social dimensions. Objectivity in the measurement of health outcomes facilitated quantitative synthesis of studies which highlighted the impact of sustainability methods on physical activity, Body Mass Index, perinatal outcomes and child health. Studies synthesized qualitatively (and also quantitatively) showed outcomes such as routines, family relations, citizenship, trust in relationships, social inclusion, neighbourhood social capital, wellbeing, habitability and family’s social processes. The synthesized evidence indicates slight effectiveness and efficacy of social and cultural sustainability on the targeted outcomes. Further synthesis revealed that such results of this study are due weak research designs and disintegrated implementations. If architects and other practitioners deliver their interventions in collaboration with research bodies and policy makers, a stronger evidence-base in this area could be generated.

Weigh-in-Motion Data Analysis Software for Developing Traffic Data for Mechanistic Empirical Pavement Design

Currently, there are few user friendly Weigh-in- Motion (WIM) data analysis softwares available which can produce traffic input data for the recently developed AASHTOWare pavement Mechanistic-Empirical (ME) design software. However, these softwares have only rudimentary Quality Control (QC) processes. Therefore, they cannot properly deal with erroneous WIM data. As the pavement performance is highly sensible to the quality of WIM data, it is highly recommended to use more refined QC process on raw WIM data to get a good result. This study develops a userfriendly software, which can produce traffic input for the ME design software. This software takes the raw data (Class and Weight data) collected from the WIM station and processes it with a sophisticated QC procedure. Traffic data such as traffic volume, traffic distribution, axle load spectra, etc. can be obtained from this software; which can directly be used in the ME design software.

Computational Assistance of the Research, Using Dynamic Vector Logistics of Processes for Critical Infrastructure Subjects Continuity

This paper deals with using of prevailing operation system MS Office (SmartArt...) for mathematical models, using DYVELOP (Dynamic Vector Logistics of Processes) method. It serves for crisis situations investigation and modelling within the organizations of critical infrastructure. In first part of paper, it will be introduced entities, operators, and actors of DYVELOP method. It uses just three operators of Boolean algebra and four types of the entities: the Environments, the Process Systems, the Cases, and the Controlling. The Process Systems (PrS) have five “brothers”: Management PrS, Transformation PrS, Logistic PrS, Event PrS and Operation PrS. The Cases have three “sisters”: Process Cell Case, Use Case, and Activity Case. They all need for the controlling of their functions special Ctrl actors, except ENV – it can do without Ctrl. Model´s maps are named the Blazons and they are able mathematically - graphically express the relationships among entities, actors and processes. In second part of this paper, the rich blazons of DYVELOP method will be used for the discovering and modelling of the cycling cases and their phases. The blazons need live PowerPoint presentation for better comprehension of this paper mission. The crisis management of energetic crisis infrastructure organization is obliged to use the cycles for successful coping of crisis situations. Several times cycling of these cases is necessary condition for the encompassment for both emergency events and the mitigation of organization´s damages. Uninterrupted and continuous cycling process brings for crisis management fruitfulness and it is good indicator and controlling actor of organizational continuity and its sustainable development advanced possibilities. The research reliable rules are derived for the safety and reliable continuity of energetic critical infrastructure organization in the crisis situation.

Risk Assessment of Lead in Egyptian Vegetables and Fruits from Different Environments

Lead being a toxic heavy metal that mankind is exposed to the highest levels of this metal. There are different sources of environmental pollution with lead as lead alkyl additives in petrol and manufacturing processes. The contaminated atmosphere in urban and industrial areas by lead in Egypt may lead to the contamination of foods beside the other different sources. The present investigation studied the risk assessment of lead in some Egyptian edible vegetables and fruits collected from different environments in Greater Cairo Governorate, i.e. industrial, heavy traffic and rural areas. A total of 325 leafy and fruity vegetables and fruits samples belonging to 11, 6 and 4 different species, respectively were randomly collected from markets of the three main models. Data indicated the variation of lead levels in different three areas. The highest levels of lead were detected in the samples collected from industrial and traffic areas. However, the lowest levels were found in the rural areas. It could be concluded that determination of lead levels in foods from different localities and environments at regularly is very important.

Defect Management Life Cycle Process for Software Quality Improvement

Software quality issues require special attention especially in view of the demands of quality software product to meet customer satisfaction. Software development projects in most organisations need proper defect management process in order to produce high quality software product and reduce the number of defects. The research question of this study is how to produce high quality software and reducing the number of defects. Therefore, the objective of this paper is to provide a framework for managing software defects by following defined life cycle processes. The methodology starts by reviewing defects, defect models, best practices, and standards. A framework for defect management life cycle is proposed. The major contribution of this study is to define a defect management roadmap in software development. The adoption of an effective defect management process helps to achieve the ultimate goal of producing high quality software products and contributes towards continuous software process improvement.

Cost Valuation Method for Development Concurrent Phase Appropriate Requirement Valuation Using the Example of Load Carrier Development in the Lithium-Ion-Battery Production

In the past years electric mobility became part of a public discussion. The trend to fully electrified vehicles instead of vehicles fueled with fossil energy has notably gained momentum. Today nearly every big car manufacturer produces and sells fully electrified vehicles, but electrified vehicles are still not as competitive as conventional powered vehicles. As the traction battery states the largest cost driver, lowering its price is a crucial objective. In addition to improvements in product and production processes a nonnegligible, but widely underestimated cost driver of production can be found in logistics, since the production technology is not continuous yet and neither are the logistics systems. This paper presents an approach to evaluate cost factors on different designs of load carrier systems. Due to numerous interdependencies, the combination of costs factors for a particular scenario is not transparent. This is effecting actions for cost reduction negatively, but still cost reduction is one of the major goals for simultaneous engineering processes. Therefore a concurrent and phase appropriate cost valuation method is necessary to serve cost transparency. In this paper the four phases of this cost valuation method are defined and explained, which based upon a new approach integrating the logistics development process in to the integrated product and process development.

Usage of Military Continuity Management System for Supporting of Emergency Management

Ensuring of continuity of business is basic strategy of every company. Continuity of organization activities includes comprehensive procedures that help in solving unexpected situations of natural and anthropogenic character (for example flood, blaze, economic situations). Planning of continuity operations is a process that helps identify critical processes and implement plans for the security and recovery of key processes. The aim of this article is to demonstrate application of system approach to managing business continuity called business continuity management systems in military issues. This article describes the life cycle of business continuity management which is based on the established cycle PDCA (Plan- Do-Check-Act). After this is carried out by activities which are making by University of Defence during activation of forces and means of the integrated rescue system in case of emergencies - accidents at a nuclear power plant in Czech Republic. Activities of various stages of deployment earmarked forces and resources are managed and evaluated by using MCMS application (Military Continuity Management System).

Advanced Manufacturing Technology Adoption in Manufacturing Companies in Kenya

Over the past few decades, manufacturing has evolved from a more labor-intensive set of mechanical processes to a sophisticated set of information based technology processes. With the existence of various advanced manufacturing technologies (AMTs), more and more functions or jobs are performed by these machines instead of human labour. This study was undertaken in order to research the extent of AMTs adoption in manufacturing companies in Kenya. In order to investigate a survey was conducted via questionnaires that were sent to 183 selected AMT manufacturing companies in Kenya. 92 companies responded positively. All the surveyed companies were found to have a measure of investment in at least two of the 14 types of AMTs investigated. In general the company surveyed showed that the level of AMT adoption in Kenya is very low with investments levels at a mean of 2.057 and integration levels at a mean of 1.639 in a scale of 1-5.

Top-Down Influences to Multistable Perception: Evidence from Temporal Dynamics

We have studied the temporal characteristics of bistable perception of the stimuli of two types: one involves alterations in a perceived depth and another one has an ambiguous content. We used the Necker lattice and lines of shadowed circles ambiguously perceived either as spheres or holes as stimuli of the first type. The Winson figure (the Eskimo/Indian picture) was a stimulus of the second type. We have analyzed how often the reversals occurred (reversal rate) and for how long each of the two interpretations, or percepts, was observed during one presentation (stability durations). For all three ambiguous images the reversal rate and the stability durations had similar values, which provide another evidence for a significant role of top-down processes in multistable perception.

Using Agility in Building Business Process Management Solutions

In turbulent modern economy, the companies need to properly manage their business processes. Well-defined and stable business processes ensure security of crucial data and applications, and provide a quality product or service to the end customer. On the other side, constant changes on the market, new regulatory provisions, and emerging new technologies require the need of issuing prompt and effective changes of business process. In this article, we explore the use of agile principles in working with business process management (BPM) solutions. We deal with difficulties in BPM development cycle, review the benefits of using agility, and choose the basic agile principles that ensure the success of a BPM project.