Growth Performance and Blood Characteristics of Broilers Chicken Fed on Diet Containing Brewer Spent Grain at Finisher Phase

This study was conducted to investigate the effects of brewer spent grain (BSG) on growth performance and serum biochemistry characteristics of blood of broilers chickens. Three hundred and fifteen (4 weeks old) Oba – Marshall Broilers were used for the experiment. Five experimental diets were formulated with diet 1 (T1) containing 100% soya bean meal as the control, Diet 2, 3, 4 and 5 had BSG as replacement for soya bean meal at 0%, 36%, 57%, 76% and 100% respectively. The birds were allocated into each dietary group in a completely randomized design with 63 chicks in 3 replicates of 21 chicks each. The birds were offered these diets ad libitum from four weeks old to nine weeks old (35 days). Feed intake, body weight, weight gain, and feed conversion ratio (FCR) were assessed. Blood samples were also collected to examine the effect of BSG waste on hematology and serum biochemistry of broilers. Result indicated that BSG did not significantly (P>0.05) affect feed intake and weight gain. However, FCR and final weight of finishing broilers differs significantly (P

Opening up Government Datasets for Big Data Analysis to Support Policy Decisions

Policy makers are increasingly looking to make evidence-based decisions. Evidence-based decisions have historically used rigorous methodologies of empirical studies by research institutes, as well as less reliable immediate survey/polls often with limited sample sizes. As we move into the era of Big Data analytics, policy makers are looking to different methodologies to deliver reliable empirics in real-time. The question is not why did these people do this for the last 10 years, but why are these people doing this now, and if the this is undesirable, and how can we have an impact to promote change immediately. Big data analytics rely heavily on government data that has been released in to the public domain. The open data movement promises greater productivity and more efficient delivery of services; however, Australian government agencies remain reluctant to release their data to the general public. This paper considers the barriers to releasing government data as open data, and how these barriers might be overcome.

A Large Ion Collider Experiment (ALICE) Diffractive Detector Control System for RUN-II at the Large Hadron Collider

The selection of diffractive events in the ALICE experiment during the first data taking period (RUN-I) of the Large Hadron Collider (LHC) was limited by the range over which rapidity gaps occur. It would be possible to achieve better measurements by expanding the range in which the production of particles can be detected. For this purpose, the ALICE Diffractive (AD0) detector has been installed and commissioned for the second phase (RUN-II). Any new detector should be able to take the data synchronously with all other detectors and be operated through the ALICE central systems. One of the key elements that must be developed for the AD0 detector is the Detector Control System (DCS). The DCS must be designed to operate safely and correctly this detector. Furthermore, the DCS must also provide optimum operating conditions for the acquisition and storage of physics data and ensure these are of the highest quality. The operation of AD0 implies the configuration of about 200 parameters, from electronics settings and power supply levels to the archiving of operating conditions data and the generation of safety alerts. It also includes the automation of procedures to get the AD0 detector ready for taking data in the appropriate conditions for the different run types in ALICE. The performance of AD0 detector depends on a certain number of parameters such as the nominal voltages for each photomultiplier tube (PMT), their threshold levels to accept or reject the incoming pulses, the definition of triggers, etc. All these parameters define the efficiency of AD0 and they have to be monitored and controlled through AD0 DCS. Finally, AD0 DCS provides the operator with multiple interfaces to execute these tasks. They are realized as operating panels and scripts running in the background. These features are implemented on a SCADA software platform as a distributed control system which integrates to the global control system of the ALICE experiment.

Formulation and Evaluation of Niosomes Containing an Antihypertensive Drug

Niosomes were formulated with an aim of enhancing the oral bioavailability of losartan potassium and formulated in different molar ratios of surfactant, cholesterol and dicetyl phosphate. The formulated niosomes were found in range of 54.98 µm to 107.85 µm in size. Formulations with 1:1 ratio of surfactant and cholesterol have shown maximum entrapment efficiencies. Niosomes with sorbitan monostearate showed maximum drug release and zero order release kinetics, at the end of 24 hours. The in vivo study has shown the significant enhancement in oral bioavailability of losartan potassium in rats, after a dose of 10 mg/kg. The average relative bioavailability in relation with pure drug solution was found 2.56, indicates more than two fold increase in oral bioavailability. A significant increment in MRT reflects the release retarding ability of the vesicles. In conclusion, niosomes could be a promising delivery of losartan potassium with improved oral bioavailability and prolonged release profiles.

Conduction Accompanied With Transient Radiative Heat Transfer Using Finite Volume Method

The objective of this research work is to investigate for one dimensional transient radiative transfer equations with conduction using finite volume method. Within the infrastructure of finite-volume, we obtain the conservative discretization of the terms in order to preserve the overall conservative property of finitevolume schemes. Coupling of conductive and radiative equation resulting in fluxes is governed by the magnitude of emissivity, extinction coefficient, and temperature of the medium as well as geometry of the problem. The problem under consideration has been solved, for a slab dominating radiation coupled with transient conduction based on finite volume method. The boundary conditions are also chosen so as to give a good model of the discretized form of radiation transfer equation. The important feature of the present method is flexibility in specifying the control angles in the FVM, while keeping the simplicity in the solution procedure. Effects of various model parameters are examined on the distributions of temperature, radiative and conductive heat fluxes and incident radiation energy etc. The finite volume method is considered to effectively evaluate the propagation of radiation intensity through a participating medium.

A Reference Framework Integrating Lean and Green Principles within Supply Chain Management

In the last decades, an increasing set of companies adopted lean philosophy to improve their productivity and efficiency promoting the so-called continuous improvement concept, reducing waste of time and cutting off no-value added activities. In parallel, increasing attention rises toward green practice and management through the spread of the green supply chain pattern, to minimise landfilled waste, drained wastewater and pollutant emissions. Starting from a review on contributions deepening lean and green principles applied to supply chain management, the most relevant drivers to measure the performance of industrial processes are pointed out. Specific attention is paid on the role of cost because it is of key importance and it crosses both lean and green principles. This analysis leads to figure out an original reference framework for integrating lean and green principles in designing and managing supply chains. The proposed framework supports the application, to the whole value chain or to parts of it, e.g. distribution network, assembly system, job-shop, storage system etc., of the lean-green integrated perspective. Evidences show that the combination of the lean and green practices lead to great results, higher than the sum of the performances from their separate application. Lean thinking has beneficial effects on green practices and, at the same time, methods allowing environmental savings generate positive effects on time reduction and process quality increase.

Status Report of the GERDA Phase II Startup

The GERmanium Detector Array (GERDA) experiment, located at the Laboratori Nazionali del Gran Sasso (LNGS) of INFN, searches for 0νββ of 76Ge. Germanium diodes enriched to ∼ 86 % in the double beta emitter 76Ge(enrGe) are exposed being both source and detectors of 0νββ decay. Neutrinoless double beta decay is considered a powerful probe to address still open issues in the neutrino sector of the (beyond) Standard Model of particle Physics. Since 2013, just after the completion of the first part of its experimental program (Phase I), the GERDA setup has been upgraded to perform its next step in the 0νββ searches (Phase II). Phase II aims to reach a sensitivity to the 0νββ decay half-life larger than 1026 yr in about 3 years of physics data taking. This exposing a detector mass of about 35 kg of enrGe and with a background index of about 10^−3 cts/(keV·kg·yr). One of the main new implementations is the liquid argon scintillation light read-out, to veto those events that only partially deposit their energy both in Ge and in the surrounding LAr. In this paper, the GERDA Phase II expected goals, the upgrade work and few selected features from the 2015 commissioning and 2016 calibration runs will be presented. The main Phase I achievements will be also reviewed.

Statistically Significant Differences of Carbon Dioxide and Carbon Monoxide Emission in Photocopying Process

Experimental results confirmed the temporal variation of carbon dioxide and carbon monoxide concentration during the working shift of the photocopying process in a small photocopying shop in Novi Sad, Serbia. The statistically significant differences of target gases were examined with two-way analysis of variance without replication followed by Scheffe's post hoc test. The existence of statistically significant differences was obtained for carbon monoxide emission which is pointed out with F-values (12.37 and 31.88) greater than Fcrit (6.94) in contrary to carbon dioxide emission (F-values of 1.23 and 3.12 were less than Fcrit).  Scheffe's post hoc test indicated that sampling point A (near the photocopier machine) and second time interval contribute the most on carbon monoxide emission.

Quantitative Assessment of Different Formulations of Antimalarials in Sentinel Sites of India

Substandard and counterfeit antimalarials is a major problem in malaria endemic areas. The availability of counterfeit/ substandard medicines is not only decreasing the efficacy in patients, but it is also one of the contributing factors for developing antimalarial drug resistance. Owing to this, a pilot study was conducted to survey quality of drugs collected from different malaria endemic areas of India. Artesunate+Sulphadoxine-Pyrimethamine (AS+SP), Artemether-Lumefantrine (AL), Chloroquine (CQ) tablets were randomly picked from public health facilities in selected states of India. The quality of antimalarial drugs from these areas was assessed by using Global Pharma Health Fund Minilab test kit. This includes physical/visual inspection and disintegration test. Thin-layer chromatography (TLC) was carried out for semi-quantitative assessment of active pharmaceutical ingredients. A total of 45 brands, out of which 21 were for CQ, 14 for AL and 10 for AS+SP were tested from Uttar Pradesh (U.P.), Mizoram, Meghalaya and Gujrat states. One out of 45 samples showed variable disintegration and retension factor. The variable disintegration and retention factor which would have been due to substandard quality or other factors including storage. However, HPLC analysis confirms standard active pharmaceutical ingredient, but may be due to humid temperature and moisture in storage may account for the observed result.

Designing Creative Events with Deconstructivism Approach

Deconstruction is an approach that is entirely incompatible with the traditional prevalent architecture. Considering the fact that this approach attempts to put architecture in sharp contrast with its opposite events and transpires with attending to the neglected and missing aspects of architecture and deconstructing its stable structures. It also recklessly proceeds beyond the existing frameworks and intends to create a different and more efficient prospect for space. The aim of deconstruction architecture is to satisfy both the prospective and retrospective visions as well as takes into account all tastes of the present in order to transcend time. Likewise, it ventures to fragment the facts and symbols of the past and extract new concepts from within their heart, which coincide with today’s circumstances. Since this approach is an attempt to surpass the limits of the prevalent architecture, it can be employed to design places in which creative events occur and imagination and ambition flourish. Thought-provoking artistic events can grow and mature in such places and be represented in the best way possible to all people. The concept of event proposed in the plan grows out of the interaction between space and creation. In addition to triggering surprise and high impressions, it is also considered as a bold journey into the suspended realms of the traditional conflicts in architecture such as architecture-landscape, interior-exterior, center-margin, product-process, and stability-instability. In this project, at first, through interpretive-historical research method and examining the inputs and data collection, recognition and organizing takes place. After evaluating the obtained data using deductive reasoning, the data is eventually interpreted. Given the fact that the research topic is in its infancy and there is not a similar case in Iran with limited number of corresponding instances across the world, the selected topic helps to shed lights on the unrevealed and neglected parts in architecture. Similarly, criticizing, investigating and comparing specific and highly prized cases in other countries with the project under study can serve as an introduction into this architecture style.

High-Rises and Urban Design: The Reasons for Unsuccessful Placemaking with Residential High-Rises in England

High-rises and placemaking is an understudied combination which receives more and more interest with the proliferation of this typology in many British cities. The reason for studying three major cities in England: London, Birmingham and Manchester, is to learn from the latest advances in urban design in well-developed and prominent urban environment. The analysis of several high-rise sites reveals the weaknesses in urban design of contemporary British cities and presents an opportunity to study from the implemented examples. Therefore, the purpose of this research is to analyze design approaches towards creating a sustainable and varied urban environment when high-rises are involved. The research questions raised by the study are: what is the quality of high-rises and their surroundings; what facilities and features are deployed in the research area; what is the role of the high-rise buildings in the placemaking process; what urban design principles are applicable in this context. The methodology utilizes observation of the researched area by structured questions, developed by the author to evaluate the outdoor qualities of the high-rise surroundings. In this context, the paper argues that the quality of the public realm around the high-rises is quite low, missing basic but vital elements such as plazas, public art, and seating, along with landscaping and pocket parks. There is lack of coherence, the rhythm of the streets is often disrupted, and even though the high-rises are very aesthetically appealing, they fail to create a sense of place on their own. The implications of the study are that future planning can take into consideration the critique in this article and provide more opportunities for urban design interventions around high-rise buildings in the British cities.

Threshold Based Region Incrementing Secret Sharing Scheme for Color Images

In this era of online communication, which transacts data in 0s and 1s, confidentiality is a priced commodity. Ensuring safe transmission of encrypted data and their uncorrupted recovery is a matter of prime concern. Among the several techniques for secure sharing of images, this paper proposes a k out of n region incrementing image sharing scheme for color images. The highlight of this scheme is the use of simple Boolean and arithmetic operations for generating shares and the Lagrange interpolation polynomial for authenticating shares. Additionally, this scheme addresses problems faced by existing algorithms such as color reversal and pixel expansion. This paper regenerates the original secret image whereas the existing systems regenerates only the half toned secret image.

Examining the Change of Power Transmission Line in Urban Regeneration with Geographical Information System

In this study, spatial differences of Power Transmission Line (PTL) and effects of the situation before and after the urban regeneration are studied by using Geographical Information System (GIS). In addition, a questionable and analyzable structure is acquired by developed system. In the study area many parcels on the PTL were analyzed. The amount of the parcels, which are affected by the negativity of PTL is clearly seen with the aid of generated maps. Some kind of changes are exhibited in the system, which are created by GIS, for instance before urban regeneration PTL was very close to people’s private properties and huge parts of PTL were among the buildings, however; after urban regeneration electricity lines were changed their locations to the underground. According to the results, GIS can be used as a device in planning and managing of PTL in urban regeneration projects and can be used for analyses. By the help of GIS technology, necessary investigations should be carried out in urban regeneration applications for creating sustainable cities.

Applying Participatory Design for the Reuse of Deserted Community Spaces

The concept of community building started in 1994 in Taiwan. After years of development, it fostered the notion of active local resident participation in community issues as co-operators, instead of minions. Participatory design gives participants more control in the decision-making process, helps to reduce the friction caused by arguments and assists in bringing different parties to consensus. This results in an increase in the efficiency of projects run in the community. Therefore, the participation of local residents is key to the success of community building. This study applied participatory design to develop plans for the reuse of deserted spaces in the community from the first stage of brainstorming for design ideas, making creative models to be employed later, through to the final stage of construction. After conducting a series of participatory designed activities, it aimed to integrate the different opinions of residents, develop a sense of belonging and reach a consensus. Besides this, it also aimed at building the residents’ awareness of their responsibilities for the environment and related issues of sustainable development. By reviewing relevant literature and understanding the history of related studies, the study formulated a theory. It took the “2012-2014 Changhua County Community Planner Counseling Program” as a case study to investigate the implementation process of participatory design. Research data are collected by document analysis, participants’ observation and in-depth interviews. After examining the three elements of “Design Participation”, “Construction Participation”, and” Follow–up Maintenance Participation” in the case, the study emerged with a promising conclusion: Maintenance works were carried out better compared to common public works. Besides this, maintenance costs were lower. Moreover, the works that residents were involved in were more creative. Most importantly, the community characteristics could be easy be recognized.

Coding Considerations for Standalone Molecular Dynamics Simulations of Atomistic Structures

The laws of Newtonian mechanics allow ab-initio molecular dynamics to model and simulate particle trajectories in material science by defining a differentiable potential function. This paper discusses some considerations for the coding of ab-initio programs for simulation on a standalone computer and illustrates the approach by C language codes in the context of embedded metallic atoms in the face-centred cubic structure. The algorithms use velocity-time integration to determine particle parameter evolution for up to several thousands of particles in a thermodynamical ensemble. Such functions are reusable and can be placed in a redistributable header library file. While there are both commercial and free packages available, their heuristic nature prevents dissection. In addition, developing own codes has the obvious advantage of teaching techniques applicable to new problems.

De-Securitizing Identity: Narrative (In)Consistency in Periods of Transition

When examining conflicts around the world, it is evident that the majority of intractable conflicts are steeped in identity. Identity seems to be not only a causal variable for conflict, but also a catalytic parameter for the process of reconciliation that follows ceasefire. This paper focuses on the process of identity securitization that occurs between rival groups of heterogeneous collective identities – ethnic, national or religious – as well as on the relationship between identity securitization and the ability of the groups involved to reconcile. Are securitized identities obstacles to the process of reconciliation, able to hinder any prospects of peace? If the level to which an identity is securitized is catalytic to a conflict’s discourse and settlement, then which factors act as indicators of identity de-securitization? The level of an in-group’s identity securitization can be estimated through a number of indicators, one of which is narrative. The stories, views and stances each in-group adopts in relation to its history of conflict and relation with their rival out-group can clarify whether that specific in-group feels victimized and threatened or safe and ready to reconcile. Accordingly, this study discusses identity securitization through narrative in relation to intractable conflicts. Are there conflicts around the world that, despite having been identified as intractable, stagnated or insoluble, show signs of identity de-securitization through narrative? This inquiry uses the case of the Cyprus conflict and its partitioned societies to present official narratives from the two communities and assess whether these narratives have transformed, indicating a less securitized in-group identity for the Greek and Turkish Cypriots. Specifically, the study compares the official historical overviews presented by each community’s Ministry of Foreign Affairs website and discusses the extent to which the two official narratives present a securitized collective identity. In addition, the study will observe whether official stances by the two communities – as adopted by community leaders – have transformed to depict less securitization over time. Additionally, the leaders’ reflection of popular opinion is evaluated through recent opinion polls from each community. Cyprus is currently experiencing renewed optimism for reunification, with the leaders of its two communities engaging in rigorous negotiations, and with rumors calling for a potential referendum for reunification to be taking place even as early as within 2016. Although leaders’ have shown a shift in their rhetoric and have moved away from narratives of victimization, this is not the case for the official narratives used by their respective ministries of foreign affairs. The study’s findings explore whether this narrative inconsistency proves that Cyprus is transitioning towards reunification, or whether the leaders are risking sending a securitized population to the polls to reject a potential reunification. More broadly, this study suggests that in the event that intractable conflicts might be moving towards viable peace, in-group narratives--official narratives in particular--can act as indicators of the extent to which rival entities have managed to reconcile.

Conflict of the Thai-Malaysian Gas Pipeline Project

This research was aimed to investigate (1) the relationship among local social movements, non-governmental Organization activities and state measures deployment; and (2) the effects of local social movements, non-governmental Organization activities, and state measures deployment on conflict of local people towards the Thai-Malaysian gas pipeline project. These people included 1,000 residents of the four districts in Songkhla province. The methods of data analysis consist of multiple regression analysis. The results of the analysis showed that: (1) local social movements depended on information, and mass communication; deployment of state measures depended on compromise, coordination, and mass communication; and (2) the conflict of local people depended on mobilization, negotiation, and campaigning for participation of people in the project. Thus, it is recommended that to successfully implement any government policy, consideration must be paid to the conflict of local people, mobilization, negotiation, and campaigning for people’s participation in the project.

X-Corner Detection for Camera Calibration Using Saddle Points

This paper discusses a corner detection algorithm for camera calibration. Calibration is a necessary step in many computer vision and image processing applications. Robust corner detection for an image of a checkerboard is required to determine intrinsic and extrinsic parameters. In this paper, an algorithm for fully automatic and robust X-corner detection is presented. Checkerboard corner points are automatically found in each image without user interaction or any prior information regarding the number of rows or columns. The approach represents each X-corner with a quadratic fitting function. Using the fact that the X-corners are saddle points, the coefficients in the fitting function are used to identify each corner location. The automation of this process greatly simplifies calibration. Our method is robust against noise and different camera orientations. Experimental analysis shows the accuracy of our method using actual images acquired at different camera locations and orientations.

Image Features Comparison-Based Position Estimation Method Using a Camera Sensor

In this paper, propose method that can user’s position that based on database is built from single camera. Previous positioning calculate distance by arrival-time of signal like GPS (Global Positioning System), RF(Radio Frequency). However, these previous method have weakness because these have large error range according to signal interference. Method for solution estimate position by camera sensor. But, signal camera is difficult to obtain relative position data and stereo camera is difficult to provide real-time position data because of a lot of image data, too. First of all, in this research we build image database at space that able to provide positioning service with single camera. Next, we judge similarity through image matching of database image and transmission image from user. Finally, we decide position of user through position of most similar database image. For verification of propose method, we experiment at real-environment like indoor and outdoor. Propose method is wide positioning range and this method can verify not only position of user but also direction.

Graphene/h-BN Heterostructure Interconnects

The material behavior of graphene, a single layer of carbon lattice, is extremely sensitive to its dielectric environment. We demonstrate improvement in electronic performance of graphene nanowire interconnects with full encapsulation by lattice-matching, chemically inert, 2D layered insulator hexagonal boron nitride (h- BN). A novel layer-based transfer technique is developed to construct the h-BN/MLG/h-BN heterostructures. The encapsulated graphene wires are characterized and compared with that on SiO2 or h-BN substrate without passivating h-BN layer. Significant improvements in maximum current-carrying density, breakdown threshold, and power density in encapsulated graphene wires are observed. These critical improvements are achieved without compromising the carrier transport characteristics in graphene. Furthermore, graphene wires exhibit electrical behavior less insensitive to ambient conditions, as compared with the non-passivated ones. Overall, h-BN/graphene/h- BN heterostructure presents a robust material platform towards the implementation of high-speed carbon-based interconnects.