Abstract: The medical model of disability, though beneficial for the medical professional, is often exclusionary, restrictive and dehumanizing when applied to the lived experience of disability. As a result, a critique of this model was constructed called the social model of disability. Much of the language used to articulate the purpose behind the social model of disability can be summed up within the word inclusion. However, this essay asserts that inclusiveness is an incomplete aspiration. The social model, as it currently stands, does not aid in creating a society where those with impairments actually belong. Rather, the social model aids in lessening the visibility, or negative consequence of, difference. Therefore, the social model does not invite society to welcome those with physical and intellectual impairments. It simply aids society in ignoring the existence of impairment by removing explicit forms of exclusion. Rather than simple inclusion, then, this essay uses John Swinton’s concept of friendship and Jean Vanier’s understanding of belonging to better articulate the intended outcome of the social model—a society where everyone can belong.
Abstract: Magnetic Resonance Imaging Contrast Agents
(MRI-CM) are significant in the clinical and biological imaging as
they have the ability to alter the normal tissue contrast, thereby
affecting the signal intensity to enhance the visibility and detectability
of images. Superparamagnetic Iron Oxide (SPIO) nanoparticles,
coated with dextran or carboxydextran are currently available for
clinical MR imaging of the liver. Most SPIO contrast agents are
T2 shortening agents and Resovist (Ferucarbotran) is one of a
clinically tested, organ-specific, SPIO agent which has a low
molecular carboxydextran coating. The enhancement effect of
Resovist depends on its relaxivity which in turn depends on factors
like magnetic field strength, concentrations, nanoparticle properties,
pH and temperature. Therefore, this study was conducted to
investigate the impact of field strength and different contrast
concentrations on enhancement effects of Resovist. The study
explored the MRI signal intensity of Resovist in the physiological
range of plasma from T2-weighted spin echo sequence at three
magnetic field strengths: 0.47 T (r1=15, r2=101), 1.5 T (r1=7.4,
r2=95), and 3 T (r1=3.3, r2=160) and the range of contrast
concentrations by a mathematical simulation. Relaxivities of r1 and r2
(L mmol-1 Sec-1) were obtained from a previous study and the selected
concentrations were 0.05, 0.06, 0.07, 0.08, 0.09, 0.1, 0.2, 0.3, 0.4, 0.5,
0.6, 0.7, 0.8, 0.9, 1.0, 2.0, and 3.0 mmol/L. T2-weighted images were
simulated using TR/TE ratio as 2000 ms /100 ms. According to the
reference literature, with increasing magnetic field strengths, the
r1 relaxivity tends to decrease while the r2 did not show any
systematic relationship with the selected field strengths. In parallel,
this study results revealed that the signal intensity of Resovist at lower
concentrations tends to increase than the higher concentrations. The
highest reported signal intensity was observed in the low field strength
of 0.47 T. The maximum signal intensities for 0.47 T, 1.5 T and 3 T
were found at the concentration levels of 0.05, 0.06 and 0.05 mmol/L,
respectively. Furthermore, it was revealed that, the concentrations
higher than the above, the signal intensity was decreased
exponentially. An inverse relationship can be found between the field
strength and T2 relaxation time, whereas, the field strength was
increased, T2 relaxation time was decreased accordingly. However,
resulted T2 relaxation time was not significantly different between
0.47 T and 1.5 T in this study. Moreover, a linear correlation of
transverse relaxation rates (1/T2, s–1) with the concentrations of
Resovist can be observed. According to these results, it can conclude
that the concentration of SPIO nanoparticle contrast agents and the
field strengths of MRI are two important parameters which can affect the signal intensity of T2-weighted SE sequence. Therefore, when MR
imaging those two parameters should be considered prudently.
Abstract: Ant algorithms are well-known metaheuristics which
have been widely used since two decades. In most of the literature,
an ant is a constructive heuristic able to build a solution from scratch.
However, other types of ant algorithms have recently emerged: the
discussion is thus not limited by the common framework of the
constructive ant algorithms. Generally, at each generation of an ant
algorithm, each ant builds a solution step by step by adding an
element to it. Each choice is based on the greedy force (also called the
visibility, the short term profit or the heuristic information) and the
trail system (central memory which collects historical information of
the search process). Usually, all the ants of the population have the
same characteristics and behaviors. In contrast in this paper, a new
type of ant metaheuristic is proposed, namely SMART (for Solution
Methods with Ants Running by Types). It relies on the use of different
population of ants, where each population has its own personality.
Abstract: Visibility problems are central to many computational geometry applications. One of the typical visibility problems is computing the view from a given point. In this paper, a linear time procedure is proposed to compute the visibility subsets from a corner of a rectangular prism in an orthogonal polyhedron. The proposed algorithm could be useful to solve classic 3D problems.
Abstract: Environmental concerns about the scarcity of marine
resources are critical driving forces for firms aiming to prepare their
supply chains for sustainability. Building on previous work, this
paper highlights the implementation of good practices geared towards
sustainable operations in the seafood department, which were
pursued in an exploratory retailer case. Outcomes of the adopted
environmentally and socially acceptable fish retailing strategies,
ranged from traceability, to self-certification and eco-labelling. The
consequences for business were, as follows: stronger collaboration
and trust across the chain of custody, improvement of sponsors’
image and of consumers’ loyalty and, progress in the Greenpeace
retailers’ evaluation ranking.
Abstract: The social logic of 'Sequina' slum area in Alexandria details the integral measure of space syntax at the room-level of twenty-building samples. The essence of spatial structure integrates the central 'visitor' domain with the 'living' frontage of the 'children' zone against the segregated privacy of the opposite 'parent' depth. Meanwhile, the multifunctioning of shallow rooms optimizes the integral 'visitor' structure through graph and visibility dimensions in contrast to the 'inhabitant' structure of graph-tails out of sight. Common theme of the layout integrity increases in compensation to the decrease of room visibility. Despite the 'pheno-type' of collective integration, the individual layouts observe 'geno-type' structure of spatial diversity per room adjoins. In this regard, the layout integrity alternates the cross-correlation of the 'kitchen & living' rooms with the 'inhabitant & visitor' domains of 'motherhood' dynamic structure. Moreover, the added 'grandparent' restructures the integral measure to become the deepest space, but opens to the 'living' of 'household' integrity. Some isomorphic layouts change the integral structure just through the 'balcony' extension of access, visual or ignored 'ringiness' of space syntax. However, the most integrated or segregated layouts invert the 'geno-type' into a shallow 'inhabitant' centrality versus the remote 'visitor' structure. Overview of the multivariate social logic of spatial integrity could never clarify without the micro-data analysis.
Abstract: The UK is leading in online retail and mobile
adoption. However, there is a dearth of information relating to mobile
apparel retail, and developing an understanding about consumer
browsing and purchase behaviour in m-retail channel would provide
apparel marketers, mobile website and app developers with the
necessary understanding of consumers’ needs. Despite the rapid
growth of mobile retail businesses, no published study has examined
shopping behaviour on fashion mobile apps and websites. A mixed method approach helped to understand why fashion
consumers prefer websites on smartphones, when diverse mobile
apps are also available. The following research methods were
employed: survey, eye-tracking experiments, observation, and
interview with retrospective think aloud. The mobile gaze tracking
device by SensoMotoric Instruments was used to understand
frustrations in navigation and other issues facing consumers in
mobile channel. This method helped to validate and compliment
other traditional user-testing approaches in order to optimize user
experience and enhance the development of mobile retail channel.
The study involved eight participants - females aged 18 to 35 years
old, who are existing mobile shoppers. The participants used the
Topshop mobile app and website on a smart phone to complete a task
according to a specified scenario leading to a purchase. The
comparative study was based on: duration and time spent at different
stages of the shopping journey, number of steps involved and product
pages visited, search approaches used, layout and visual clues, as
well as consumer perceptions and expectations. The results from the data analysis show significant differences in
consumer behaviour when using a mobile app or website on a smart
phone. Moreover, two types of problems were identified, namely
technical issues and human errors. Having a mobile app does not
guarantee success in satisfying mobile fashion consumers. The
differences in the layout and visual clues seem to influence the
overall shopping experience on a smart phone. The layout of search
results on the website was different from the mobile app. Therefore,
participants, in most cases, behaved differently on different
platforms. The number of product pages visited on the mobile app
was triple the number visited on the website due to a limited visibility
of products in the search results. Although, the data on traffic trends
held by retailers to date, including retail sector breakdowns for visits
and views, data on device splits and duration, might seem a valuable
source of information, it cannot explain why consumers visit many
product pages, stay longer on the website or mobile app, or abandon
the basket. A comprehensive list of pros and cons was developed by
highlighting issues for website and mobile app, and recommendations
provided. The findings suggest that fashion retailers need to be aware of
actual consumers’ behaviour on the mobile channel and their expectations in order to offer a seamless shopping experience. Added
to which is the challenge of retaining existing and acquiring new
customers. There seem to be differences in the way fashion
consumers search and shop on mobile, which need to be explored in
further studies.
Abstract: Automated Teller Machines (ATMs) can be
considered among one of the most important service facilities in the
banking industry. The investment in ATMs and the impact on the
banking industry is growing steadily in every part of the world. The
banks take into consideration many factors like safety, convenience,
visibility, and cost in order to determine the optimum locations of
ATMs. Today, ATMs are not only available in bank branches but
also at retail locations. Another important factor is the cash
management in ATMs. A cash demand model for every ATM is
needed in order to have an efficient cash management system. This
forecasting model is based on historical cash demand data which is
highly related to the ATMs location. So, the location and the cash
management problem should be considered together. This paper
provides a general review on studies, efforts and development in
ATMs location and cash management problem.
Abstract: To mitigate the urban heat island effect has become a
global issue when we are faced with the challenge of climate change.
Through literature review, plant photosynthesis can reduce the carbon
dioxide and mitigate the urban heat island effect to a degree. Because
there are not enough open space and parks, green roof has become an
important policy in Taiwan.
We selected elementary school buildings in northern New Taipei
City as research subjects since elementary schools are asked with
priority to build green roof and important educational place to promote
green roof concept. Testo175-H1 recording device was used to record
the temperature and humidity differences between roof surface and
interior space below roof with and without green roof in the long-term.
We also use questionnaires to investigate the awareness of comfort
level of green roof and sensation of teachers and students of the
elementary schools.
The results indicated that the temperature of roof without greening
was higher than that with greening by about 2°C. But sometimes
during noontime, the temperature of green roof was higher than that of
non-green roof probably because of the character of the accumulation
and dissipation of heat of greening. The temperature of the interior
space below green roof was normally lower than that without green
roof by about 1°C, showing that green roof could lower the
temperature. The humidity of the green roof was higher than the one
without greening also indicated that green roof retained water better.
Teachers liked to combine green roof concept in the curriculum,
and students wished all classes can take turns to maintain the green
roof. Teachers and students whose school had integrated green roof
concept in the curriculum were more willing to participate in the
maintenance work of green roof. Teachers and students who may have
access to and touch the green roof can be more aware of the green roof
benefit. We suggest architects to increase the accessibility and
visibility of green roof, such as use it as a part of the activity space.
This idea can be a reference to the green roof curriculum design.
Abstract: The detection of moving objects from a video image
sequences is very important for object tracking, activity recognition,
and behavior understanding in video surveillance.
The most used approach for moving objects detection / tracking is
background subtraction algorithms. Many approaches have been
suggested for background subtraction. But, these are illumination
change sensitive and the solutions proposed to bypass this problem
are time consuming.
In this paper, we propose a robust yet computationally efficient
background subtraction approach and, mainly, focus on the ability to
detect moving objects on dynamic scenes, for possible applications in
complex and restricted access areas monitoring, where moving and
motionless persons must be reliably detected. It consists of three
main phases, establishing illumination changes invariance,
background/foreground modeling and morphological analysis for
noise removing.
We handle illumination changes using Contrast Limited Histogram
Equalization (CLAHE), which limits the intensity of each pixel to
user determined maximum. Thus, it mitigates the degradation due to
scene illumination changes and improves the visibility of the video
signal. Initially, the background and foreground images are extracted
from the video sequence. Then, the background and foreground
images are separately enhanced by applying CLAHE.
In order to form multi-modal backgrounds we model each channel
of a pixel as a mixture of K Gaussians (K=5) using Gaussian Mixture
Model (GMM). Finally, we post process the resulting binary
foreground mask using morphological erosion and dilation
transformations to remove possible noise.
For experimental test, we used a standard dataset to challenge the
efficiency and accuracy of the proposed method on a diverse set of
dynamic scenes.
Abstract: This paper presents a new meta-heuristic bio-inspired
optimization algorithm which is called Cuttlefish Algorithm (CFA).
The algorithm mimics the mechanism of color changing behavior of
the cuttlefish to solve numerical global optimization problems. The
colors and patterns of the cuttlefish are produced by reflected light
from three different layers of cells. The proposed algorithm considers
mainly two processes: reflection and visibility. Reflection process
simulates light reflection mechanism used by these layers, while
visibility process simulates visibility of matching patterns of the
cuttlefish. To show the effectiveness of the algorithm, it is tested with
some other popular bio-inspired optimization algorithms such as
Genetic Algorithms (GA), Particle Swarm Optimization (PSO) and
Bees Algorithm (BA) that have been previously proposed in the
literature. Simulations and obtained results indicate that the proposed
CFA is superior when compared with these algorithms.
Abstract: In this paper, we propose a method that allows faster and more accurate detection of traffic lights by a vision sensor during driving, DGPS is used to obtain physical location of a traffic light, extract from the image information of the vision sensor only the traffic light area at this location and ascertain if the sign is in operation and determine its form. This method can solve the problem in existing research where low visibility at night or reflection under bright light makes it difficult to recognize the form of traffic light, thus making driving unstable. We compared our success rate of traffic light recognition in day and night road environments. Compared to previous researches, it showed similar performance during the day but 50% improvement at night.
Abstract: There are many drivers who feel right A pillar of Japanese right-hand-drive car preventing visibility on turning right or left at intersection. On the other hand, there is a report that almost pedestrian accident is caused by the delay of finding pedestrian by drivers and this is found by drivers’ eye movement. Thus, we developed the evaluation method of quantification using drivers’ eye movement data by least squares estimation and we applied this method to commercial vehicle and evaluation the visibility. It is suggested that visibility of vehicle can be quantified and estimated by linear model obtained from experimental eye fixation data and information of vehicle dimensions.
Abstract: In this paper, we present an activity diagram model for double-loop control self-adaptive braking system. Since activity diagram helps to improve visibility of self-adaption. We can easily find where improvement is needed on double-loop control. Double-loop control is adopted since the design conditions and actual conditions can be different. The system is reconfigured in runtime by using double-loop control. We simulated to verify and validate our model by using MATLAB. We compared single-loop control model with double-loop control model. Simulation results show that double-loop control provides more consistent brake power control than single-loop control.
Abstract: In this work, we study elliptic divisibility sequences
over finite fields. Morgan Ward in [14], [15] gave arithmetic theory
of elliptic divisibility sequences and formulas for elliptic divisibility
sequences with rank two over finite field Fp. We study elliptic
divisibility sequences with rank three, four and five over a finite field
Fp, where p > 3 is a prime and give general terms of these sequences
and then we determine elliptic and singular curves associated with
these sequences.
Abstract: In this paper, an experimentation to enhance the
visibility of hot objects in a thermal image acquired with ordinary
digital camera is reported, after the applications of lowpass and
median filters to suppress the distracting granular noises. The
common thresholding and slicing techniques were used on the
histogram at different gray levels, followed by a subjective
comparative evaluation. The best result came out with the threshold
level 115 and the number of slices 3.
Abstract: Lighting is not only important for the safety of traffic,
but also it is very important for the protection of pedestrians.
Improvement on visibility in a long distance, lighting, signing,
reduces considerably the risk of accidents in crosswalks. This paper
evaluates different aspects of crosswalks including signing and
lighting to improve road safety.
Abstract: Both image steganography and image encryption have
advantages and disadvantages. Steganograhy allows us to hide a
desired image containing confidential information in a covered or
host image while image encryption is decomposing the desired image
to a non-readable, non-comprehended manner. The encryption
methods are usually much more robust than the steganographic ones.
However, they have a high visibility and would provoke the attackers
easily since it usually is obvious from an encrypted image that
something is hidden! The combination of steganography and
encryption will cover both of their weaknesses and therefore, it
increases the security. In this paper an image encryption method
based on sinc-convolution along with using an encryption key of 128
bit length is introduced. Then, the encrypted image is covered by a
host image using a modified version of JSteg steganography
algorithm. This method could be applied to almost all image formats
including TIF, BMP, GIF and JPEG. The experiment results show
that our method is able to hide a desired image with high security and
low visibility.
Abstract: The aim of this study was to compare the
sensitometric properties of commonly used radiographic films
processed with chemical solutions in different workload hospitals.
The effect of different processing conditions on induced densities on
radiologic films was investigated. Two accessible double emulsions
Fuji and Kodak films were exposed with 11-step wedge and
processed with Champion and CPAC processing solutions. The
mentioned films provided in both workloads centers, high and low.
Our findings displays that the speed and contrast of Kodak filmscreen
in both work load (high and low) is higher than Fuji filmscreen
for both processing solutions. However there was significant
differences in films contrast for both workloads when CPAC solution
had been used (p=0.000 and 0.028). The results showed base plus
fog density for Kodak film was lower than Fuji. Generally Champion
processing solution caused more speed and contrast for investigated
films in different conditions and there was significant differences in
95% confidence level between two used processing solutions
(p=0.01). Low base plus fog density for Kodak films provide more
visibility and accuracy and higher contrast results in using lower
exposure factors to obtain better quality in resulting radiographs. In
this study we found an economic advantages since Champion
solution and Kodak film are used while it makes lower patient dose.
Thus, in a radiologic facility any change in film processor/processing
cycle or chemistry should be carefully investigated before
radiological procedures of patients are acquired.
Abstract: Effective estimation of just noticeable distortion (JND) for images is helpful to increase the efficiency of a compression algorithm in which both the statistical redundancy and the perceptual redundancy should be accurately removed. In this paper, we design a DCT-based model for estimating JND profiles of color images. Based on a mathematical model of measuring the base detection threshold for each DCT coefficient in the color component of color images, the luminance masking adjustment, the contrast masking adjustment, and the cross masking adjustment are utilized for luminance component, and the variance-based masking adjustment based on the coefficient variation in the block is proposed for chrominance components. In order to verify the proposed model, the JND estimator is incorporated into the conventional JPEG coder to improve the compression performance. A subjective and fair viewing test is designed to evaluate the visual quality of the coding image under the specified viewing condition. The simulation results show that the JPEG coder integrated with the proposed DCT-based JND model gives better coding bit rates at visually lossless quality for a variety of color images.