Are XBRL-based Financial Reports Better than Non-XBRL Reports? A Quality Assessment

Using a scoring system, this paper provides a comparative assessment of the quality of data between XBRL formatted financial reports and non-XBRL financial reports. It shows a major improvement in the quality of data of XBRL formatted financial reports. Although XBRL formatted financial reports do not show much advantage in the quality at the beginning, XBRL financial reports lately display a large improvement in the quality of data in almost all aspects. With the improved XBRL web data managing, presentation and analysis applications, XBRL formatted financial reports have a much better accessibility, are more accurate and better in timeliness.

The Role of Velocity Map Quality in Estimation of Intravascular Pressure Distribution

Phase-Contrast MR imaging methods are widely used for measurement of blood flow velocity components. Also there are some other tools such as CT and Ultrasound for velocity map detection in intravascular studies. These data are used in deriving flow characteristics. Some clinical applications are investigated which use pressure distribution in diagnosis of intravascular disorders such as vascular stenosis. In this paper an approach to the problem of measurement of intravascular pressure field by using velocity field obtained from flow images is proposed. The method presented in this paper uses an algorithm to calculate nonlinear equations of Navier- Stokes, assuming blood as an incompressible and Newtonian fluid. Flow images usually suffer the lack of spatial resolution. Our attempt is to consider the effect of spatial resolution on the pressure distribution estimated from this method. In order to achieve this aim, velocity map of a numerical phantom is derived at six different spatial resolutions. To determine the effects of vascular stenoses on pressure distribution, a stenotic phantom geometry is considered. A comparison between the pressure distribution obtained from the phantom and the pressure resulted from the algorithm is presented. In this regard we also compared the effects of collocated and staggered computational grids on the pressure distribution resulted from this algorithm.

Robot Vision Application based on Complex 3D Pose Computation

The paper presents a technique suitable in robot vision applications where it is not possible to establish the object position from one view. Usually, one view pose calculation methods are based on the correspondence of image features established at a training step and exactly the same image features extracted at the execution step, for a different object pose. When such a correspondence is not feasible because of the lack of specific features a new method is proposed. In the first step the method computes from two views the 3D pose of feature points. Subsequently, using a registration algorithm, the set of 3D feature points extracted at the execution phase is aligned with the set of 3D feature points extracted at the training phase. The result is a Euclidean transform which have to be used by robot head for reorientation at execution step.

Design of Liquids Mixing Control System using Fuzzy Time Control Discrete Event Model for Industrial Applications

This paper presents a time control liquids mixing system in the tanks as an application of fuzzy time control discrete model. The system is designed for a wide range of industrial applications. The simulation design of control system has three inputs: volume, viscosity, and selection of product, along with the three external control adjustments for the system calibration or to take over the control of the system autonomously in local or distributed environment. There are four controlling elements: rotatory motor, grinding motor, heating and cooling units, and valves selection, each with time frame limit. The system consists of three controlled variables measurement through its sensing mechanism for feed back control. This design also facilitates the liquids mixing system to grind certain materials in tanks and mix with fluids under required temperature controlled environment to achieve certain viscous level. Design of: fuzzifier, inference engine, rule base, deffuzifiers, and discrete event control system, is discussed. Time control fuzzy rules are formulated, applied and tested using MATLAB simulation for the system.

Fatigue Analysis of Crack Growing Rate and Stress Intensity Factor for Stress Corrosion Cracking in a Pipeline System

Environment-assisted cracking (EAC) is one of the most serious causes of structural failure over a broad range of industrial applications including offshore structures. In EAC condition there is not a definite relation such as Paris equation in Linear Elastic Fracture Mechanics (LEFM). According to studying and searching a lot what the researchers said either a material has contact with hydrogen or any other corrosive environment, phenomenon of electrical and chemical reactions of material with its environment will be happened. In the literature, there are many different works to consider fatigue crack growing and solve it but they are experimental works. Thus, in this paper, authors have an aim to evaluate mathematically the pervious works in LEFM. Obviously, if an environment is more sour and corrosive, the changes of stress intensity factor is more and the calculation of stress intensity factor is difficult. A mathematical relation to deal with the stress intensity factor during the diffusion of sour environment especially hydrogen in a marine pipeline is presented. By using this relation having and some experimental relation an analytical formulation will be presented which enables the fatigue crack growth and critical crack length under cyclic loading to be predicted. In addition, we can calculate KSCC and stress intensity factor in the pipeline caused by EAC.

A Computational Model of Minimal Consciousness Functions

Interest in Human Consciousness has been revived in the late 20th century from different scientific disciplines. Consciousness studies involve both its understanding and its application. In this paper, a computational model of the minimum consciousness functions necessary in my point of view for Artificial Intelligence applications is presented with the aim of improving the way computations will be made in the future. In section I, human consciousness is briefly described according to the scope of this paper. In section II, a minimum set of consciousness functions is defined - based on the literature reviewed - to be modelled, and then a computational model of these functions is presented in section III. In section IV, an analysis of the model is carried out to describe its functioning in detail.

Multi-Font Farsi/Arabic Isolated Character Recognition Using Chain Codes

Nowadays, OCR systems have got several applications and are increasingly employed in daily life. Much research has been done regarding the identification of Latin, Japanese, and Chinese characters. However, very little investigation has been performed regarding Farsi/Arabic characters recognition. Probably the reason is difficulty and complexity of those characters identification compared to the others and limitation of IT activities in Farsi and Arabic speaking countries. In this paper, a technique has been employed to identify isolated Farsi/Arabic characters. A chain code based algorithm along with other significant peculiarities such as number and location of dots and auxiliary parts, and the number of holes existing in the isolated character has been used in this study to identify Farsi/Arabic characters. Experimental results show the relatively high accuracy of the method developed when it is tested on several standard Farsi fonts.

An Experimental Comparison of Unsupervised Learning Techniques for Face Recognition

Face Recognition has always been a fascinating research area. It has drawn the attention of many researchers because of its various potential applications such as security systems, entertainment, criminal identification etc. Many supervised and unsupervised learning techniques have been reported so far. Principal Component Analysis (PCA), Self Organizing Maps (SOM) and Independent Component Analysis (ICA) are the three techniques among many others as proposed by different researchers for Face Recognition, known as the unsupervised techniques. This paper proposes integration of the two techniques, SOM and PCA, for dimensionality reduction and feature selection. Simulation results show that, though, the individual techniques SOM and PCA itself give excellent performance but the combination of these two can also be utilized for face recognition. Experimental results also indicate that for the given face database and the classifier used, SOM performs better as compared to other unsupervised learning techniques. A comparison of two proposed methodologies of SOM, Local and Global processing, shows the superiority of the later but at the cost of more computational time.

Applications of Entropy Measures in Field of Queuing Theory

In the present communication, we have studied different variations in the entropy measures in the different states of queueing processes. In case of steady state queuing process, it has been shown that as the arrival rate increases, the uncertainty increases whereas in the case of non-steady birth-death process, it is shown that the uncertainty varies differently. In this pattern, it first increases and attains its maximum value and then with the passage of time, it decreases and attains its minimum value.

A Force Measurement Evaluation Tool for Telerobotic Cutting Applications: Development of an Effective Characterization Platform

Sensorized instruments that accurately measure the interaction forces (between biological tissue and instrument endeffector) during surgical procedures offer surgeons a greater sense of immersion during minimally invasive robotic surgery. Although there is ongoing research into force measurement involving surgical graspers little corresponding effort has been carried out on the measurement of forces between scissor blades and tissue. This paper presents the design and development of a force measurement test apparatus, which will serve as a sensor characterization and evaluation platform. The primary aim of the experiments is to ascertain whether the system can differentiate between tissue samples with differing mechanical properties in a reliable, repeatable manner. Force-angular displacement curves highlight trends in the cutting process as well the forces generated along the blade during a cutting procedure. Future applications of the test equipment will involve the assessment of new direct force sensing technologies for telerobotic surgery.

A Mobile Multihop Relay Dynamic TDD Scheme for Cellular Networks

In this paper, we present an analytical framework for the evaluation of the uplink performance of multihop cellular networks based on dynamic time division duplex (TDD). New wireless broadband protocols, such as WiMAX, WiBro, and 3G-LTE apply TDD, and mobile communication protocols under standardization (e.g., IEEE802.16j) are investigating mobile multihop relay (MMR) as a future technology. In this paper a novel MMR TDD scheme is presented, where the dynamic range of the frame is shared to traffic resources of asymmetric nature and multihop relaying. The mobile communication channel interference model comprises of inner and co-channel interference (CCI). The performance analysis focuses on the uplink due to the fact that the effects of dynamic resource allocation show significant performance degradation only in the uplink compared to time division multiple access (TDMA) schemes due to CCI [1-3], where the downlink results to be the same or better.The analysis was based on the signal to interference power ratio (SIR) outage probability of dynamic TDD (D-TDD) and TDMA systems,which are the most widespread mobile communication multi-user control techniques. This paper presents the uplink SIR outage probability with multihop results and shows that the dynamic TDD scheme applying MMR can provide a performance improvement compared to single hop applications if executed properly.

Artificial Neural Networks Modeling in Water Resources Engineering: Infrastructure and Applications

The use of artificial neural network (ANN) modeling for prediction and forecasting variables in water resources engineering are being increasing rapidly. Infrastructural applications of ANN in terms of selection of inputs, architecture of networks, training algorithms, and selection of training parameters in different types of neural networks used in water resources engineering have been reported. ANN modeling conducted for water resources engineering variables (river sediment and discharge) published in high impact journals since 2002 to 2011 have been examined and presented in this review. ANN is a vigorous technique to develop immense relationship between the input and output variables, and able to extract complex behavior between the water resources variables such as river sediment and discharge. It can produce robust prediction results for many of the water resources engineering problems by appropriate learning from a set of examples. It is important to have a good understanding of the input and output variables from a statistical analysis of the data before network modeling, which can facilitate to design an efficient network. An appropriate training based ANN model is able to adopt the physical understanding between the variables and may generate more effective results than conventional prediction techniques.

Quantum Computing: A New Era of Computing

Nature conducts its action in a very private manner. To reveal these actions classical science has done a great effort. But classical science can experiment only with the things that can be seen with eyes. Beyond the scope of classical science quantum science works very well. It is based on some postulates like qubit, superposition of two states, entanglement, measurement and evolution of states that are briefly described in the present paper. One of the applications of quantum computing i.e. implementation of a novel quantum evolutionary algorithm(QEA) to automate the time tabling problem of Dayalbagh Educational Institute (Deemed University) is also presented in this paper. Making a good timetable is a scheduling problem. It is NP-hard, multi-constrained, complex and a combinatorial optimization problem. The solution of this problem cannot be obtained in polynomial time. The QEA uses genetic operators on the Q-bit as well as updating operator of quantum gate which is introduced as a variation operator to converge toward better solutions.

Moving Vehicles Detection Using Automatic Background Extraction

Vehicle detection is the critical step for highway monitoring. In this paper we propose background subtraction and edge detection technique for vehicle detection. This technique uses the advantages of both approaches. The practical applications approved the effectiveness of this method. This method consists of two procedures: First, automatic background extraction procedure, in which the background is extracted automatically from the successive frames; Second vehicles detection procedure, which depend on edge detection and background subtraction. Experimental results show the effective application of this algorithm. Vehicles detection rate was higher than 91%.

A Comparative Study on the Dimensional Error of 3D CAD Model and SLS RP Model for Reconstruction of Cranial Defect

Rapid Prototyping (RP) is a technology that produces models and prototype parts from 3D CAD model data, CT/MRI scan data, and model data created from 3D object digitizing systems. There are several RP process like Stereolithography (SLA), Solid Ground Curing (SGC), Selective Laser Sintering (SLS), Fused Deposition Modeling (FDM), 3D Printing (3DP) among them SLS and FDM RP processes are used to fabricate pattern of custom cranial implant. RP technology is useful in engineering and biomedical application. This is helpful in engineering for product design, tooling and manufacture etc. RP biomedical applications are design and development of medical devices, instruments, prosthetics and implantation; it is also helpful in planning complex surgical operation. The traditional approach limits the full appreciation of various bony structure movements and therefore the custom implants produced are difficult to measure the anatomy of parts and analyze the changes in facial appearances accurately. Cranioplasty surgery is a surgical correction of a defect in cranial bone by implanting a metal or plastic replacement to restore the missing part. This paper aims to do a comparative study on the dimensional error of CAD and SLS RP Models for reconstruction of cranial defect by comparing the virtual CAD with the physical RP model of a cranial defect.

Mechanical Properties of Ultra High Performance Concrete

A research program is conducted to evaluate the mechanical properties of Ultra High Performance Concrete, target compressive strength at the age of 28 days being more than 150 MPa. The methodology to develop such mix has been explained. The material properties, mix design and curing regime are determined. The material attributes are understood by studying the stress strain behaviour of UHPC cylinders under uniaxial compressive loading. The load –crack mouth opening displacement (cmod) of UHPC beams, flexural strength and fracture energy was evaluated using third point loading test. Compressive strength and Split tensile strength results are determined to find out the compressive and tensile behaviour. Residual strength parameters are presented vividly explaining the flexural performance, toughness of concrete.Durability studies were also done to compare the effect of fibre to that of a control mix For all the studies the Mechanical properties were evaluated by varying the percentage and aspect ratio of steel fibres The results reflected that higher aspect ratio and fibre volume produced drastic changes in the cube strength, cylinder strength, post peak response, load-cmod, fracture energy flexural strength, split tensile strength, residual strength and durability. In regards to null application of UHPC in India, an initiative is undertaken to comprehend the mechanical behaviour of UHPC, which will be vital for longer run in commercialization for structural applications.

New Device for Enhancement of Liposomal Magnetofection Efficiency of Cancer Cells

Liposomal magnetofection is the most powerful nonviral method for the nucleic acid delivery into the cultured cancer cells and widely used for in vitro applications. Use of the static magnetic field condition may result in non-uniform distribution of aggregate complexes on the surface of cultured cells. To prevent this, we developed the new device which allows to concentrate aggregate complexes under dynamic magnetic field, assisting more contact of these complexes with cellular membrane and, possibly, stimulating endocytosis. Newly developed device for magnetofection under dynamic gradient magnetic field, “DynaFECTOR", was used to compare transfection efficiency of human liver hepatocellular carcinoma cell line HepG2 with that obtained by lipofection and magnetofection. The effect of two parameters on transfection efficiency, incubation time under dynamic magnetic field and rotation frequency of magnet, was estimated. Liposomal magnetofection under dynamic gradient magnetic field showed the highest transfection efficiency for HepG2 cells.

An Intelligent Approach of Rough Set in Knowledge Discovery Databases

Knowledge Discovery in Databases (KDD) has evolved into an important and active area of research because of theoretical challenges and practical applications associated with the problem of discovering (or extracting) interesting and previously unknown knowledge from very large real-world databases. Rough Set Theory (RST) is a mathematical formalism for representing uncertainty that can be considered an extension of the classical set theory. It has been used in many different research areas, including those related to inductive machine learning and reduction of knowledge in knowledge-based systems. One important concept related to RST is that of a rough relation. In this paper we presented the current status of research on applying rough set theory to KDD, which will be helpful for handle the characteristics of real-world databases. The main aim is to show how rough set and rough set analysis can be effectively used to extract knowledge from large databases.

Evaluation of the Impact of Dataset Characteristics for Classification Problems in Biological Applications

Availability of high dimensional biological datasets such as from gene expression, proteomic, and metabolic experiments can be leveraged for the diagnosis and prognosis of diseases. Many classification methods in this area have been studied to predict disease states and separate between predefined classes such as patients with a special disease versus healthy controls. However, most of the existing research only focuses on a specific dataset. There is a lack of generic comparison between classifiers, which might provide a guideline for biologists or bioinformaticians to select the proper algorithm for new datasets. In this study, we compare the performance of popular classifiers, which are Support Vector Machine (SVM), Logistic Regression, k-Nearest Neighbor (k-NN), Naive Bayes, Decision Tree, and Random Forest based on mock datasets. We mimic common biological scenarios simulating various proportions of real discriminating biomarkers and different effect sizes thereof. The result shows that SVM performs quite stable and reaches a higher AUC compared to other methods. This may be explained due to the ability of SVM to minimize the probability of error. Moreover, Decision Tree with its good applicability for diagnosis and prognosis shows good performance in our experimental setup. Logistic Regression and Random Forest, however, strongly depend on the ratio of discriminators and perform better when having a higher number of discriminators.