Design and Performance Improvement of Three-Dimensional Optical Code Division Multiple Access Networks with NAND Detection Technique
In this paper, we have presented and analyzed three-dimensional (3-D) matrices of wavelength/time/space code for optical code division multiple access (OCDMA) networks with NAND subtraction detection technique. The 3-D codes are constructed by integrating a two-dimensional modified quadratic congruence (MQC) code with one-dimensional modified prime (MP) code. The respective encoders and decoders were designed using fiber Bragg gratings and optical delay lines to minimize the bit error rate (BER). The performance analysis of the 3D-OCDMA system is based on measurement of signal to noise ratio (SNR), BER and eye diagram for a different number of simultaneous users. Also, in the analysis, various types of noises and multiple access interference (MAI) effects were considered. The results obtained with NAND detection technique were compared with those obtained with OR and AND subtraction techniques. The comparison results proved that the NAND detection technique with 3-D MQC\MP code can accommodate more number of simultaneous users for longer distances of fiber with minimum BER as compared to OR and AND subtraction techniques. The received optical power is also measured at various levels of BER to analyze the effect of attenuation.
A Route Guidance System for Car Finding in Indoor Parking Garages
This paper presents a route guidance system for car owners to find their cars in parking garages. The presents system comprises a positioning-assisting subsystem and a car-finding mobile app. The positioning-assisting subsystem mainly uses the iBeacon technology for indoor positioning. The car-finding mobile app guides car owners to their cars based on a non-map navigation strategy. This study also designs a virtual coordinate system to support identifying the locations of parking spaces and iBeacon devices. We use Arduino and Android as the platforms to implement the proposed positioning-assisting subsystem and car-finding mobile app, respectively. We have also deployed the system in a parking garage in our campus for testing. Experimental results verify that our system can efficiently and correctly guide car owners to the parking spaces of their cars.
Ambient Notifications and the Interruption Effect
The technology of mobile devices has changed our daily lives. Since smartphone have become a multi-functional device, many people spend unnecessary time on them, and could be interrupted by inappropriate notifications such as unimportant messages from social media. Notifications from smartphone could draw people’s attention and distract them from their priorities and current tasks. This research investigated that if the users were notified by their surroundings instead of smartphone, would it create less distraction and keep their focus on the present task. The experiment was a simulation of a lamp and door notification. Notifications related to work will be embedded in the lamp such as an email from a colleague. A notification that is useful when going outside such as weather information, traffic information, and schedule reminder will be embedded in the door. The experiment was conducted by sending notifications to the participant while he or she was working on a primary task and the working performance was measured. The results show that the lamp notification had fewer interruption effects than the smartphone. For the door notification, it was simulated in order to gain opinions and insights on ambient notifications from participants. Many participants agreed that the ambient notifications are useful and being informed by them could lessen the usage of their smartphone. The results and insights from this research could be used to guide the design process of ambient notifications.
Performance Comparison of Different Regression Methods for a Polymerization Process with Adaptive Sampling
Developing complete mechanistic models for polymerization reactors is not easy, because complex reactions occur simultaneously; there is a large number of kinetic parameters involved and sometimes the chemical and physical phenomena for mixtures involving polymers are poorly understood. To overcome these difficulties, empirical models based on sampled data can be used instead, namely regression methods typical of machine learning field. They have the ability to learn the trends of a process without any knowledge about its particular physical and chemical laws. Therefore, they are useful for modeling complex processes, such as the free radical polymerization of methyl methacrylate achieved in a batch bulk process. The goal is to generate accurate predictions of monomer conversion, numerical average molecular weight and gravimetrical average molecular weight. This process is associated with non-linear gel and glass effects. For this purpose, an adaptive sampling technique is presented, which can select more samples around the regions where the values have a higher variation. Several machine learning methods are used for the modeling and their performance is compared: support vector machines, k-nearest neighbor, k-nearest neighbor and random forest, as well as an original algorithm, large margin nearest neighbor regression. The suggested method provides very good results compared to the other well-known regression algorithms.
Efficient Semi-Systolic Finite Field Multiplier Using Redundant Basis
The arithmetic operations over GF(2m) have been
extensively used in error correcting codes and public-key
cryptography schemes. Finite field arithmetic includes addition,
multiplication, division and inversion operations. Addition is very
simple and can be implemented with an extremely simple circuit.
The other operations are much more complex. The multiplication
is the most important for cryptosystems, such as the elliptic
curve cryptosystem, since computing exponentiation, division, and
computing multiplicative inverse can be performed by computing
multiplication iteratively. In this paper, we present a parallel
computation algorithm that operates Montgomery multiplication over
finite field using redundant basis. Also, based on the multiplication
algorithm, we present an efficient semi-systolic multiplier over finite
field. The multiplier has less space and time complexities compared
to related multipliers. As compared to the corresponding existing
structures, the multiplier saves at least 5% area, 50% time, and 53%
area-time (AT) complexity. Accordingly, it is well suited for VLSI
implementation and can be easily applied as a basic component for
computing complex operations over finite field, such as inversion and
Faster Pedestrian Recognition Using Deformable Part Models
Deformable part models achieve high precision in
pedestrian recognition, but all publicly available implementations are
too slow for real-time applications. We implemented a deformable
part model algorithm fast enough for real-time use by exploiting
information about the camera position and orientation. This
implementation is both faster and more precise than alternative
DPM implementations. These results are obtained by computing
convolutions in the frequency domain and using lookup tables to
speed up feature computation. This approach is almost an order of
magnitude faster than the reference DPM implementation, with no
loss in precision. Knowing the position of the camera with respect to
horizon it is also possible prune many hypotheses based on their
size and location. The range of acceptable sizes and positions is
set by looking at the statistical distribution of bounding boxes in
labelled images. With this approach it is not needed to compute the
entire feature pyramid: for example higher resolution features are
only needed near the horizon. This results in an increase in mean
average precision of 5% and an increase in speed by a factor of
two. Furthermore, to reduce misdetections involving small pedestrians
near the horizon, input images are supersampled near the horizon.
Supersampling the image at 1.5 times the original scale, results in
an increase in precision of about 4%. The implementation was tested
against the public KITTI dataset, obtaining an 8% improvement in
mean average precision over the best performing DPM-based method.
By allowing for a small loss in precision computational time can be
easily brought down to our target of 100ms per image, reaching a
solution that is faster and still more precise than all publicly available
Big Data: Concepts, Technologies and Applications in the Public Sector
Big Data (BD) is associated with a new generation of technologies and architectures which can harness the value of extremely large volumes of very varied data through real time processing and analysis. It involves changes in (1) data types, (2) accumulation speed, and (3) data volume. This paper presents the main concepts related to the BD paradigm, and introduces architectures and technologies for BD and BD sets. The integration of BD with the Hadoop Framework is also underlined. BD has attracted a lot of attention in the public sector due to the newly emerging technologies that allow the availability of network access. The volume of different types of data has exponentially increased. Some applications of BD in the public sector in Romania are briefly presented.
Using Multi-Arm Bandits to Optimize Game Play Metrics and Effective Game Design
Game designers have the challenging task of building games that engage players to spend their time and money on the game. There are an infinite number of game variations and design choices, and it is hard to systematically determine game design choices that will have positive experiences for players. In this work, we demonstrate how multi-arm bandits can be used to automatically explore game design variations to achieve improved player metrics. The advantage of multi-arm bandits is that they allow for continuous experimentation and variation, intrinsically converge to the best solution, and require no special infrastructure to use beyond allowing minor game variations to be deployed to users for evaluation. A user study confirms that applying multi-arm bandits was successful in determining the preferred game variation with highest play time metrics and can be a useful technique in a game designer's toolkit.
Study of Icons in Enterprise Application Software Context
Icons are not merely decorative elements in enterprise applications but very often used because of their many advantages such as compactness, visual appeal, etc. Despite these potential advantages, icons often cause usability problems when they are designed without consideration for their many potential downsides. The aim of the current study was to examine the effect of articulatory distance – the distance between the physical appearance of an interface element and what it actually means. In other words, will the subject find the association of the function and its appearance on the interface natural or is the icon difficult for them to associate with its function. We have calculated response time and quality of identification by varying icon concreteness, the context of usage and subject experience in the enterprise context. The subjects were asked to associate icons (prepared for study purpose) with given function options in context and out of context mode. Response time and their selection were recorded for analysis.
Design of a Service-Enabled Dependable Integration Environment
The aim of information systems integration is to make all the data sources, applications and business flows integrated into the new environment so that unwanted redundancies are reduced and bottlenecks and mismatches are eliminated. Two issues have to be dealt with to meet such requirements: the software architecture that supports resource integration, and the adaptor development tool that help integration and migration of legacy applications. In this paper, a service-enabled dependable integration environment (SDIE), is presented, which has two key components, i.e., a dependable service integration platform and a legacy application integration tool. For the dependable platform for service integration, the service integration bus, the service management framework, the dependable engine for service composition, and the service registry and discovery components are described. For the legacy application integration tool, its basic organization, functionalities and dependable measures taken are presented. Due to its service-oriented integration model, the light-weight extensible container, the service component combination-oriented p-lattice structure, and other features, SDIE has advantages in openness, flexibility, performance-price ratio and feature support over commercial products, is better than most of the open source integration software in functionality, performance and dependability support.
Design and Implementation of a Counting and Differentiation System for Vehicles through Video Processing
This paper presents a self-sustaining mobile system for
counting and classification of vehicles through processing video. It
proposes a counting and classification algorithm divided in four steps
that can be executed multiple times in parallel in a SBC (Single
Board Computer), like the Raspberry Pi 2, in such a way that it
can be implemented in real time. The first step of the proposed
algorithm limits the zone of the image that it will be processed.
The second step performs the detection of the mobile objects using
a BGS (Background Subtraction) algorithm based on the GMM
(Gaussian Mixture Model), as well as a shadow removal algorithm
using physical-based features, followed by morphological operations.
In the first step the vehicle detection will be performed by using
edge detection algorithms and the vehicle following through Kalman
filters. The last step of the proposed algorithm registers the vehicle
passing and performs their classification according to their areas.
An auto-sustainable system is proposed, powered by batteries and
photovoltaic solar panels, and the data transmission is done through
GPRS (General Packet Radio Service)eliminating the need of using
external cable, which will facilitate it deployment and translation to
any location where it could operate. The self-sustaining trailer will
allow the counting and classification of vehicles in specific zones
with difficult access.
Information Technology and Business Alignments among Different Divisions: A Comparative Analysis of Japan and South Korea
This paper empirically investigates whether information technology (IT) strategies, business strategies, and divisions are aligned to meet overall business goals for Korean Small and medium-sized enterprises (SMEs), based on structure based Strategic Alignment Model, and make comparison with those of Japanese SMEs. Using 2,869 valid responses of Korean Human Capital Corporate Panel survey, a result of this study suggests that Korean human resources (HR) departments have a major influence over IT strategy, which is the same as Japanese SMEs, even though their management styles are quite different. As for IT strategy, it is not related to other departments at all for Korean SMEs. The Korean management seems to possess a great power over each division, such as Sales/Service, Research and Development/Technical Experts, HR, and Production.
Portable Hands-Free Process Assistant for Gas Turbine Maintenance
This paper presents how smart glasses and voice commands can be used for improving the maintenance process of industrial gas turbines. It presents the process of inspecting a gas turbine’s combustion chamber and how it is currently performed using a set of paper-based documents. In order to improve this process, a portable hands-free process assistance system has been conceived. In the following, it will be presented how the approach of user-centered design and the method of paper prototyping have been successfully applied in order to design a user interface and a corresponding workflow model that describes the possible interaction patterns between the user and the interface. The presented evaluation of these results suggests that the assistance system could help the user by rendering multiple manual activities obsolete, thus allowing him to work hands-free and to save time for generating protocols.
Adapting Tools for Text Monitoring and for Scenario Analysis Related to the Field of Social Disasters
Humanity faces more and more often with different social disasters, which in turn can generate new accidents and catastrophes. To mitigate their consequences, it is important to obtain early possible signals about the events which are or can occur and to prepare the corresponding scenarios that could be applied. Our research is focused on solving two problems in this domain: identifying signals related that an accident occurred or may occur and mitigation of some consequences of disasters. To solve the first problem, methods of selecting and processing texts from global network Internet are developed. Information in Romanian is of special interest for us. In order to obtain the mentioned tools, we should follow several steps, divided into preparatory stage and processing stage. Throughout the first stage, we manually collected over 724 news articles and classified them into 10 categories of social disasters. It constitutes more than 150 thousand words. Using this information, a controlled vocabulary of more than 300 keywords was elaborated, that will help in the process of classification and identification of the texts related to the field of social disasters. To solve the second problem, the formalism of Petri net has been used. We deal with the problem of inhabitants’ evacuation in useful time. The analysis methods such as reachability or coverability tree and invariants technique to determine dynamic properties of the modeled systems will be used. To perform a case study of properties of extended evacuation system by adding time, the analysis modules of PIPE such as Generalized Stochastic Petri Nets (GSPN) Analysis, Simulation, State Space Analysis, and Invariant Analysis have been used. These modules helped us to obtain the average number of persons situated in the rooms and the other quantitative properties and characteristics related to its dynamics.
Image Retrieval Based on Multi-Feature Fusion for Heterogeneous Image Databases
Selecting an appropriate image representation is the most important factor in implementing an effective Content-Based Image Retrieval (CBIR) system. This paper presents a multi-feature fusion approach for efficient CBIR, based on the distance distribution of features and relative feature weights at the time of query processing. It is a simple yet effective approach, which is free from the effect of features' dimensions, ranges, internal feature normalization and the distance measure. This approach can easily be adopted in any feature combination to improve retrieval quality. The proposed approach is empirically evaluated using two benchmark datasets for image classification (a subset of the Corel dataset and Oliva and Torralba) and compared with existing approaches. The performance of the proposed approach is confirmed with the significantly improved performance in comparison with the independently evaluated baseline of the previously proposed feature fusion approaches.
Development of Configuration Software of Space Environment Simulator Control System Based on Linux
This paper presents a configuration software solution in Linux, which is used for the control of space environment simulator. After introducing the structure and basic principle, it is said that the developing of QT software frame and the dynamic data exchanging between PLC and computer. The OPC driver in Linux is also developed. This driver realizes many-to-many communication between hardware devices and SCADA software. Moreover, an algorithm named “Scan PRI” is put forward. This algorithm is much more optimizable and efficient compared with "Scan in sequence" in Windows. This software has been used in practical project. It has a good control effect and can achieve the expected goal.
Analysis of Urban Population Using Twitter Distribution Data: Case Study of Makassar City, Indonesia
In the past decade, the social networking app has been growing very rapidly. Geolocation data is one of the important features of social media that can attach the user's location coordinate in the real world. This paper proposes the use of geolocation data from the Twitter social media application to gain knowledge about urban dynamics, especially on human mobility behavior. This paper aims to explore the relation between geolocation Twitter with the existence of people in the urban area. Firstly, the study will analyze the spread of people in the particular area, within the city using Twitter social media data. Secondly, we then match and categorize the existing place based on the same individuals visiting. Then, we combine the Twitter data from the tracking result and the questionnaire data to catch the Twitter user profile. To do that, we used the distribution frequency analysis to learn the visitors’ percentage. To validate the hypothesis, we compare it with the local population statistic data and land use mapping released by the city planning department of Makassar local government. The results show that there is the correlation between Twitter geolocation and questionnaire data. Thus, integration the Twitter data and survey data can reveal the profile of the social media users.
Communicative and Artistic Machines: A Survey of Models and Experiments on Artificial Agents
Machines can be either tool, media, or social agents. Advances in technology have been delivering machines capable of autonomous expression, both through communication and art. This paper deals with models (theoretical approach) and experiments (applied approach) related to artificial agents. On one hand it traces how social sciences' scholars have worked with topics such as text automatization, man-machine writing cooperation, and communication. On the other hand it covers how computer sciences' scholars have built communicative and artistic machines, including the programming of creativity. The aim is to present a brief survey on artificially intelligent communicators and artificially creative writers, and provide the basis to understand the meta-authorship and also to new and further man-machine co-authorship.
Proposal of Commutation Protocol in Hybrid Sensors and Vehicular Networks for Intelligent Transport Systems
Hybrid Sensors and Vehicular Networks (HSVN), represent a hybrid network, which uses several generations of Ad-Hoc networks. It is used especially in Intelligent Transport Systems (ITS). The HSVN allows making collaboration between the Wireless Sensors Network (WSN) deployed on the border of the road and the Vehicular Network (VANET). This collaboration is defined by messages exchanged between the two networks for the purpose to inform the drivers about the state of the road, provide road safety information and more information about traffic on the road. Moreover, this collaboration created by HSVN, also allows the use of a network and the advantage of improving another network. For example, the dissemination of information between the sensors quickly decreases its energy, and therefore, we can use vehicles that do not have energy constraint to disseminate the information between sensors. On the other hand, to solve the disconnection problem in VANET, the sensors can be used as gateways that allow sending the messages received by one vehicle to another. However, because of the short communication range of the sensor and its low capacity of storage and processing of data, it is difficult to ensure the exchange of road messages between it and the vehicle, which can be moving at high speed at the time of exchange. This represents the time where the vehicle is in communication range with the sensor. This work is the proposition of a communication protocol between the sensors and the vehicle used in HSVN. The latter has as the purpose to ensure the exchange of road messages in the available time of exchange.
Topology-Based Character Recognition Method for Coin Date Detection
For recognizing coins, the graved release date is important information to identify precisely its monetary type. However, reading characters in coins meets much more obstacles than traditional character recognition tasks in the other fields, such as reading scanned documents or license plates. To address this challenging issue in a numismatic context, we propose a training-free approach dedicated to detection and recognition of the release date of the coin. In the first step, the date zone is detected by comparing histogram features; in the second step, a topology-based algorithm is introduced to recognize coin numbers with various font types represented by binary gradient map. Our method obtained a recognition rate of 92% on synthetic data and of 44% on real noised data.
A Spatial Hypergraph Based Semi-Supervised Band Selection Method for Hyperspectral Imagery Semantic Interpretation
Hyperspectral imagery (HSI) typically provides a
wealth of information captured in a wide range of the
electromagnetic spectrum for each pixel in the image. Hence, a
pixel in HSI is a high-dimensional vector of intensities with a
large spectral range and a high spectral resolution. Therefore, the
semantic interpretation is a challenging task of HSI analysis. We
focused in this paper on object classification as HSI semantic
interpretation. However, HSI classification still faces some issues,
among which are the following: The spatial variability of spectral
signatures, the high number of spectral bands, and the high cost
of true sample labeling. Therefore, the high number of spectral
bands and the low number of training samples pose the problem of
the curse of dimensionality. In order to resolve this problem, we
propose to introduce the process of dimensionality reduction trying
to improve the classification of HSI. The presented approach is a
semi-supervised band selection method based on spatial hypergraph
embedding model to represent higher order relationships with
different weights of the spatial neighbors corresponding to the
centroid of pixel. This semi-supervised band selection has been
developed to select useful bands for object classification. The
presented approach is evaluated on AVIRIS and ROSIS HSIs
and compared to other dimensionality reduction methods. The
experimental results demonstrate the efficacy of our approach
compared to many existing dimensionality reduction methods for
A Neural Approach for Color-Textured Images Segmentation
In this paper, we present a neural approach for
unsupervised natural color-texture image segmentation, which is
based on both Kohonen maps and mathematical morphology, using
a combination of the texture and the image color information of the
image, namely, the fractal features based on fractal dimension are
selected to present the information texture, and the color features
presented in RGB color space. These features are then used to train
the network Kohonen, which will be represented by the underlying
probability density function, the segmentation of this map is made
by morphological watershed transformation. The performance of our
color-texture segmentation approach is compared first, to color-based
methods or texture-based methods only, and then to k-means method.
Working Mode and Key Technology of Thermal Vacuum Test Software for Spacecraft Test
A universal software platform is developed for improving the defects in the practical one. This software platform has distinct advantages in modularization, information management, and the interfaces. Several technologies such as computer technology, virtualization technology, network technology, etc. are combined together in this software platform, and four working modes are introduced in this article including single mode, distributed mode, cloud mode, and the centralized mode. The application area of the software platform is extended through the switch between these working modes. The software platform can arrange the thermal vacuum test process automatically. This function can improve the reliability of thermal vacuum test.
Comparison of Power Consumption of WiFi Inbuilt Internet of Things Device with Bluetooth Low Energy
The Internet of things (IoT) is currently a highly
researched topic, especially within the context of the smart home.
These are small sensors that are capable of gathering data and
transmitting it to a server. The majority of smart home products use
protocols such as ZigBee or Bluetooth Low Energy (BLE). As these
small sensors are increasing in number, the need to implement these
with much more capable and ubiquitous transmission technology is
necessary. The high power consumption is the reason that holds
these small sensors back from using other protocols such as the
most ubiquitous form of communication, WiFi. Comparing the power
consumption of existing transmission technologies to one with WiFi
inbuilt, would provide a better understanding for choosing between
these technologies. We have developed a small IoT device with WiFi
capability and proven that it is much more efficient than the first
protocol, 433 MHz. We extend our work in this paper and compare
WiFi power consumption with the other most widely used protocol
BLE. The experimental results in this paper would conclude whether
the developed prototype is capable in terms of power consumption to
replace the existing protocol BLE with WiFi.
Knowledge Management in a Combined/Joint Environment
In the current era of shrinking budgets, increasing amounts of worldwide natural disasters, state and non-state initiated conflicts within the world. The response has involved multinational coalitions to conduct effective military operations. The need for a Knowledge Management strategy when developing these coalitions have been overlooked in the past and the need for developing these accords early on will save time and help shape the way information and knowledge are transferred from the staff and action officers of the coalition to the decision-makers in order to make timely decisions within an ever changing environment. The aim of this paper is to show how Knowledge Management has developed within the United States military and how the transformation of working within a Combined/ Joint environment in both the Middle East and the Far East has improved relations between members of the coalitions as well as being more effective as a military force. These same principles could be applied to multinational corporations when dealing with cultures and decision-making processes.
Energy Efficient Resource Allocation and Scheduling in Cloud Computing Platform
There has been renewal of interest in the relation between Green IT and cloud computing in recent years. Cloud computing has to be a highly elastic environment which provides stable services to users. The growing use of cloud computing facilities has caused marked energy consumption, putting negative pressure on electricity cost of computing center or data center. Each year more and more network devices, storages and computers are purchased and put to use, but it is not just the number of computers that is driving energy consumption upward. We could foresee that the power consumption of cloud computing facilities will double, triple, or even more in the next decade. This paper aims at resource allocation and scheduling technologies that are short of or have not well developed yet to reduce energy utilization in cloud computing platform. In particular, our approach relies on recalling services dynamically onto appropriate amount of the machines according to user’s requirement and temporarily shutting down the machines after finish in order to conserve energy. We present initial work on integration of resource and power management system that focuses on reducing power consumption such that they suffice for meeting the minimizing quality of service required by the cloud computing platform.
A Multi-Modal Virtual Walkthrough of the Virtual Past and Present Based on Panoramic View, Crowd Simulation and Acoustic Heritage on Mobile Platform
This research presents a multi-modal simulation in the reconstruction of the past and the construction of present in digital cultural heritage on mobile platform. In bringing the present life, the virtual environment is generated through a presented scheme for rapid and efficient construction of 360° panoramic view. Then, acoustical heritage model and crowd model are presented and improvised into the 360° panoramic view. For the reconstruction of past life, the crowd is simulated and rendered in an old trading port. However, the keystone of this research is in a virtual walkthrough that shows the virtual present life in 2D and virtual past life in 3D, both in an environment of virtual heritage sites in George Town through mobile device. Firstly, the 2D crowd is modelled and simulated using OpenGL ES 1.1 on mobile platform. The 2D crowd is used to portray the present life in 360° panoramic view of a virtual heritage environment based on the extension of Newtonian Laws. Secondly, the 2D crowd is animated and rendered into 3D with improved variety and incorporated into the virtual past life using Unity3D Game Engine. The behaviours of the 3D models are then simulated based on the enhancement of the classical model of Boid algorithm. Finally, a demonstration system is developed and integrated with the models, techniques and algorithms of this research. The virtual walkthrough is demonstrated to a group of respondents and is evaluated through the user-centred evaluation by navigating around the demonstration system. The results of the evaluation based on the questionnaires have shown that the presented virtual walkthrough has been successfully deployed through a multi-modal simulation and such a virtual walkthrough would be particularly useful in a virtual tour and virtual museum applications.
Urban Growth Analysis Using Multi-Temporal Satellite Images, Non-stationary Decomposition Methods and Stochastic Modeling
Remotely sensed data are a significant source for monitoring and updating databases for land use/cover. Nowadays, changes detection of urban area has been a subject of intensive researches. Timely and accurate data on spatio-temporal changes of urban areas are therefore required. The data extracted from multi-temporal satellite images are usually non-stationary. In fact, the changes evolve in time and space. This paper is an attempt to propose a methodology for changes detection in urban area by combining a non-stationary decomposition method and stochastic modeling. We consider as input of our methodology a sequence of satellite images I1, I2, … In at different periods (t = 1, 2, ..., n). Firstly, a preprocessing of multi-temporal satellite images is applied. (e.g. radiometric, atmospheric and geometric). The systematic study of global urban expansion in our methodology can be approached in two ways: The first considers the urban area as one same object as opposed to non-urban areas (e.g. vegetation, bare soil and water). The objective is to extract the urban mask. The second one aims to obtain a more knowledge of urban area, distinguishing different types of tissue within the urban area. In order to validate our approach, we used a database of Tres Cantos-Madrid in Spain, which is derived from Landsat for a period (from January 2004 to July 2013) by collecting two frames per year at a spatial resolution of 25 meters. The obtained results show the effectiveness of our method.
Head-Mounted Displays for HCI Validations While Driving
To provide reliable and valid findings when evaluating innovative in-car devices in the automotive context highly realistic driving environments are recommended. Nowadays, in-car devices are mostly evaluated due to driving simulator studies followed by real car driving experiments. Driving simulators are characterized by high internal validity, but weak regarding ecological validity. Real car driving experiments are ecologically valid, but difficult to standardize, more time-robbing and costly. One economizing suggestion is to implement more immersive driving environments when applying driving simulator studies. This paper presents research comparing non-immersive standard PC conditions with mobile and highly immersive Oculus Rift conditions while performing the Lane Change Task (LCT). Subjective data with twenty participants show advantages regarding presence and immersion experience when performing the LCT with the Oculus Rift, but affect adversely cognitive workload and simulator sickness, compared to non-immersive PC condition.
Context Detection in Spreadsheets Based on Automatically Inferred Table Schema
Programming requires years of training. With natural language and end user development methods, programming could become available to everyone. It enables end users to program their own devices and extend the functionality of the existing system without any knowledge of programming languages. In this paper, we describe an Interactive Spreadsheet Processing Module (ISPM), a natural language interface to spreadsheets that allows users to address ranges within the spreadsheet based on inferred table schema. Using the ISPM, end users are able to search for values in the schema of the table and to address the data in spreadsheets implicitly. Furthermore, it enables them to select and sort the spreadsheet data by using natural language. ISPM uses a machine learning technique to automatically infer areas within a spreadsheet, including different kinds of headers and data ranges. Since ranges can be identified from natural language queries, the end users can query the data using natural language. During the evaluation 12 undergraduate students were asked to perform operations (sum, sort, group and select) using the system and also Excel without ISPM interface, and the time taken for task completion was compared across the two systems. Only for the selection task did users take less time in Excel (since they directly selected the cells using the mouse) than in ISPM, by using natural language for end user software engineering, to overcome the present bottleneck of professional developers.
Markov Chain Based QoS Support for Wireless Body Area Network Communication in Health Monitoring Services
Wireless Body Area Networks (WBANs) are essential for real-time health monitoring of patients and in diagnosing of many diseases. WBANs comprise many sensors to monitor a large range of ambient conditions. Quality of Service (QoS) is a key challenge in WBAN, because the different state information of the neighboring nodes has to be monitored in an accurate manner. However, energy consumption gets increased while predicting and maintaining the exact information in highly dynamic environments. In order to reduce energy consumption and end to end delay, Markov Chain Based Quality of Service Support (MC-QoSS) method is designed in the health monitoring services of WBAN communication. The energy consumption gets reduced by forming a Markov chain with high energy nodes in the sensor networks communication path. The low energy level sensor nodes are removed using transitional probability in order to reduce end to end delay. High energy nodes are formed in the chain structure of its corresponding path to enhance communication. After choosing the communication path through high energy nodes, the packets are sent to the sink node from the source node with a higher Packet Delivery Ratio. The simulation result shows that MC-QoSS method improves the packet delivery ratio and reduces energy consumption with minimum end to end delay, compared to existing methods.
A Robust Eyelashes and Eyelid Detection in Transformation Invariant Iris Recognition: In Application with LRC Security System
Biometric authentication is an essential task for any
kind of real-life applications. In this paper, we contribute two
primary paradigms to Iris recognition such as Robust Eyelash
Detection (RED) using pathway kernels and hair curve fitting
synthesized model. Based on these two paradigms, rotation invariant
iris recognition is enhanced. In addition, the presented framework
is tested with real-life iris data to provide the authentication for
LRC (Learning Resource Center) users. Recognition performance
is significantly improved based on the contributed schemes by
evaluating real-life irises. Furthermore, the framework has been
implemented using Java programming language. Experiments are
performed based on 1250 diverse subjects in different angles of
variations on the authentication process. The results revealed that the
methodology can deploy in the process on LRC management system
and other security required applications.