Excellence in Research and Innovation for Humanity

International Science Index

Commenced in January 1999 Frequency: Monthly Edition: International Abstract Count: 50097

Computer and Systems Engineering

133
85662
A Unified Approach to Support the Coordination of Usability Work in Agile Software Development
Abstract:
Usability evaluation is essential for developing usable software systems, yet its integration within agile software development remains a challenging interdisciplinary endeavour. In this paper, the authors present a study to investigate obstacles of such integration from the management perspective. The study incorporates two methods, namely an online questionnaire survey and a series of interviews with participants that answered the questionnaire. Based on the obtained results, a unified approach is proposed for enabling coordinate the efforts of agile developers and usability engineers to produce usable software systems.
Digital Article Identifier (DAI):
132
83545
Seawater Changes' Estimation at Tidal Flat in Korean Peninsula Using Drone Stereo Images
Abstract:
Tidal flat in Korean peninsula is one of the largest biodiversity tidal flats in the world. Therefore, digital elevation models (DEM) is continuously demanded to monitor of the tidal flat. In this study, DEM of tidal flat, according to different times, was produced by means of the Drone and commercial software in order to measure seawater change during high tide at water-channel in tidal flat. To correct the produced DEMs of the tidal flat where is inaccessible to collect control points, the DEM matching method was applied by using the reference DEM instead of the survey. After the ortho-image was made from the corrected DEM, the land cover classified image was produced. The changes of seawater amount according to the times were analyzed by using the classified images and DEMs. As a result, it was confirmed that the amount of water rapidly increased as the time passed during high tide.
Digital Article Identifier (DAI):
131
83453
The Design Process of an Interactive Seat for Improving Workplace Productivity
Abstract:
Creative industries’ workers are becoming more prominent as countries move towards intellectual-based economies. Consequently, the nature and essence of the workplace needs to be reconfigured so that creativity and productivity can be better promoted at these spaces. Using a multidisciplinary approach and a user-centered methodology, combining product design, electronic engineering, software and human-computer interaction, we have designed and developed a new seat that uses embedded sensors and actuators to increase the overall well-being of its users, their productivity and their creativity. Our contribution focuses on the parameters that most affect the user’s work on these kinds of spaces, which are, according to our study, noise and temperature. We describe the design process for a new interactive seat targeted at improving workspace productivity.
Digital Article Identifier (DAI):
130
82893
Implementation of an IoT Sensor Data Collection and Analysis Library
Abstract:
Due to the development of information technology and wireless Internet technology, various data are being generated in various fields. These data are advantageous in that they provide real-time information to the users themselves. However, when the data are accumulated and analyzed, more various information can be extracted. In addition, development and dissemination of boards such as Arduino and Raspberry Pie have made it possible to easily test various sensors, and it is possible to collect sensor data directly by using database application tools such as MySQL. These directly collected data can be used for various research and can be useful as data for data mining. However, there are many difficulties in using the board to collect data, and there are many difficulties in using it when the user is not a computer programmer, or when using it for the first time. Even if data are collected, lack of expert knowledge or experience may cause difficulties in data analysis and visualization. In this paper, we aim to construct a library for sensor data collection and analysis to overcome these problems.
Digital Article Identifier (DAI):
129
82653
A POX Controller Module to Prepare a List of Flow Header Information Extracted from SDN Traffic
Abstract:
Software Defined Networking (SDN) is a paradigm designed to facilitate the way of controlling the network dynamically and with more agility. Network traffic is a set of flows, each of which contains a set of packets. In SDN, a matching process is performed on every packet coming to the network in the SDN switch. Only the headers of the new packets will be forwarded to the SDN controller. In terminology, the flow header fields are called tuples. Basically, these tuples are 5-tuple: the source and destination IP addresses, source and destination ports, and protocol number. This flow information is used to provide an overview of the network traffic. Our module is meant to extract this 5-tuple with the packets and flows numbers and show them as a list. Therefore, this list can be used as a first step in the way of detecting the DDoS attack. Thus, this module can be considered as the beginning stage of any flow-based DDoS detection method.
Digital Article Identifier (DAI):
128
82341
Automated Java Testing: JUnit versus AspectJ
Abstract:
Growing dependency of mankind on software technology increases the need for thorough testing of the software applications and automated testing techniques that support testing activities. We have outlined our testing strategy for performing various types of automated testing of Java applications using AspectJ which has become the de-facto standard for Aspect Oriented Programming (AOP). Likewise JUnit, a unit testing framework is the most popular Java testing tool. In this paper, we have evaluated our proposed AOP approach for automated testing and JUnit on various parameters. First we have provided the similarity between the two approaches and then we have done a detailed comparison of the two testing techniques on factors like lines of testing code, learning curve, testing of private members etc. We established that our AOP testing approach using AspectJ has got several advantages and is thus particularly more effective than JUnit.
Digital Article Identifier (DAI):
127
81960
Developing Co-Creation Monitoring Technique for Civic Tech
Abstract:
The main task of the paper is to offer a scientific evidence-based Co-creation Monitoring Technique for Civic Tech (civic technologies platforms) as a tool to evaluate, effectively manage and standardize digital supported co-creation processes, and multiply successful models of collective decision making and transparent management in other sectors. The co-creation concept fundamentally differs from traditional public engagement approach, while it focuses on the collective influence and responsibility of all stakeholders by creating the public good. While traditional approaches to public engagement and governmental reforms remain relevant, this paper focuses towards the growing potential of networked society to solve their social problems. It expands co-creation field to the citizens co-initiated, heavily technology supported, and systems oriented co-creation approaches. Around the world, the civic organizations, individual citizens, and even businesses experiment with the ICT tools and available open resources to collaborate with each other and with the government to find innovative solutions for societal problems. To support this, the international scientific society publishes the research results about the creative power of networked systems and their potential to grow 'collective intelligence.' The current research project relates co-creation and collective intelligence concepts and supplements the knowledge in the field of collective intelligence with the new aspects. The co-creation is defined as the new form of collective intelligence, which influences an internal and external motivation of the platforms' users to act for the public good. The both mentioned concepts were strongly influenced by technological progress but were developed in science parallel. The Civic Tech was investigated as collective intelligence systems, which integrate all criteria inherent for such kind of systems (openness, dynamism, decentralisation, critical mass for 'swarm effect', etc.). The challenging task for the proposed methodology was to correlate different factors and to find realizable possibilities for the system performance in these causal relationships. The Co-Creation Monitoring Technique evaluates the basic characteristics, functionality, and technological design of civic tech using a set of integral socio-technological indicators.
Digital Article Identifier (DAI):
126
81920
Point-of-Interest Recommender Systems for Location-Based Social Network Services
Abstract:
Location Based Social Network services (LBSNs) is a new term that combines location based service and social network service (SNS). Unlike traditional SNS, LBSNs emphasizes empirical elements in the user's actual physical location. Point-of-Interest (POI) is the most important factor to implement LBSNs recommendation system. POI information is the most popular spot in the area. In this study, we would like to recommend POI to users in a specific area through recommendation system using collaborative filtering. The process is as follows: first, we will use different data sets based on Seoul and New York to find interesting results on human behavior. Secondly, based on the location-based activity information obtained from the personalized LBSNs, we have devised a new rating that defines the user's preference for the area. Finally, we have developed an automated rating algorithm from massive raw data using distributed systems to reduce advertising costs of LBSNs.
Digital Article Identifier (DAI):
125
80276
Topographic Mapping of Farmland by Integration of Multiple Sensors on Board Low-Altitude Unmanned Aerial System
Abstract:
This paper introduced a topographic mapping system with time-saving and simplicity advantages based on integration of Light Detection and Ranging (LiDAR) data and Post Processing Kinematic Global Positioning System (PPK GPS) data. This topographic mapping system used a low-altitude Unmanned Aerial Vehicle (UAV) as a platform to conduct land survey in a low-cost, efficient, and totally autonomous manner. An experiment in a small-scale sugarcane farmland was conducted in Queensland, Australia. Subsequently, we synchronized LiDAR distance measurements that were corrected by using attitude information from gyroscope with PPK GPS coordinates for generation of precision topographic maps, which could be further utilized for such applications like precise land leveling and drainage management. The results indicated that LiDAR distance measurements and PPK GPS altitude reached good accuracy of less than 0.015 m.
Digital Article Identifier (DAI):
124
80157
From E-Government to Cloud-Government Challenges of Jordanian Citizens' Acceptance for Public Services
Abstract:
On the inception of the third millennium, there is much evidence that cloud technologies have become the strategic trend for many governments not only developed countries (e.g., UK, Japan, and USA), but also developing countries (e.g. Malaysia and the Middle East region), who have launched cloud computing movements for enhanced standardization of IT resources, cost reduction, and more efficient public services. Therefore, cloud-based e-government services considered as one of the high priorities for government agencies in Jordan. Although of their phenomenal evolution, government cloud-services still suffering from the adoption challenges of e-government initiatives (e.g. technological, human-aspects, social, and financial) which need to be considered carefully by governments contemplating its implementation. This paper presents a pilot study to investigate the citizens' perception of the extent in which these challenges affect the acceptance and use of cloud computing in Jordanian public sector. Based on the data analysis collected using online survey some important challenges were identified. The results can help to guide successful acceptance of cloud-based e-government services in Jordan.
Digital Article Identifier (DAI):
123
78014
Multilayer Neural Network and Fuzzy Logic Based Software Quality Prediction
Abstract:
In the software development lifecycle, the quality prediction techniques hold a prime importance in order to minimize future design errors and expensive maintenance. There are many techniques proposed by various researchers, but with the increasing complexity of the software lifecycle model, it is crucial to develop a flexible system which can cater for the factors which in result have an impact on the quality of the end product. These factors include properties of the software development process and the product along with its operation conditions. In this paper, a neural network (perceptron) based software quality prediction technique is proposed. Using this technique, the stakeholders can predict the quality of the resulting software during the early phases of the lifecycle saving time and resources on future elimination of design errors and costly maintenance. This technique can be brought into practical use using successful training.
Digital Article Identifier (DAI):
122
75889
D-Care: Diabetes Care Application to Enhance Diabetic Awareness to Diabetes in Indonesia
Abstract:
Diabetes is a common disease in Indonesia. One of the risk factors of diabetes is an unhealthy diet which is consuming food that contains too much glucose, one of glucose sources presents in food containing carbohydrate. The purpose of this study is to identify the amount of glucose level in the consumed food. The authors use literature studies for this research method. For the results of this study, the authors expect diabetics to be more aware of diabetes by applying daily dietary regulation through D-Care. D-Care is an application that can enhance people awareness to diabetes in Indonesia. D-Care provides two menus; there are nutrition calculation and healthy food. Nutrition calculation menu is used for knowing estimated glucose intake level by calculating food that consumed each day. Whereas healthy food menu, it provides a combination of healthy food menu for diabetic. The conclusion is D-Care is useful to be used for reducing diabetes prevalence in Indonesia.
Digital Article Identifier (DAI):
121
75514
Water End-Use Classification with Contemporaneous Water-Energy Data and Deep Learning Network
Abstract:
‘Water-related energy’ is energy use which is directly or indirectly influenced by changes to water use. Informatics applying a range of mathematical, statistical and rule-based approaches can be used to reveal important information on demand from the available data provided at second, minute or hourly intervals. This study aims to combine these two concepts to improve the current water end use disaggregation problem through applying a wide range of most advanced pattern recognition techniques to analyse the concurrent high-resolution water-energy consumption data. The obtained results have shown that recognition accuracies of all end-uses have significantly increased, especially for mechanised categories, including clothes washer, dishwasher and evaporative air cooler where over 95% of events were correctly classified.
Digital Article Identifier (DAI):
120
74866
Gradient Boosted Trees on Spark Platform for Supervised Learning in Health Care Big Data
Abstract:
Health care is one of the prominent industries that generate voluminous data thereby finding the need of machine learning techniques with big data solutions for efficient processing and prediction. Missing data, incomplete data, real time streaming data, sensitive data, privacy, heterogeneity are few of the common challenges to be addressed for efficient processing and mining of health care data. In comparison with other applications, accuracy and fast processing are of higher importance for health care applications as they are related to the human life directly. Though there are many machine learning techniques and big data solutions used for efficient processing and prediction in health care data, different techniques and different frameworks are proved to be effective for different applications largely depending on the characteristics of the datasets. In this paper, we present a framework that uses ensemble machine learning technique gradient boosted trees for data classification in health care big data. The framework is built on Spark platform which is fast in comparison with other traditional frameworks. Unlike other works that focus on a single technique, our work presents a comparison of six different machine learning techniques along with gradient boosted trees on datasets of different characteristics. Five benchmark health care datasets are considered for experimentation, and the results of different machine learning techniques are discussed in comparison with gradient boosted trees. The metric chosen for comparison is misclassification error rate and the run time of the algorithms. The goal of this paper is to i) Compare the performance of gradient boosted trees with other machine learning techniques in Spark platform specifically for health care big data and ii) Discuss the results from the experiments conducted on datasets of different characteristics thereby drawing inference and conclusion. The experimental results show that the accuracy is largely dependent on the characteristics of the datasets for other machine learning techniques whereas gradient boosting trees yields reasonably stable results in terms of accuracy without largely depending on the dataset characteristics.
Digital Article Identifier (DAI):
119
74464
Programming without Code: An Approach and Environment to Conditions-On-Data Programming
Abstract:
This paper presents the concept of an object-based programming language where tests (if... then... else) and control structures (while, repeat, for...) disappear and are replaced by conditions on data. According to the object paradigm, by using this concept, data are still embedded inside objects, as variable-value couples, but object methods are expressed into the form of logical propositions (‘conditions on data’ or COD).For instance : variable1 = value1 AND variable2 > value2 => variable3 = value3. Implementing this approach, a central inference engine turns and examines objects one after another, collecting all CODs of each object. CODs are considered as rules in a rule-based system: the left part of each proposition (left side of the ‘=>‘ sign) is the premise and the right part is the conclusion. So, premises are evaluated and conclusions are fired. Conclusions modify the variable-value couples of the object and the engine goes to examine the next object. The paper develops the principles of writing CODs instead of complex algorithms. Through samples, the paper also presents several hints for implementing a simple mechanism able to process this ‘COD language’. The proposed approach can be used within the context of simulation, process control, industrial systems validation, etc. By writing simple and rigorous conditions on data, instead of using classical and long-to-learn languages, engineers and specialists can easily simulate and validate the functioning of complex systems.
Digital Article Identifier (DAI):
118
73667
Foslip Loaded and CEA-Affimer Functionalised Silica Nanoparticles for Fluorescent Imaging of Colorectal Cancer Cells
Abstract:
Introduction: There is a need for real-time imaging of colorectal cancer (CRC) to allow tailored surgery to the disease stage. Fluorescence guided laparoscopic imaging of primary colorectal cancer and the draining lymphatics would potentially bring stratified surgery into clinical practice and realign future CRC management to the needs of patients. Fluorescent nanoparticles can offer many advantages in terms of intra-operative imaging and therapy (theranostic) in comparison with traditional soluble reagents. Nanoparticles can be functionalised with diverse reagents and then targeted to the correct tissue using an antibody or Affimer (artificial binding protein). We aimed to develop and test fluorescent silica nanoparticles and targeted against CRC using an anti-carcinoembryonic antigen (CEA) Affimer (Aff). Methods: Anti-CEA and control Myoglobin Affimer binders were subcloned into the expressing vector pET11 followed by transformation into BL21 Star™ (DE3) E.coli. The expression of Affimer binders was induced using 0.1 mM isopropyl β-D-1-thiogalactopyranoside (IPTG). Cells were harvested, lysed and purified using nickle chelating affinity chromatography. The photosensitiser Foslip (soluble analogue of 5,10,15,20-Tetra(m-hydroxyphenyl) chlorin) was incorporated into the core of silica nanoparticles using water-in-oil microemulsion technique. Anti-CEA or control Affs were conjugated to silica nanoparticles surface using sulfosuccinimidyl-4-(N-maleimidomethyl) cyclohexane-1-carboxylate (sulfo SMCC) chemical linker. Binding of CEA-Aff or control nanoparticles to colorectal cancer cells (LoVo, LS174T and HC116) was quantified in vitro using confocal microscopy. Results: The molecular weights of the obtained band of Affimers were ~12.5KDa while the diameter of functionalised silica nanoparticles was ~80nm. CEA-Affimer targeted nanoparticles demonstrated 9.4, 5.8 and 2.5 fold greater fluorescence than control in, LoVo, LS174T and HCT116 cells respectively (p < 0.002) for the single slice analysis. A similar pattern of successful CEA-targeted fluorescence was observed in the maximum image projection analysis, with CEA-targeted nanoparticles demonstrating 4.1, 2.9 and 2.4 fold greater fluorescence than control particles in LoVo, LS174T, and HCT116 cells respectively (p < 0.0002). There was no significant difference in fluorescence for CEA-Affimer vs. CEA-Antibody targeted nanoparticles. Conclusion: We are the first to demonstrate that Foslip-doped silica nanoparticles conjugated to anti-CEA Affimers via SMCC allowed tumour cell-specific fluorescent targeting in vitro, and had shown sufficient promise to justify testing in an animal model of colorectal cancer. CEA-Affimer appears to be a suitable targeting molecule to replace CEA-Antibody. Targeted silica nanoparticles loaded with Foslip photosensitiser is now being optimised to drive photodynamic killing, via reactive oxygen generation.
Digital Article Identifier (DAI):
117
73588
Clustering Categorical Data Using the K-Means Algorithm and the Attribute’s Relative Frequency
Abstract:
Clustering is a well known data mining technique used in pattern recognition and information retrieval. The initial dataset to be clustered can either contain categorical or numeric data. Each type of data has its own specific clustering algorithm. In this context, two algorithms are proposed: the k-means for clustering numeric datasets and the k-modes for categorical datasets. The main encountered problem in data mining applications is clustering categorical dataset so relevant in the datasets. One main issue to achieve the clustering process on categorical values is to transform the categorical attributes into numeric measures and directly apply the k-means algorithm instead the k-modes. In this paper, it is proposed to experiment an approach based on the previous issue by transforming the categorical values into numeric ones using the relative frequency of each modality in the attributes. The proposed approach is compared with a previously method based on transforming the categorical datasets into binary values. The scalability and accuracy of the two methods are experimented. The obtained results show that our proposed method outperforms the binary method in all cases.
Digital Article Identifier (DAI):
116
72645
Assembly Training: An Augmented Reality Approach Using Design Science Research
Abstract:
Augmented Reality (AR) is a strong growing research topic. This innovative technology is interesting for several training domains like education, medicine, military, sports and industrial use cases like assembly and maintenance tasks. AR can help to improve the efficiency, quality and transfer of training tasks. Due to these reasons, AR becomes more interesting for big companies and researchers because the industrial domain is still an unexplored field. This paper presents the research proposal of a PhD thesis which is done in cooperation with the BMW Group, aiming to explore head-mounted display (HMD) based training in industrial environments. We give a short introduction, describing the motivation, the underlying problems as well as the five formulated research questions we want to clarify along this thesis. We give a brief overview of the current assembly training in industrial environments and present some AR-based training approaches, including their research deficits. We use the Design Science Research (DSR) framework for this thesis and describe how we want to realize the seven guidelines, mandatory from the DSR. Furthermore, we describe each methodology which we use within that framework and present our approach in a comprehensive figure, representing the entire thesis.
Digital Article Identifier (DAI):
115
72016
A Computational Cost-Effective Clustering Algorithm in Multidimensional Space Using the Manhattan Metric: Application to the Global Terrorism Database
Abstract:
The increasing amount of collected data has limited the performance of the current analyzing algorithms. Thus, developing new cost-effective algorithms in terms of complexity, scalability, and accuracy raised significant interests. In this paper, a modified effective k-means based algorithm is developed and experimented. The new algorithm aims to reduce the computational load without significantly affecting the quality of the clusterings. The algorithm uses the City Block distance and a new stop criterion to guarantee the convergence. Conducted experiments on a real data set show its high performance when compared with the original k-means version.
Digital Article Identifier (DAI):
114
71998
Extraction of Forest Plantation Resources in Selected Forest of San Manuel, Pangasinan, Philippines Using LiDAR Data for Forest Status Assessment
Abstract:
Forest inventories are essential to assess the composition, structure and distribution of forest vegetation that can be used as baseline information for management decisions. Classical forest inventory is labor intensive and time-consuming and sometimes even dangerous. The use of Light Detection and Ranging (LiDAR) in forest inventory would improve and overcome these restrictions. This study was conducted to determine the possibility of using LiDAR derived data in extracting high accuracy forest biophysical parameters and as a non-destructive method for forest status analysis of San Manual, Pangasinan. Forest resources extraction was carried out using LAS tools, GIS, Envi and .bat scripts with the available LiDAR data. The process includes the generation of derivatives such as Digital Terrain Model (DTM), Canopy Height Model (CHM) and Canopy Cover Model (CCM) in .bat scripts followed by the generation of 17 composite bands to be used in the extraction of forest classification covers using ENVI 4.8 and GIS software. The Diameter in Breast Height (DBH), Above Ground Biomass (AGB) and Carbon Stock (CS) were estimated for each classified forest cover and Tree Count Extraction was carried out using GIS. Subsequently, field validation was conducted for accuracy assessment. Results showed that the forest of San Manuel has 73% Forest Cover, which is relatively much higher as compared to the 10% canopy cover requirement. On the extracted canopy height, 80% of the tree&rsquo;s height ranges from 12 m to 17 m. CS of the three forest covers based on the AGB were: 20819.59 kg/20x20 m for closed broadleaf, 8609.82 kg/20x20 m for broadleaf plantation and 15545.57 kg/20x20m for open broadleaf. Average tree counts for the tree forest plantation was 413 trees/ha. As such, the forest of San Manuel has high percent forest cover and high CS.
Digital Article Identifier (DAI):
113
71996
Landcover Mapping Using Lidar Data and Aerial Image and Soil Fertility Degradation Assessment for Rice Production Area in Quezon, Nueva Ecija, Philippines
Abstract:
Land-cover maps were important for many scientific, ecological and land management purposes and during the last decades, rapid decrease of soil fertility was observed to be due to land use practices such as rice cultivation. High-precision land-cover maps are not yet available in the area which is important in an economy management. To assure&nbsp;&nbsp; accurate mapping of land cover to provide information, remote sensing is a very suitable tool to carry out this task and automatic land use and cover detection. The study did not only provide high precision land cover maps but it also provides estimates of rice production area that had undergone chemical degradation due to fertility decline. Land-cover were delineated and classified into pre-defined classes to achieve proper detection features. After generation of Land-cover map, of high intensity of rice cultivation, soil fertility degradation assessment in rice production area due to fertility decline was created to assess the impact of soils used in agricultural production. Using Simple spatial analysis functions and ArcGIS, the Land-cover map of Municipality of Quezon in Nueva Ecija, Philippines was overlaid to the fertility decline maps from Land Degradation Assessment Philippines- Bureau of Soils and Water Management (LADA-Philippines-BSWM) to determine the area of rice crops that were most likely where nitrogen, phosphorus, zinc and sulfur deficiencies were induced by high dosage of urea and imbalance N:P fertilization. The result found out that 80.00 % of fallow and 99.81% of rice production area has high soil fertility decline.
Digital Article Identifier (DAI):
112
71469
A Neural Network Based Clustering Approach for Imputing Multivariate Values in Big Data
Abstract:
The treatment of incomplete data is an important step in the data pre-processing. Missing values creates a noisy environment in all applications and it is an unavoidable problem in big data management and analysis. Numerous techniques likes discarding rows with missing values, mean imputation, expectation maximization, neural networks with evolutionary algorithms or optimized techniques and hot deck imputation have been introduced by researchers for handling missing data. Among these, imputation techniques plays a positive role in filling missing values when it is necessary to use all records in the data and not to discard records with missing values. In this paper we propose a novel artificial neural network based clustering algorithm, Adaptive Resonance Theory-2(ART2) for imputation of missing values in mixed attribute data sets. The process of ART2 can recognize learned models fast and be adapted to new objects rapidly. It carries out model-based clustering by using competitive learning and self-steady mechanism in dynamic environment without supervision. The proposed approach not only imputes the missing values but also provides information about handling the outliers.
Digital Article Identifier (DAI):
111
70645
Road Accidents Bigdata Mining and Visualization Using Support Vector Machines
Abstract:
Useful information has been extracted from the road accident data in United Kingdom (UK), using data analytics method, for avoiding possible accidents in rural and urban areas. This analysis make use of several methodologies such as data integration, support vector machines (SVM), correlation machines and multinomial goodness. The entire datasets have been imported from the traffic department of UK with due permission. The information extracted from these huge datasets forms a basis for several predictions, which in turn avoid unnecessary memory lapses. Since data is expected to grow continuously over a period of time, this work primarily proposes a new framework model which can be trained and adapt itself to new data and make accurate predictions. This work also throws some light on use of SVM&rsquo;s methodology for text classifiers from the obtained traffic data. Finally, it emphasizes the uniqueness and adaptability of SVMs methodology appropriate for this kind of research work.
Digital Article Identifier (DAI):
110
68301
Usability and Biometric Authentication of Electronic Voting System
Abstract:
In this paper, a new voting system is developed and its usability is evaluated. The main feature of this system is the biometric verification of the voter and then a few easy steps to cast a vote. As compared to existing systems available, e.g dual vote, the new system requires no training in advance. The security is achieved via multiple key concept (another part of this project). More than 100 student voters were participated in the election from University of Malakanad, Chakdara, PK. To achieve the reliability, the voters cast their votes in two ways, i.e. paper based and electronic based voting using our new system. The results of paper based and electronic voting system are compared and it is concluded that the voters cast their votes for the intended candidates on the electronic voting system. The voters were requested to fill a questionnaire and the results of the questionnaire are carefully analyzed. The results show that the new system proposed in this paper is more secure and usable than other systems.
Digital Article Identifier (DAI):
109
67755
Assessing the Spatial Distribution of Urban Parks Using Remote Sensing and Geographic Information Systems Techniques
Abstract:
Urban parks and open spaces play a significant role in improving physical and mental health of the citizens, strengthen the societies and make the cities more attractive places to live and work. As the world’s cities continue to grow, continuing to value green space in cities is vital but is also a challenge, particularly in developing countries where there is pressure for space, resources, and development. Offering equal opportunity of accessibility to parks is one of the important issues of park distribution. The distribution of parks should allow all inhabitants to have close proximity to their residence. Remote sensing and Geographic information systems (GIS) can provide decision makers with enormous opportunities to improve the planning and management of Park facilities. This study exhibits the capability of GIS and RS techniques to provide baseline knowledge about the distribution of parks, level of accessibility and to help in identification of potential areas for such facilities. For this purpose Landsat OLI imagery for year 2016 was acquired from USGS Earth Explorer. Preprocessing models were applied using Erdas Imagine 2014v for the atmospheric correction and NDVI model was developed and applied to quantify the land use/land cover classes including built up, barren land, water, and vegetation. The parks amongst total public green spaces were selected based on their signature in remote sensing image and distribution. Percentages of total green and parks green were calculated for each town of Lahore City and results were then synchronized with the recommended standards. ANGSt model was applied to calculate the accessibility from parks. Service area analysis was performed using Network Analyst tool. Serviceability of these parks has been evaluated by employing statistical indices like service area, service population and park area per capita. Findings of the study may contribute in helping the town planners for understanding the distribution of parks, demands for new parks and potential areas which are deprived of parks. The purpose of present study is to provide necessary information to planners, policy makers and scientific researchers in the process of decision making for the management and improvement of urban parks.
Digital Article Identifier (DAI):
108
66941
Spatio-Temporal Variation of Suspended Sediment Concentration in the near Shore Waters, Southern Karnataka, India
Abstract:
Suspended Sediment Concentration (SSC) was estimated for the period of four months (November, 2013 to February 2014) using Oceansat-2 (Ocean Colour Monitor) satellite images to understand the coastal dynamics and regional sediment transport, especially distribution and budgeting in coastal waters. The coastal zone undergoes continuous changes due to natural processes and anthropogenic activities. The importance of the coastal zone, with respect to safety, ecology, economy and recreation, demands a management strategy in which each of these aspects is taken into account. Monitoring and understanding the sediment dynamics and suspended sediment transport is an important issue for coastal engineering related activities. A study of the transport mechanism of suspended sediments in the near shore environment is essential not only to safeguard marine installations or navigational channels, but also for the coastal structure design, environmental protection and disaster reduction. Such studies also help in assessment of pollutants and other biological activities in the region. An accurate description of the sediment transport, caused by waves and tidal or wave-induced currents, is of great importance in predicting coastal morphological changes. Satellite-derived SSC data have been found to be useful for Indian coasts because of their high spatial (360 m), spectral and temporal resolutions. The present paper outlines the applications of state‐of‐the‐art operational Indian Remote Sensing satellite, Oceansat-2 to study the dynamics of sediment transport.
Digital Article Identifier (DAI):
107
66927
Evaluate the Possibility of Using ArcGIS Basemaps as GCP for Large Scale Maps
Abstract:
Awareness of the importance large-scale maps for development of a country is growing in all walks of life, especially for governments in Indonesia. Various parties, especially local governments throughout Indonesia demanded for immediate availability the large-scale maps of 1:5000 for regional development. But in fact, the large-scale maps of 1:5000 is only available less than 5% of the entire territory of Indonesia. Unavailability precise GCP at the entire territory of Indonesia is one of causes of slow availability the large scale maps of 1:5000. This research was conducted to find an alternative solution to this problem. This study was conducted to assess the accuracy of ArcGIS base maps coordinate when it shall be used as GCP for creating a map scale of 1:5000. The study was conducted by comparing the GCP coordinate from Field survey using GPS Geodetic than the coordinate from ArcGIS basemaps in various locations in Indonesia. Some areas are used as a study area are Lombok Island, Kupang City, Surabaya City and Kediri District. The differences value of the coordinates serve as the basis for assessing the accuracy of ArcGIS basemaps coordinates. The results of the study at various study area show the variation of the amount of the coordinates value given. Differences coordinate value in the range of millimeters (mm) to meters (m) in the entire study area. This is shown the inconsistency quality of ArcGIS base maps coordinates. This inconsistency shows that the coordinate value from ArcGIS Basemaps is careless. The Careless coordinate from ArcGIS Basemaps indicates that its cannot be used as GCP for large-scale mapping on the entire territory of Indonesia.
Digital Article Identifier (DAI):
106
66737
Use of Personal Rhythm to Authenticate Encrypted Messages
Abstract:
When communicating using private and secure keys, there is always the doubt as to the identity of the message creator. We introduce an algorithm that uses the personal typing rhythm (keystroke dynamics) of the message originator to increase the trust of the authenticity of the message originator by the message recipient. The methodology proposes the use of a Rhythm Certificate Authority (RCA) to validate rhythm information. An illustrative example of the communication between Bob and Alice and the RCA is included. An algorithm of how to communicate with the RCA is presented. This RCA can be an independent authority or an enhanced Certificate Authority like the one used in public key infrastructure (PKI).
Digital Article Identifier (DAI):
105
66615
Installing Cloud Computing Model for E-Businesses in Small Organizations
Authors:
Abstract:
Information technology developments have changed the way how businesses are working. Organizations are required to become visible online and stay connected to take advantages of costs reduction and improved operation of existing resources. The approval and the application areas of the cloud computing has significantly increased since it was presented by Google in 2007. Internet Cloud computing has attracted the IT enterprise attention especially the e-business enterprise. At this time, there is a great issue of environmental costs during the enterprises apply the e- business, but with the coming of cloud computing, most of the problem will be solved. Organizations around the world are facing with the continued budget challenges and increasing in the size of their computational data so, they need to find a way to deliver their services to clients as economically as possible without negotiating the achievement of anticipated outcomes. E- business companies need to provide better services to satisfy their clients. In this research, the researcher proposed a paradigm that use and deploy cloud computing technology environment to be used for e-business in small enterprises. Cloud computing might be a suitable model for implementing e-business and e-commerce architecture to improve efficiency and user satisfaction.
Digital Article Identifier (DAI):
104
66491
Learning Dynamic Representations of Nodes in Temporally Variant Graphs
Abstract:
In many industries, including telecommunications, churn prediction has been a topic of active research. A lot of attention has been drawn on devising the most informative features, and this area of research has gained even more focus with spread of (social) network analytics. The call detail records (CDRs) have been used to construct customer networks and extract potentially useful features. However, to the best of our knowledge, no studies including network features have yet proposed a generic way of representing network information. Instead, ad-hoc and dataset dependent solutions have been suggested. In this work, we build upon a recently presented method (node2vec) to obtain representations for nodes in observed network. The proposed approach is generic and applicable to any network and domain. Unlike node2vec, which assumes a static network, we consider a dynamic and time-evolving network. To account for this, we propose an approach that constructs the feature representation of each node by generating its node2vec representations at different timestamps, concatenating them and finally compressing using an auto-encoder-like method in order to retain reasonably long and informative feature vectors. We test the proposed method on churn prediction task in telco domain. To predict churners at timestamp ts+1, we construct training and testing datasets consisting of feature vectors from time intervals [t1, ts-1] and [t2, ts] respectively, and use traditional supervised classification models like SVM and Logistic Regression. Observed results show the effectiveness of proposed approach as compared to ad-hoc feature selection based approaches and static node2vec.
Digital Article Identifier (DAI):