Other Volumes


2007 2008 2009 2010 2012



Editorial Board

Prof. Metin Demiralp (Turkey),

Prof. A.Kuri-Morales (Mexico),

Prof. V.Mladenov (Bulgaria),

Prof. G.Bognar (Hyngary),

Prof. Lotfi A. Zadeh (USA),

Prof. Leonid Kazovsky (USA),

Prof. T.Kaczorek (Poland),

Prof. L.Chua (USA),

Prof. O.Martin (Romania),

Prof. C.Udriste (Romania),

Prof. N.Mastorakis (Greece),

Prof. D.Bertsekas (USA),

Prof. R.Yager (USA),

Prof. Anping Xu (China),  

Prof. M. A. Breuer (USA),

Prof. M.Wasfy (USA)



Topics

All aspects of classic and modern applied mathematics are covered including algebra, differential equations, probability, statistics, operational research, optimization, algorthms theory, computational complexity, control . It appears quarterly.  Special Issues are specially encouraged.

Format

Download Format... 

 

Contact us

Issue 1, Volume 5, 2011

Consumer Profile and Tipping Habits. A Romanian Framework Using Fuzzy Method

by Ada Mirela Tomescu, Olimpia I. Ban

Abstract: Tipping is an interesting consumer behavior, because the tips are voluntary payments given after services are completed. Our paper tries to establish the Romanian sample profile and their perception regarding tipping phenomena. The Romanian consumer is difficult to investigate because of resistance to being fully honest in this problem. Thus the present study aim is to identify whether there are differences between respondents` attitude toward idea of tipping in various units, and whether these attitudes are influenced by respondent incomes, this last objective it was accomplished using Fuzzy method. A thesis could be: a higher income leads to a more positive attitude towards the idea of giving tips; another is to record the profile and the attitude to this issue of a certain sample of Romanian consumers.
Keywords:
Fuzzy method, Interaction, Romanian consumer, Tips
Full Paper, pp. 1-8

 

Parametric Modeling for Advanced Architecture

by M. Stavric, O. Marina

Abstract: Computer simulation of evolutionary processes is already well established technique for the study of environmental, biological and economic dynamics. Use of algorithms for generation of virtual entities that will develop its functional and formal properties within the non-linear process of adaptation of complex system is a foundation for new point of view in understanding architecture and urban environment. This paper describes parametric approach in architectural design through elaboration of shift in paradigms in architecture that has brought to the idea of use of parametric modeling with emphasis on two different groups of parametric software and presents the possibilities of generative algorithms in modeling architectural form and development of cities and urban design.
Keywords:
Geometry, parametric design, generative algorithms urbanism, architectural design
Full Paper, pp. 9-16

 

Assessment of Public Perception toward Mobile Tagging System for Retrieving Information of Genetically Modified Food

by Tan Shiang-Yen, Long Yoon Foo, Rosnah Idrus

Abstract: The purpose of this paper is to assess the public's perception and acceptance toward the proposed mobile tagging system, which can be used by consumers to retrieve product information of Genetically Modified Food (GMF) in the market. In the proposed mobile tagging system, consumers can use the camera on their mobile phone to capture the image of the Quick Response (QR) codes on the product packaging and send to the server for decoding. Subsequently, the server sends back the details of the GMF product in the form of Multimedia Message Service (MMS). Findings of this study indicate that albert not much the public is currently using mobile tagging service, but they possess positive attitude to find out more and to try the mobile tagging services. Most importantly, acceptance of the proposed mobile tagging service for retrieving product information of genetically modified food is considerably high. Additionally, the finding suggests that marketing campaign is vital to increase awareness among the consumers, and customer education is important to improve the acceptance level toward the mobile tagging service. This study is one of the pioneer studies in mobile tagging system. It provides the feasibility analysis of applying mobile tagging system in genetically modified food area and assesses the marketability of the system.
Keywords:
2D Codes, Genetically Modified Food, Mobile Tagging, Quick Response (QR) Code
Full Paper, pp. 17-25

 

Issue 3, Volume 5, 2011

Statistical Causality and Stochastic Dynamic Systems

by Ljiljana Petrovic

Abstract: In this paper we consider a problem (that follows directly from realization problem): how to find a possible states (even minimal) of a stochastic dynamic system S1 with known outputs, provided it is in a certain causality relationship with another stochastic dynamic system S2 whose states (or some information about them) are given. This paper is continuation of the papers (Gill and Petrovi´c 1987), (Petrovi´c 1996) and (Petrovi´c 2005).
Keywords:
Hilbert space, causality, stochastic dynamic system, realization.
Full Paper, pp. 153-156

 

The HOSVD Based Domain and the Related Image Processing Techniques

by Andras Rovid, Laszlo Szeidl, Peter Varlaki

Abstract: In the framework of the paper an improved method for image resolution enhancement is introduced. The enhancement is performed through another representation domain, where the image function is expressed with the help of polylinear functions on higher order singular value decomposition (HOSVD) basis. The paper gives a detailed description on how to determine the polylinear functions corresponding to an image and how to process them in order to obtain a higher resolution image. Furthermore, the proposed approach and the Fourier-based approach will be compared from the point of view of their effectiveness.
Keywords:
Image resolution, HOSVD, approximation, numerical reconstruction
Full Paper, pp. 157-164

 

Achieving the Efficiency for Information Systems - Possibilities and Prospects

by D. Litan, M. Velicanu, L. Copcea Teohari, M. Teohari, A. M. Mocanu Virgolici, I. Surugiu, O. Raduta

Abstract: Nowadays companies cannot cope with competition successfully unless they have performing information systems. Therefore, the need for efficient information systems is very high. The efficiency can be achieved by using various methods/ techniques and technologies “to build” the information system. In this paper we are going to present and analyze to what extent the new technologies and methods/techniques: methods of integrating the information systems; techniques of databases optimization; “special” types of databases (mobile, multimedia and spatial databases); data warehouses, have increased or could increase the efficiency of information systems.
Keywords:
Database optimization, data warehouse, Enterprise Application Integration, mobile databases, multimedia databases, spatial databases, Oracle Warehouse Builder
Full Paper, pp. 165-179

 

Nonlocal Symmetries for a Family Benjamin-Bona-Mahony-Burgers Equations. Some Exact Solutions

by M. S. Bruzon, M. L. Gandarias

Abstract: In this work the nonlocal symmetries of a family Benjamin-Bona-Mahony-Burgers equations are studied. The partial differential equation written as a conservation law can be transformed into an equivalent system by introducing a suitable potential. The nonlocal symmetry group generators of original partial differential equation can be obtained through their equivalent system. We have proved that the nonclassical method applied to this system leads to new symmetries, which are not solutions arising from potential symmetries of the Benjamin- Bona-Mahony-Burgers equations. We also have derived traveling wave solutions for the Benjamin-Bona-Mahony-Burgers equations by using a direct method. Among them we nd a solution which describes a kink solution.
Keywords:
Nonlocal symmetries, Partial differential equations, Nonlinear equations, MAXIMA Software, kink solutions
Full Paper, pp. 180-187

 

Capacity Recovery of Very Noisy Optical Quantum Channels

by Laszlo Gyongyosi, Sandor Imre

Abstract: The number of efficient approximation algorithms for quantum informational distances is very small, because of the special properties of quantum informational generator functions and of asymmetric quantum informational distances. If we wish to analyze the properties of quantum channels using today’s classical computer architectures, an extremely efficient algorithm is needed. The capacity recovery of very noisy communication channels cannot be imagined for classical systems, and this effect has no analogue in classical systems. The capacity recovery of very noisy quantum channels makes it possible to use two very noisy optical-fiber based quantum channels with a positive joint capacity at the output. Here we define a fundamentally new approach of capacity recovery of very noisy, practically completely useless optical quantum channels. We show an algorithmic solution to the capacity recovery problem, and provide an efficient algorithmic solution for finding the set of recoverable very noisy optical quantum channels. The calculations are based on the asymptotic classical capacity of the quantum channel.
Keywords:
Capacity Recovery, Noisy Optical Communications, Quantum Channels, Quantum Communications
Full Paper, pp. 188-199

 

Software Quality Evaluation: User's View

by Anas Bassam AL-Badareen, Mohd Hasan Selamat, Jamilah Din, Marzanah A. Jabar, Sherzod Turaev

Abstract: Nowadays, software products rapidly increased and it is usage not limited to specific people or corporation. It is now used in the most of human life activities. Therefore, the quality of the software product is increasingly being important and the users demanding higher quality than ever before. However, most of this research is focused on the internal/development view of quality. Hence, in software development, strong attention must be given to the user’s satisfaction. Recently, the studies intend to understand the users’ perspective of the software quality. In this study, we intend to discuss the characteristics of the software products that influence the users’ satisfaction on software quality. Based on the well-known software quality models and the emotion of the software users, a model of software quality evaluation based on users’ views is proposed.
Keywords:
Software Quality, Quality Model, Software Evaluation, User Perspective
Full Paper, pp. 200-207

 

Effectiveness of Stemming and n-Grams String Similarity Matching on Malay Documents

by Tengku Mohd T. Sembok, Zainab Abu Bakar

Abstract: There are two main classes of conflation algorithms, namely, string-similarity algorithms and stemming algorithms. String-similarity matching algorithms, bi-grams and tri-grams, are used in the experiments conducted on Malay texts. Malay stemming algorithms used in the experiments is developed by Fatimah et al. Inherent characteristics of n-grams on Malay documents are discussed in this paper. Retrieval effectiveness experiments using several variations of combinations between n-grams and stemming algorithms are performed in order to find the best combination. The variations experimented are: both nonstemmed queries and documents; stemmed queries and nonstemmed documents; and both stemmed queries and documents. Further experiments are then carried out by removing the most frequently occurring n-grams. Besides using dice coefficients to rank documents, inverse document frequency (idf) weights are also used. Interpolation technique and standard recall-precision functions are used to calculate recall-precision values. It is found that using combined search, n-gram matching and stemming, improves retrieval effectiveness. Removing the most frequently occurring n-gram that appears in about 46% of the words also improve the retrieval effectiveness.
Keywords:
Information retrieval, string similarity matching, stemming algorithms
Full Paper, pp. 208-215

 

Effective Algorithms for Improving the Performance of Search Engine Results

by G. Poonkuzhali, R. Kishore Kumar, R. Kripa Keshav, K.Thiagarajan, K.Sarukesi

Abstract: Search engine has become an important tool in todayΤs world for searching various data like text, audio, video and images. While searching for information many users end up with irrelevant information causing a waste in user time and accessing time of the search engine. So to narrow down this problem, many researchers are involved in web mining, still the algorithms developed by them contains nearly 30% of outlaid content such as irrelevant and redundant information. In this paper, two Statistical approaches based on Proportions (Z-test hypothesis) and chi square test (T-test) is developed for mining this outlaid content. Also comparative studies between these two methods are presented. Elimination of this outlaid content during a searching process improves the quality of search engines further. The results show that the system easily provides relevancies and delivers dominant text extraction, supporting users in their query to efficiently examine and make the most of available web data sources. Experimental results revealed that statistical approach produces better results than chi square test.
Keywords:
Chi Square Test, critical value, degrees of confidence, relevant, test statistic, web document
Full Paper, pp. 216-223

 

Sharing Behaviors in Games and Social Media

by Harri Ketamo

Abstract: User generated media highlights sharing: Sharing videos, images and texts in social media, as well as sharing character outfits and maps in games. However, behavior is one of the aspects that are not shared. The aim of this study is to show how user generated behaviors can be shared in different types of games and social media applications. The examples about sport games and educational games are based on MindFarm AI technology that enables end users to construct human-like behavior by only teaching. MindFarm emulates the human way to learn: According to cognitive psychology of learning, our thinking is based on conceptual representations of our experiences and complex relations between these concepts and experiences. Phenomena when the mental structure change is called learning. In the near future user-generated behaviors can be developed and shared, as all other user-generated content. According to examples on this study, game- and media developers can design extensions that enable users to easily construct behaviors. From a game consumer point of view, the most interesting part is in developing behaviors, sharing them and finally playing with them, or against them.
Keywords:
Artificial Intelligence, Games, Learning, Social Media, User Generated Content
Full Paper, pp. 224-232

 

Deformations of Monic Polynomial Matrices. Analysis of Perturbation of Eigenvalues

by M. Isabel Garcia-Planas, Sonia Tarragona

Abstract: In this work we study behavior of a multiple eigenvalue of the monic polynomial family P(λ) as well as we study of behavior of a simple eigenvalue of a family of 1-degree singular
polynomial matrices representing families of singular linear systems.
Keywords:
Polynomial matrix, Eigenvalues, Perturbation
Full Paper, pp. 233-239

 

Monocriteria and Multicriteria based Placement of Reactive Power Sources in Distribution Systems

by Whester J. Araujo, Petr Ya. Ekel, Rafael P. Falcao Filho, Illya Kokshenev, Henrique S. Schuffner

Abstract: This paper presents results of research on formulating and solving the problem of capacitor placement in distribution systems within the framework of monocriteria and multicriteia models. The application of the multicriteria approach is aimed, first of all, at overcoming the difficulties of simultaneous observation of contradictory constraints for the upper and lower permissible voltage limits at different buses of distribution systems as well as other important conditions in operating capacitors. The solution of the capacitor placement problem in the monocriteria statement is based on applying the generalized algorithms of discrete optimization. The solution of the problem in the multicriteria statement is based on combining the Bellman-Zadeh approach to decision making in a fuzzy environment with the application of the generalized algorithms of discrete optimization. The paper results are illustrated by computational experiments with a real distribution system.
Keywords:
Discrete optimization, Distribution systems, Muticriteria decision making, Reactive power compensation
Full Paper, pp. 240-248

 

Integrated Study of Automated Checkout Shopping Trolley

by Koh Peik See, Tan Shiang-Yen, Nasuha Lee Abdullah, Rosnah Idrus

Abstract: Recently, Automated checkout shopping trolley technologies have taken an advance step than the previous introduced self-checkout system in terms of greater convenience, ease-of-use and greater efficiency. As this technology is considerable new in the market of developing countries, hence it is vital to explore the perception of consumers toward the use of automated checkout trolley. The purpose of this paper is to examine the perceptions and expectations of both enterprise users and end users toward the proposed checkout system for hypermarket in the context of Malaysia, as a typical paradigm of developing countries. Findings from this study indicate that both consumers and retailers possess positive attitude toward the automated checkout trolley system, and are enthusiastic to try this emerging technology. The main concerns of the users are the secureness of payment and the variability of payment methods, which forms their core criteria in determining whether to adopt the system. This study provides an understanding about the practicality and market condition of automated trolley system in the developing countries.
Keywords:
Technology-based self-service (TBSS), Automated checkout, Market survey, Technology and innovation management
Full Paper, pp. 249-257

 

A Prediction Method for Nonlinear Time Series Analysis by Combining the False Nearest Neighbors and Subspace Identification Methods

by I. Marin Carrion, E. Arias Antunez, M. M. Artigao Castillo, J. J. Miralles Canals

Abstract: Nonlinear time series analysis is a powerful methodology that permits to predict the temporal evolution of some kinds of dynamical systems from characteristic quantities, such as the minimal embedding dimension or the maximal Lyapunov exponent. In fact, one of the most important goals of nonlinear analysis of experimental time series is the prediction. The subspace identification methods provides a good framework to model a system, both deterministic and stochastic, in a easily way. In order to make predictions, we propose a method which combine the minimal embedding dimension obtained by the method of false nearest neighbors and a model estimated by means of a subspace identification method. The results, in terms of predicted error, show the reliability of this new approach.
Keywords:
Nonlinear Time Series Analysis, False Nearest Neighbors Method, Subspace Identification Methods, Prediction
Full Paper, pp. 258-265

 

The Quickest Maximum Dynamic Flow of Minimum Cost

by Mircea Parpalea, Eleonor Ciurea

Abstract: This article states and solves the multi-criteria maximum flow problem in discrete dynamic networks for the case of two objective functions. This represents a generalisation of the maximum flow of minimum cost problem for the case of minimizing the travelling cost (minimum cost flow) and travelling time (quickest flow). The approach is actually based on generating efficient extreme points in the objective space by iteratively solving a series of maximum flow problems with different single objective functions. Each time, the dynamic flow is augmented along a cheapest (minimum cost) path from the source node to the sink node in the time-space network avoiding the explicit time expansion of the network.
Keywords:
Bi-criteria flow, Discrete dynamic network, Dynamic maximum flow, Successive shortest path
Full Paper, pp. 266-274

 

 

 

 

 

Issue 2, Volume 5, 2011

Time Varying Causality between Stock Exchanges in the CEE Region

by Cristiana Tudor

Abstract: This paper analyzes the evolution of linkages and causality between six Central and Eastern European (Bulgaria, Czech Republic, Hungary, Poland, Romania, Russia) and the USA stock exchanges. The effects of the recent global financial crisis and possible changes in inter-market relationships as a result of the crisis are topics of special interest in this research. For this purpose, the panel data sample (daily observations for the 2006-2009 time period) is divided into two sub-periods corresponding to the pre-crisis and crisis period. In order to separate the two sub-periods, a secondary investigation is conducted which shows that the beginning of July 2007 is the moment when the global financial crisis began to show its full manifestations on international stock exchanges. The study concludes that stock markets in the CEE region have became increasingly integrated during crisis, while before the crisis the markets appear to be segmented, as both contemporaneous correlations and causality relationships are mostly insignificant. Also, before the crisis CEE markets were significantly influenced by innovations in the USA market, thus explaining why they were affected heavily by the crisis, which has managed to spread immediately in the region. As far as the risk-adjusted performance is concerned, the Czech market realized the best risk-adjusted performance due to its average rate of return and low risk, s followed by the Romanian, Russian and Polish stock markets, while the lowest risk-adjusted performance, as represented by the coefficient of variation was found in the case of the US stock market.
Keywords:
Granger causality, contemporaneous correlation, risk-adjusted performance, Central and Eastern European Stock exchanges
Full Paper, pp. 27-34

 

Market Research – A Significant Step in Propelling a Business or a Solution for the Egovernment Systems to the Top

by D. Litan, A. Mocanu, S. Olaru, A. Apostu

Abstract: Market research... a real treasure which can trigger performance in any business, but most of the times it seems unexploited at its true value. The marketing analysis brings information about competitors, clients, business partners and so on, guiding the business towards success. Because during the last years the information technologies have been applied in so many aspects of our lives, they have also been used for marketing research. The most used marketing research method is the questionnaire, that is why many of the well known companies from this field have left the “real” environment and they are focusing now to the “online” environment. Therefore, today one can not do a marketing analysis without specialized IT tools (for the online surveys or for data analysis). In the current paper, we search and offer some answers regarding the information technologies used, or which could be used, for developing and implementing information systems for marketing research. In the same time, on the one hand, we are opening new perspectives and propose new ways of considering the marketing research: used to design an e-government system about the citizens needs, expectations or preferences; and on the other hand: online survey systems “seen” like a first step for a large scale usage of evoting concept.
Keywords:
E-government, geographical information system, information security, information technology, market research
Full Paper, pp. 35-45

 

The Image Recognition of Brain-Stem Ultrasound Images with using a Neural Network based on PCA

by Jiri Blahuta, Tomas Soukup, Petr Cermak

Abstract: This paper shows how to solve the recognition of ultrasound brain-stem images. Our work is based on PCA method that is very useful and known method for image processing. We will demonstrate a solving with C application, Gnumeric spreadsheet as a standard solution. Also we will use an artificial neural networks (ANN) for this problem and we will compare a results. The ANN are generally very usable for image processing. It has been demonstrated with NeuroSolutions software that is very sophisticated simulator of ANN with PCA multilayer (ML) NN topology.
Keywords:
PCA, ultrasound, neural network, image processing
Full Paper, pp. 46-54

 

Quality of Life for People with Disabilities - the Study Realized on the NUTS3 Regional Level in the Czech Republic

by Pavel Jirava, Jiri Krupka, Miloslava Kasparova , Jan Mandys

Abstract: This article is focused on the problem of the data acquisition and analysis of data about quality of life of hearing disabled people in the region. Data were obtained from disabled citizens of Pardubice region, Czech Republic. To obtain the data was used a questionnaire survey. For processing the obtained data were used the statistical methods, selected visualization techniques and data mining method. The achievements of this research are relations and links influencing quality of life of disabled people and basic characteristics of hearing disabled people. This is the basis for future social policy making in the Pardubice region. The methods used in solving the problem can be implemented also in other regions (in the level Nomenclature of Units for Territorial Statistics 3).
Keywords:
Quality of life, people with disabilities, data mining, cluster analysis, regional development, sustainable development, questionnaire survey.
Full Paper, pp. 55-62

 

Analysis of Complexity of One-dimensional Chaotic Maps with Entropy Characteristic

by Sen Pei, Zhiming Zheng

Abstract: Chaotic phenomena are widespread in a variety of research fields. In this paper, we investigate the complexity of one-dimensional chaotic maps with entropy characteristic. First we analyze the dynamics of one-dimensional chaotic maps. Then we analogize the iteration of maps to a one-sided shift transformation, and give a probability explanation of Lyapunov exponent by Ergodic Theory. We use the entropy of corresponding shift to reflect the complexity of these maps, providing a new way to characterize complexity of chaotic systems.
Keywords:
Chaos, Complexity, Entropy, Ergodicity, Lyapunov Exponent
Full Paper, pp. 63-70

 

AES Algorithm Adapted on GPU Using CUDA for Small Data and Large Data Volume Encryption

by Tomoiaga Radu Daniel, Stratulat Mircea

Abstract: The goal of this paper is to study the possibility of using alternative computing solution in cryptography, the use of a graphic processing unit in non graphical calculations. We tried to use the graphic processing unit as a cryptographic coprocessor in order to obtain more computing power and better runtime for AES. In this paper we present an implementation of AES on NVIDIA GPU using CUDA. The results of our tests show that the CUDA implementation can offer speedups of almost 40 times in comparison to the CPU. The tests are conducted in two directions: running the tests on small amount of data that is located in memory and a big amount of data that are stored in files on hard drives and as news, the access time for the information on the hard disk is added to the encrypting time.
Keywords:
Benchmark, AES, CUDA, GPU, Cryptography
Full Paper, pp. 71-81

 

Derivations and Iterated Skew Polynomial Rings

by Michael Gr. Voskoglou

Abstract: Two are the objectives of the present paper. First we study properties of a differentially simple commutative ring R with respect to a set D of derivations of R. Among the others we investigate the relation between the D-simplicity of R and that of the local ring RP with respect to a prime ideal P of R and we prove a criterion about the D- simplicity of R in case where R is a 1-dimensional (Krull dimension) finitely generated algebra over a field of characteristic zero and D is a singleton set. The above criterion was quoted without proof in an earlier paper of the author. Second we construct a special class of iterated skew polynomial rings defined with respect to finite sets of derivations of a ring R (not necessarily commutative) commuting to each other. The important thing in this class is that, if R is a commutative ring, then its differential simplicity is the necessary and sufficient condition for the simplicity of the corresponding skew polynomial ring.
Keywords:
Derivations, Differentially simple rings, Finitely-generated algebras, Iterated skew polynomial rings, Simple rings
Full Paper, pp. 82-90

 

Computational and Mathematical Modeling of an Industrial-Automobile Robot: A Multi-Purpose Case of Study

by J. Alejandro Betancur

Abstract: Nowadays, in automobile industry are found many working situations, in which the use of an articulated arm that supports the activities developed by operators is inevitable; out side the assembly line of a car, activities such as searching for parts needed to produce a vehicle, handling the cars tires, among others, are performed manually in many different companies. In this work is exposed the design of a robot that helps the operator to develop those activities, which is determined by a mechanical arm, made up by four (4) segments and four (4) joints, located on a mobile platform with two (2) wheels, each controlled by an independent engine. Through this work is intended to cover mathematics and physics present in the above mentioned device, taking advantage of the computational power symbolic of the software MAPLE. Although this development is focused on the automotive sector, applications of this kind of device in other fields of industry are innumerable.
Keywords:
Arm, Automobile, Denavit, Electromechanical, Euler, Hartenberg, Langrange
Full Paper, pp. 91-100

 

Considerations Regarding Formal Languages Generation Using Labelled Stratified Graphs

by Tudor (Preda) Irina-Valentina

Abstract: The concept of labelled stratified graph was introduced in order to obtain the concept of knowledge base with output and various applications of this concept were presented ever since. This paper studies another application of this structure: generating formal languages by means of labelled stratified graphs. Various mechanisms to define and generate formal languages are known and we show that we can obtain different types of languages such as: regular languages or context sensitive language. We also give an example of context sensitive language, but not a context-free language (according to Chomsky hierarchy), that can be generated by labelled stratified graphs. The concepts introduced in this paper can initiate a possible research line concerning the generative power of the formal languages generated by labelled stratified graphs.
Keywords:
Context-free language, interpretation, labelled graph, labelled stratified graph, Peano algebra, regular language
Full Paper, pp. 101-108

 

DiaSys – A Dialogue System based on Conditional Knowledge in Client-Server Technology

by Cristina Zamfir (Tudorache)

Abstract: A conditional schema is a graph-based structure which is able to represent conditional knowledge. In [1] the structure of a conditional schema was introduced and in [2] the inference mechanism corresponding to the conditional schema representations was developed. The Knowledge Representation and Reasoning (KR&R) is based on the idea that propositional content can be rigorously represented in formal languages, such that the resulted representations can be productively reasoned over by humans and machines [16]. Conditional Knowledge Representation and Reasoning represents a new brand of KR&R, for which several formalisms have been developed. The DiaSys application based on conditional knowledge was presented in [4]. In this paper we describe the evolution of that dialogue system into a client-server application. The conditional schema will be defined and exemplified. The server and the client will be presented step by step in two separate sections. The two sections are oriented specifically towards the task that the described part of the application performs.
Keywords:
Client-server, Dialog System, Knowledge base, Natural language, XML
Full Paper, pp. 109-116

 

MAGS – An Approach Using Multi-Objective Evolutionary Algorithms for Grid Task Scheduling

by Miguel Camelo, Yezid Donoso, Harold Castro

Abstract: Grid task scheduling problem has been a research focus in grid computing for the past years. Some Deterministic, Heuristics or Metaheuristic scheduling approaches have been proposed to solve this NP-complete problem. However, these algorithms do not take the Multi-Objective nature of Grid Computing performance into account. In this paper we present a Multi-Objective approach using Evolutionary Algorithm (MOEA) to efficiently solve such kind of scheduling problems. Our proposal is based on NSGA-II MOEA algorithm combined with a set of Heuristics in different evolutionary operators which allow a fast convergence to optimal (or near-optimal) solutions. The results obtained by our proposed algorithm were compared and evaluated against Mono-Objective and Multi-Objective algorithms used for Grid task scheduling. The main contributions of this paper are the proposed mathematical model, the optimization model and the algorithm to solve it. Additionally we show the effectiveness and robustness of the proposed algorithm.
Keywords:
Evolutionary Algorithms, Grid, Task Scheduling, Heuristics, Metaheuristics, Meta-Scheduler, Multi-Objective Optimization, NP-Complete, NSGA-II, Pareto Front
Full Paper, pp. 117-126

 

Scaled Image Edge Detection Based on The Total Variation Functional

by Peter Ndajah, Hisakazu Kikuchi

Abstract: We present total variation anisotropic edge detection. We derive the total variation functional from measure theory, distribution theory and vector gradient method. The Euler-Lagrange equation of the total variation functional gives a steady state equation. The steady state equation acts as an anisotropic filter on an image. The total variation filtered images are compared to Laplacian filtered images. A subsequent application of the zero crossing algorithm works quite well for the traditional Marr-Hildreth method but gives poorer results for the total variation filtered images. It was found that thresholding methods work better and saves computational time. Also, our results show that total variation edge detection overcomes some drawbacks associated with the Marr-Hildreth method.
Keywords:
Anisotropic, Euler-Lagrange, LoG, Total Variation
Full Paper, pp. 127-136

 

Technologies for Development of the Information Systems: from ERP to e-Government

by D. Litan, A. Apostu, L. Copcea (Teohari), M. Teohari

Abstract: Although some ERP (Enterprise Resource Planning) software products are available on the market, they do not entirely fit each and every type of business. That is the main reason why companies prefer to develop their own ERP type of systems, thus eliminating “barriers” such as long time and high costs required to develop the software, compared with the shorter time and lower costs of just purchasing the ERP system and customizing it to the targeted business. However, in the short run, lower cost will eventually result into losses in the long run, due to that, one the one hand, the system is not perfectly adapted to business specifics and, on the other hand, customizing an ERP package can influence implementation of the standard benefits of an integrated system. Therefore, many times, the best choice for a company would be to develop an ERP type of system that caters perfectly for the needs of the organization. In this paper, we will proceed to the analysis, presentation and comparison of the current most important information methodologies (for business analysis and for IT projects – people-oriented methodologies) and databases/ data warehouses which are being used in development of the ERP systems. Also, another aspect that we will consider, in our paper is to examine how some of the technologies used in the development of ERP systems can be applied in the development of e-government type systems.
Keywords:
Data warehouse, e-government, Enterprise Resource Planning, information methodology, query optimization, information system, information technology
Full Paper, pp. 137-152

 

Issue 4, Volume 5, 2011

Efficient ID-Based Multi-Proxy Multi-Signature Scheme Based on CDHP

by Rajeev Anand Sahu, Sahadeo Padhye

Abstract: Proxy signature is useful in situation when a user wants to authorize an agent called proxy signer to sign any document on his behalf. Multi-proxy multi-signature is one of the primitives of proxy signature. Bilinear pairing makes the system efficient and provides an ease in computation. In this paper, we propose an ID-based multi-proxy multi-signature scheme from bilinear pairings based on Computational Diffie-Hellman Problem (CDHP), replacing the certificate generation of Li and Chen’s scheme by delegation generation. Our scheme is proxy protected and computationally more efficient than the ID-based multi-proxy multi-signature scheme of Li and Chen. We also analyze the security properties of our scheme and show that the proposed scheme satisfies all the security requirements of a safe proxy signature scheme.
Keywords:
Bilinear pairings, Cryptography, ID-based signature scheme, Multi-proxy multi-signature, Security analysis
Full Paper, pp. 275-282

 

Elman Networks for the Prediction of Inventory Levels and Capacity Utilization

by F. Harjes, B. Scholz-Reiter, A. Kaviani Mehr

Abstract: Today's production processes face an increase in dynamics and complexity. Therefore, production control techniques face a demand for continuous advancement. Methods from the field of artificial intelligence, such as neural networks, have proven their applicability in this area. They are applied for optimization, prediction, classification, control and many other production related areas. This paper introduces an approach using Elman Networks for the workstation-specific prediction of inventory levels and capacity utilization within a shop floor environment. It includes the selection of the appropriate network architecture, the determination of suitable input variables as well as the training and validation process. The evaluation of the proposed approach takes place by means of a generic shop floor model.
Keywords:
Artificial neural networks, Elman networks, predictive control, shop floor production
Full Paper, pp. 283-290

 

Diagnosis of Perceived Attitude, Importance, and Knowledge in Statistics Based on Rasch Probabilistic Model

by Zamalia Mahmud

Abstract: Postgraduate students who are involved in the analyiss of survey research data are expected to have a sound knowledge in statistics. However, statistics learners should firstly inculcate interest, feel the importance of statistics and instill positive attitude toward learning statistics. Based on Rasch Unidimensional Measurement Model, this study shall investigate the attitude, importance and knowledge in statistics among postgraduate students. The investigation involved a small classroom experiment measuring students’ perceived attitude prior to and after attending the statistics remedial classes and then their perceived knowledge on various statistical tools and concepts. The validity of the response patterns were measured and whether their responses can be taken as measuring the latent trait or otherwise. This paper also describes how Rasch probabilistic model is used to measure the latent trait of students’ attitude in response to the agent or items measuring the trait. This involves a calibration between person ability to respond towards the five-point Likert scale items and whether the items function in unison to form a single underlying pattern. Logit scores as derived from the Rasch measurement model, Person-Item Distribution map and other related Rasch statistics are executed to endorse the trait. The study shows that student demonstrate a higher cognitive positive attitude toward learning statistics after attending the remedial classes with most post-attitude items superseding the pre-attitude items during the calibration. However, there are traces of misfit responses as identified by the fit indices due to unusual or/and inappropriate responses which are highlighted in this paper.
Keywords:
Attitude towards learning statistics, Pre-post attitude Rasch probabilistic model, PIDM, Person-Item misfit, Perceived Importance and Knowledge
Full Paper, pp. 291-298

 

The Relationship Profitability - Risk for an Optimal Portfolio Building with Two Risky Assets and a Risk-Free Asset

by Florentin Serban, Maria Viorica Stefanescu, Silvia Dedu

Abstract: This paper present a combination of classification theory with risk estimation theory to determine the best assets through those that we analyzed. We present also building to the efficient frontier for a portfolio with 2 or 3 assets We use a data analysis method to obtain two classes of assets and then we estimate the risk of each asset corresponding to each class. Thus, we get the best two assets among those considered risky for which we build the efficient frontier if we consider that the portfolio consists of these two risky assets and a risk-free asset. To illustrate the effectiveness of the method used, we present a case study that refers to the Bucharest Stock Exchange. We construct the efficient frontier, based on the correlation of the best stocks that we have obtained through data analysis (for classification), and by assessing the loss distribution (for risk assessment), taking into account that the portfolio contains an asset without risk.
Keywords:
Efficient frontier, Optimization, Portfolio selection, Principal component analysis, Risk estimation
Full Paper, pp. 299-306