Volume 2, Number 2, August 2004

Computing and Information Sciences is a peer reviewed journal that is committed to timely publication of original research, surveying and tutorial contributions on the analysis and development of computing and information science. The journal is designed mainly to serve researchers and developers, dealing with information and computing. Papers that can provide both theoretical analysis, along with carefully designed computational experiments, are particularly welcome. The journal is published 2-3 times per year with distribution to librarians, universities, research centers, researchers in computing, mathematics, and information science. The journal maintains strict refereeing procedures through its editorial policies in order to publish papers of only the highest quality. The refereeing is done by anonymous Reviewers. Often, reviews take four months to six months to obtain, occasionally longer, and it takes an additional several months for the publication process.
Paper 1: Numerical Solving of Geometric Constraints By Bisection
Numerical Solving of Geometric Constraints By Bisection
Samy Ait-Aoudia and Imad Mana
Abstract:
In
computer-aided design, geometric modeling by constraints enables users
to describe shapes by relationships called constraints between geometric
elements. The problem is to derive automatically these geometric
elements. Many resolution methods have been proposed for solving systems
of geometric constraints .Geometric methods can be very efficient but
are only applicable to particular kinds of problems (typically rule and
compass constructive problems). Some schemes can't be solved by any
geometric method. A numerical method is then necessary. We use the
bisection method to solve these cases. The bisection method enables
users to reliably find all solutions to a system of non-linear equations
within a region defined by bounds on each individual co-ordinate of the
geometric objects. But bisection is very time consuming. To overcome
this problem, we present first a method that speed-up the resolution
process. This improvement is still insufficient; we propose then to
distribute the processing on several PC machines operating in parallel.
Paper 2: Modeling and Analysis of an On-Line FMS Scheduler Using Colored Petri Nets
Modeling and Analysis of an On-Line FMS Scheduler Using Colored Petri Nets
Ali A. Al-Titinchi and Kasim M. Al-Aubidy
Abstract:
An interactive hierarchical model based on colored Petri net
(CPN) for general flexible manufacturing system (FMS) scheduler is
presented in this paper. The proposed model is partitioned into two
levels to face the complexity of the manufacturing systems. The first
level (cell-level) supervises the jobs scheduling between cells. The
second level (machine-level) supervises the scheduling of the
manufacturing jobs between machines inside the active cells. The goal of
the scheduler model is to interpret the output of the manufacturing
route planner into graphical representation and to achieve the
sequencing, dispatching and monitoring the real time execution of these
manufacturing plans. Also, the proposed CPN model generates feedback
status to monitor the abnormal conditions. Sub CPN priority structures
are embedded in the main model to control the election from the
competitive jobs to use the limited resources in the system. The
hierarchy organization will enhance the real time response of the
sequencing and monitoring capabilities. The designed scheduler has been
evaluated for different case studies and the achieved results are very
encouraging.
Paper 3: Iterative Window Size Estimation on Self-Similarity Measurement for Network ...
Iterative
Window Size Estimation on Self-Similarity Measurement for Network
Traffic Anomaly Detection
Abstract: An iterative method for estimating the optimum sample time (or simply window size) in self-similarity measurement of network traffic is introduced. The main purpose of this measurement is to identify anomaly in network traffic. When the network traffic is close to the self-similarity model, it is considered as normal while otherwise it is not. Since, this model is related to a long-range dependence process, providing data in long period of time will increase the closeness of the network traffic towards the model. On the other hand, increasing the time range is one of the factors that will increase detection loss probability where an intrusive pattern may hide inside the normal data. Thus, the purpose of this method is to minimize the curve-fitting error on self-similarity measurement and detection loss probability in anomaly detection. This iterative method was applied to network traffic data provided by Lincoln Lab, Massachusetts Institute of Technology (MIT). The result has shown, that this method is able to estimate an optimum window size that is capable to reduce detection loss probability and maintain a low error rate.
Paper 4: Efficient Electronic Payment Systems by Using a Sparse Elliptic ...
Efficient Electronic Payment Systems by Using a Sparse Elliptic Curve Cryptography
Essam
Al-Daoud, Khalid Al-Tahat and Hamed Al-Fawareh
Abstract:
This
paper introduces new techniques and algorithms to reduce the size of the
elliptic curve parameters. The basic idea is to use sparse coefficients
and sparse base points. The sparse elements are introduced with a
compact representation, thus the public key parameters are reduced about
37-49 percent. The elliptic curve application such as e-payment and
e-commerce can be implemented with better performance using the
suggested approach.
Paper 5: A Heuristic for Minimizing Makespan in Proportional Parallel Flowshops
A
Heuristic for Minimizing Makespan in Proportional Parallel Flowshops
Abstract: This paper addresses the problem of minimizing the makespan on a two-stage parallel flow shops with proportional processing times. The problem of scheduling n independent jobs on m proportional parallel flow shops to minimize makespan is known to be NP-hard and hence too difficult to solve in polynomial time. A multi-phase heuristic algorithm to minimize the makespan on two proportional parallel flow shops is proposed. A simulation study is conducted to examine the effectiveness of the proposed heuristic algorithm for small and large problems. The results and analysis of quite extensive computational experiments for small size problems show that the proposed algorithm can yield solutions within a few percent of optimal solutions. The computational results for large size problems show that the performance of the proposed algorithm is relatively more effective than the existing SKV algorithm.
Paper 6: Running Multicast Applications Using Reconfigurable Parallel System
Running
Multicast Applications Using Reconfigurable Parallel System
Buhari and Hamid R. Arabnia
Abstract:
In this paper we address the important problem of Multicasting
using a reconfigurable network topology called the Multi-Ring network.
The Multi-Ring network is shown to have a rich set of nice properties
that support applications in medical imaging, scientific computing, and
artificial intelligence. The most important property of the Multi-Ring
topology is “scalability”. It is feasible and cost-effective to
interconnect hundreds of thousands of computing notes using the
Multi-Ring network topology. Efficient Multicasting operation on a
platform composed of thousands of computing nodes is a challenge that is
being addressed in this paper.
Paper 7: Constructing An Automatic Lexicon for Arabic Language
Constructing
An Automatic Lexicon for Arabic Language
Riyad Al-Shalabi and Ghassan Kanaan
Abstract:
In this
paper, we
have designed
and implemented
a system
for building an
automatic lexicon for the
Arabic language. Our Arabic Lexicon
contains word specific information. These pieces of information
include: morphological information such as the root (stem) of the word,
its pattern and its affixes, the part-of-speech tag of the word, which
classifies it as a noun, verb
or particle; lexical attributes such as gender, number, person,
case, definiteness, aspect, and mood are also extracted and stored with
the word in the lexicon. A lexicon is a collection of representations or
words used by a natural
language processor as a source of words specific information; this
representation may contain information about the morphology, phonology,
syntactic argument structure and semantics of the word. A good
lexicon is badly needed for many natural language applications such as:
parsing, text generation,
noun phrase and verb phrase construction and
so on. Many rules
based on the grammar of the Arabic language were used in our
system to identity the part-of-speech tag and the related lexical
attributes of the word [13].
We have tested our system
using a vowelized and
non-vowelized Arabic text
documents taken from the holly Qur'an and 242 Arabic
abstracts chosen
randomly from the proceedings of the Saudi Arabian national
computer conference, and we achieved an accuracy of about 96%. We
discuss the actors behind
these errors and how this accuracy rate can be enhanced.
Contacts
Editor-in-Chief
Prof. Jihad Mohamad Alja'am
Email: journal.editor.ijcis@gmail.com
The Journal Secretary
Eng. Dana Bandok
Ontario, Canada
Email: sec.ijcis@gmail.com