Newer
Older
Discussion_Papers / Website / dp1999-abstracts-contents.htm
nstanger on 11 Jun 2009 37 KB - Regenerated 1997--1999 DPs.
<div class="sectionTitle">Information Science Discussion Papers Series: 1999 Abstracts</div>

<hr>

<a name="dp1999-01"></a><h3>99/01: UML as an ontology modelling language</h3> 

<h4>S. Cranefield and M. K. Purvis</h4> 

<p>Current tools and techniques for ontology development are based on the traditions of AI knowledge representation research. This research has led to popular formalisms such as KIF and KL-ONE style languages. However, these representations are little known outside AI research laboratories. In contrast, commercial interest has resulted in ideas from the object-oriented programming community maturing into industry standards and powerful tools for object-oriented analysis, design and implementation. These standards and tools have a wide and rapidly growing user community. This paper examines the potential for object-oriented standards to be used for ontology modelling, and in particular presents an ontology representation language based on a subset of the Unified Modeling Language together with its associated Object Constraint Language.</p> 

<p><a href="papers/dp1999-01.pdf">Download</a> (PDF, 204 KB)</p> 

<hr>

<a name="dp1999-02"></a><h3>99/02: Evolving connectionist systems for on-line, knowledge-based learning: Principles and applications</h3> 

<h4>N. Kasabov</h4> 

<p>The paper introduces evolving connectionist systems (ECOS) as an effective approach to building on-line, adaptive intelligent systems. ECOS evolve through incremental, hybrid (supervised/unsupervised), on-line learning. They can accommodate new input data, including new features, new classes, etc. through local element tuning. New connections and new neurons are created during the operation of the system. The ECOS framework is presented and illustrated on a particular type of evolving neural networks&mdash;evolving fuzzy neural networks (EFuNNs). EFuNNs can learn spatial-temporal sequences in an adaptive way, through one pass learning. Rules can be inserted and extracted at any time of the system operation. The characteristics of ECOS and EFuNNs are illustrated on several case studies that include: adaptive pattern classification; adaptive, phoneme-based spoken language recognition; adaptive dynamic time-series prediction; intelligent agents.</p> 

<p><strong>Keywords:</strong> evolving connectionist systems, evolving fuzzy neural networks, on-line learning, spatial-temporal adaptation, adaptive speech recognition</p> 

<p><a href="papers/dp1999-02.pdf">Download</a> (PDF, 944 KB)</p> 

<hr>

<a name="dp1999-03"></a><h3>99/03: Spatial-temporal adaptation in evolving fuzzy neural networks for on-line adaptive phoneme recognition</h3> 

<h4>N. Kasabov and M. Watts</h4> 

<p>The paper is a study on a new class of spatial-temporal evolving fuzzy neural network systems (EFuNNs) for on-line adaptive learning, and their applications for adaptive phoneme recognition. The systems evolve through incremental, hybrid (supervised / unsupervised) learning. They accommodate new input data, including new features, new classes, etc. through local element tuning. Both feature-based similarities and temporal dependencies, that are present in the input data, are learned and stored in the connections, and adjusted over time. This is an important requirement for the task of adaptive, speaker independent spoken language recognition, where new pronunciations and new accents need to be learned in an on-line, adaptive mode. Experiments with EFuNNs, and also with multi-layer perceptrons, and fuzzy neural networks (FuNNs), conducted on the whole set of New Zealand English phonemes, show the superiority and the potential of EFuNNs when used for the task. Spatial allocation of nodes and their aggregation in EFuNNs allow for similarity preserving and similarity observation within one phoneme data and across phonemes, while subtle temporal variations within one phoneme data can be learned and adjusted through temporal feedback connections. The experimental results support the claim that spatial-temporal organisation in EFuNNs can lead to a significant improvement in the recognition rate especially for the diphthong and the vowel phonemes in English, which in many cases are problematic for a system to learn and adjust in an adaptive way.</p> 

<p><a href="papers/dp1999-03.pdf">Download</a> (PDF, 560 KB)</p> 

<hr>

<a name="dp1999-04"></a><h3>99/04: Dynamic evolving fuzzy neural networks with &lsquo;m-out-of-n&rsquo; activation nodes for on-line adaptive systems</h3> 

<h4>N. Kasabov and Q. Song</h4> 

<p>The paper introduces a new type of evolving fuzzy neural networks (EFuNNs), denoted as mEFuNNs, for on-line learning and their applications for dynamic time series analysis and prediction. mEFuNNs evolve through incremental, hybrid (supervised/unsupervised), on-line learning, like the EFuNNs. They can accommodate new input data, including new features, new classes, etc. through local element tuning. New connections and new neurons are created during the operation of the system. At each time moment the output vector of a mEFuNN is calculated based on the m-most activated rule nodes. Two approaches are proposed: (1) using weighted fuzzy rules of Zadeh-Mamdani type; (2) using Takagi-Sugeno fuzzy rules that utilise dynamically changing and adapting values for the inference parameters. It is proved that the mEFuNNs can effectively learn complex temporal sequences in an adaptive way and outperform EFuNNs, ANFIS and other neural network and hybrid models. Rules can be inserted, extracted and adjusted continuously during the operation of the system. The characteristics of the mEFuNNs are illustrated on two bench-mark dynamic time series data, as well as on two real case studies for on-line adaptive control and decision making. Aggregation of rule nodes in evolved mEFuNNs can be achieved through fuzzy C-means clustering algorithm which is also illustrated on the bench mark data sets. The regularly trained and aggregated in an on-line, self-organised mode mEFuNNs perform as well, or better, than the mEFuNNs that use fuzzy C-means clustering algorithm for off-line rule node generation on the same data set.</p> 

<p><strong>Keywords:</strong> dynamic evolving fuzzy neural networks, on-line learning, adaptive control, dynamic time series prediction, fuzzy clustering</p> 

<p><a href="papers/dp1999-04.pdf">Download</a> (PDF, 2 MB)</p> 

<hr>

<a name="dp1999-05"></a><h3>99/05: Hybrid neuro-fuzzy inference systems and their application for on-line adaptive learning of nonlinear dynamical systems</h3> 

<h4>J. Kim and N. Kasabov</h4> 

<p>In this paper, an adaptive neuro-fuzzy system, called HyFIS, is proposed to build and optimise fuzzy models. The proposed model introduces the learning power of neural networks into the fuzzy logic systems and provides linguistic meaning to the connectionist architectures. Heuristic fuzzy logic rules and input-output fuzzy membership functions can be optimally tuned from training examples by a hybrid learning scheme composed of two phases: the phase of rule generation from data, and the phase of rule tuning by using the error backpropagation learning scheme for a neural fuzzy system. In order to illustrate the performance and applicability of the proposed neuro-fuzzy hybrid model, extensive simulation studies of nonlinear complex dynamics are carried out. The proposed method can be applied to on-line incremental adaptive leaning for the purpose of prediction and control of non-linear dynamical systems.</p> 

<p><strong>Keywords:</strong> neuro-fuzzy systems, neural networks, fuzzy logic, parameter and structure learning, knowledge acquisition, adaptation, time series</p> 


<hr>

<a name="dp1999-06"></a><h3>99/06: A distributed architecture for environmental information systems</h3> 

<h4>M. K. Purvis and S. Cranefield and M. Nowostawski</h4> 

<p>The increasing availability and variety of large environmental data sets is opening new opportunities for data mining and useful cross-referencing of disparate environmental data sets distributed over a network. In order to take advantage of these opportunities, environmental information systems will need to operate effectively in a distributed, open environment. In this paper, we describe the New Zealand Distributed Information System (NZDIS) software architecture for environmental information systems. In order to optimise extensibility, openness, and flexible query processing, the architecture is organised into collaborating software agents that communicate by means of a standard declarative agent communication language. The metadata of environmental data sources are stored as part of agent ontologies, which represent information models of the domain of the data repository. The agents and the associated ontological framework are designed as much as possible to take advantage of standard object-oriented technology, such as CORBA, UML, and OQL, in order to enhance the openness and accessibility of the system.</p> 

<p><a href="papers/dp1999-06.pdf">Download</a> (PDF, 208 Kb)</p> 

<hr>

<a name="dp1999-07"></a><h3>99/07: From hybrid adjustable neuro-fuzzy systems to adaptive connectionist-based systems for phoneme and word recognition</h3> 

<h4>N. Kasabov and R. Kilgour and S. Sinclair</h4> 

<p>This paper discusses the problem of adaptation in automatic speech recognition systems (ASRS) and suggests several strategies for adaptation in a modular architecture for speech recognition. The architecture allows for adaptation at different levels of the recognition process, where modules can be adapted individually based on their performance and the performance of the whole system. Two realisations of this architecture are presented along with experimental results from small-scale experiments. The first realisation is a hybrid system for speaker-independent phoneme-based spoken word recognition, consisting of neural networks for recognising English phonemes and fuzzy systems for modelling acoustic and linguistic knowledge. This system is adjustable by additional training of individual neural network modules and tuning the fuzzy systems. The increased accuracy of the recognition through appropriate adjustment is also discussed. The second realisation of the architecture is a connectionist system that uses fuzzy neural networks FuNNs to accommodate both a prior linguistic knowledge and data from a speech corpus. A method for on-line adaptation of FuNNs is also presented.</p> 

<p><strong>Keywords:</strong> pattern recognition, artificial intelligence, neural networks, speech recognition</p> 


<hr>

<a name="dp1999-08"></a><h3>99/08: Adaptive, evolving, hybrid connectionist systems for image pattern recognition</h3> 

<h4>N. Kasabov and S. Israel and B. Woodford</h4> 

<p>The chapter presents a new methodology for building adaptive, incremental learning systems for image pattern classification. The systems are based on dynamically evolving fuzzy neural networks that are neural architectures to realise connectionist learning, fuzzy logic inference, and case-based reasoning. The methodology and the architecture are applied on two sets of real data&mdash;one of satellite image data, and the other of fruit image data. The proposed method and architecture encourage fast learning, life-long learning and on-line learning when the system operates in a changing environment of image data.</p> 

<p><strong>Keywords:</strong> image classification, evolving fuzzy neural networks, case-based reasoning</p> 

<p><a href="papers/dp1999-08.pdf">Download</a> (PDF, 1.3 MB)</p> 

<hr>

<a name="dp1999-09"></a><h3>99/09: The concepts of hidden Markov model in speech recognition</h3> 

<h4>W. Abdulla and N. Kasabov</h4> 

<p>The speech recognition field is one of the most challenging fields that has faced scientists for a long time. The complete solution is still far from reach. The efforts are concentrated with huge funds from the companies to different related and supportive approaches to reach the final goal. Then, apply it to the enormous applications that are still waiting for the successful speech recognisers that are free from the constraints of speakers, vocabularies or environment. This task is not an easy one due to the interdisciplinary nature of the problem and as it requires speech perception to be implied in the recogniser (Speech Understanding Systems) which in turn point strongly to the use of intelligence within the systems.</p>

<p>The bare techniques of recognisers (without intelligence) are following wide varieties of approaches with different claims of success by each group of authors who put their faith in their favourite way. However, the sole technique that gains the acceptance of the researchers to be the state of the art is the Hidden Markov Model (HMM) technique. HMM is agreed to be the most promising one. It might be used successfully with other techniques to improve the performance, such as hybridising the HMM with Artificial Neural Networks (ANN) algorithms. This does not mean that the HMM is pure from approximations that are far from reality, such as the successive observations independence, but the results and potential of this algorithm is reliable. The modifications on HMM take the burden of releasing it from these poorly representative approximations hoping for better results.</p>

<p>In this report we are going to describe the backbone of the HMM technique with the main outlines for successful implementation. The representation and implementation of HMM varies in one way or another but the main idea is the same as well as the results and computation costs, it is a matter of preferences to choose one. Our preference here is that adopted by Ferguson and Rabiner et al.</p>

<p>In this report we will describe the Markov Chain, and then investigate a very popular model in the speech recognition field (the Left-Right HMM Topology). The mathematical formulations needed to be implemented will be fully explained as they are crucial in building the HMM. The prominent factors in the design will also be discussed. Finally we conclude this report by some experimental results to see the practical outcomes of the implemented model.</p> 

<p><a href="papers/dp1999-09.pdf">Download</a> (PDF, 556 KB)</p> 

<hr>

<a name="dp1999-10"></a><h3>99/10: Finding medical information on the Internet: Who should do it and what should they know</h3> 

<h4>D. Parry</h4> 

<p>More and more medical information is appearing on the Internet, but it is not easy to get at the nuggets amongst all the spoil. Bruce McKenzie&rsquo;s editorial in the December 1997 edition of SIM Quarterly dealt very well with the problems of quality, but I would suggest that the problem of accessibility is as much of a challenge. As ever-greater quantities of high quality medical information are published electronically, the need to be able to find it becomes imperative. There are a number of tools to find what you want on the Internet&mdash;search engines, agents, indexing and classification schemes and hyperlinks, but their use requires care, skill and experience.</p> 

<p><a href="papers/dp1999-10.pdf">Download</a> (PDF, 196 KB)</p> 

<hr>

<a name="dp1999-11"></a><h3>99/11: Software metrics data analysis&mdash;Exploring the relative performance of some commonly used modeling techniques</h3> 

<h4>A. Gray and S. MacDonell</h4> 

<p>Whilst some software measurement research has been unquestionably successful, other research has struggled to enable expected advances in project and process management. Contributing to this lack of advancement has been the incidence of inappropriate or non-optimal application of various model-building procedures. This obviously raises questions over the validity and reliability of any results obtained as well as the conclusions that may have been drawn regarding the appropriateness of the techniques in question. In this paper we investigate the influence of various data set characteristics and the purpose of analysis on the effectiveness of four model-building techniques&mdash;three statistical methods and one neural network method. In order to illustrate the impact of data set characteristics, three separate data sets, drawn from the literature, are used in this analysis. In terms of predictive accuracy, it is shown that no one modeling method is best in every case. Some consideration of the characteristics of data sets should therefore occur before analysis begins, so that the most appropriate modeling method is then used. Moreover, issues other than predictive accuracy may have a significant influence on the selection of model-building methods. These issues are also addressed here and a series of guidelines for selecting among and implementing these and other modeling techniques is discussed.</p> 

<p><strong>Keywords:</strong> software metrics, analysis, statistical methods, connectionist methods</p> 

<p><a href="papers/dp1999-11.pdf">Download</a> (PDF, 292 KB)</p> 

<hr>

<a name="dp1999-12"></a><h3>99/12: Software forensics for discriminating between program authors using case-based reasoning, feed-forward neural networks and multiple discriminant analysis</h3> 

<h4>S. MacDonell and A. Gray and G. MacLennan and P. Sallis</h4> 

<p>Software forensics is a research field that, by treating pieces of program source code as linguistically and stylistically analyzable entities, attempts to investigate aspects of computer program authorship. This can be performed with the goal of identification, discrimination, or characterization of authors. In this paper we extract a set of 26 standard authorship metrics from 351 programs by 7 different authors. The use of feed-forward neural networks, multiple discriminant analysis, and case-based reasoning is then investigated in terms of classification accuracy for the authors on both training and testing samples. The first two techniques produce remarkably similar results, with the best results coming from the case-based reasoning models. All techniques have high prediction accuracy rates, supporting the feasibility of the task of discriminating program authors based on source-code measurements.</p> 

<p><a href="papers/dp1999-12.pdf">Download</a> (PDF, 148 KB)</p> 

<hr>

<a name="dp1999-13"></a><h3>99/13: FULSOME: Fuzzy logic for software metric practitioners and researchers</h3> 

<h4>S. MacDonell and A. Gray and J. Calvert</h4> 

<p>There has been increasing interest in recent times for using fuzzy logic techniques to represent software metric models, especially those predicting development effort. The use of fuzzy logic for this application area offers several advantages when compared to other commonly used techniques. These include the use of a single model with different levels of precision for inputs and outputs used throughout the development life cycle, the possibility of model development with little or no data, and its effectiveness when used as a communication tool. The use of fuzzy logic in any applied field however requires that suitable tools are available for both practitioners and researchers&mdash;satisfying both interface and functionality related requirements. After outlining some of the specific needs of the software metrics community, including results from a survey of software developers on this topic, the paper describes the use of a set of tools called FULSOME (Fuzzy Logic for Software Metrics). The development of a simple fuzzy logic system by a software metrician and subsequent tuning are then discussed using a real-world set of software metric data. The automatically generated fuzzy model performs acceptably when compared to regression-based models.</p> 

<p><a href="papers/dp1999-13.pdf">Download</a> (PDF, 236 KB)</p> 

<hr>

<a name="dp1999-14"></a><h3>99/14: Assessing prediction systems</h3> 

<h4>B. Kitchenham and S. MacDonell and L. Pickard and M. Shepperd</h4> 

<p>For some years software engineers have been attempting to develop useful prediction systems to estimate such attributes as the effort to develop a piece of software and the likely number of defects. Typically, prediction systems are proposed and then subjected to empirical evaluation. Claims are then made with regard to the quality of the prediction systems. A wide variety of prediction quality indicators have been suggested in the literature. Unfortunately, we believe that a somewhat confusing state of affairs prevails and that this impedes research progress. This paper aims to provide the research community with a better understanding of the meaning of, and relationship between, these indicators. We critically review twelve different approaches by considering them as descriptors of the residual variable. We demonstrate that the two most popular indicators MMRE and pred(25) are in fact indicators of the spread and shape respectively of prediction accuracy where prediction accuracy is the ratio of estimate to actual (or actual to estimate). Next we highlight the impact of the choice of indicator by comparing three prediction systems derived using four different simulated datasets. We demonstrate that the results of such a comparison depend upon the choice of indicator, the analysis technique, and the nature of the dataset used to derive the predictive model. We conclude that prediction systems cannot be characterised by a single summary statistic. We suggest that we need indicators of the central tendency and spread of accuracy as well as indicators of shape and bias. For this reason, boxplots of relative error or residuals are useful alternatives to simple summary metrics.</p> 

<p><strong>Keywords:</strong> prediction systems, estimation, empirical analysis, metrics, goodness-of-fit statistics</p> 

<p><a href="papers/dp1999-14.pdf">Download</a> (PDF, 304 KB)</p> 

<hr>

<a name="dp1999-15"></a><h3>99/15: Industry practices in project management for multimedia information systems</h3> 

<h4>S. MacDonell and T. Fletcher</h4> 

<p>This paper describes ongoing research directed at formulating a set of appropriate measures for assessing and ultimately predicting effort requirements for multimedia systems development. Whilst significant advances have been made in the determination of measures for both transaction-based and process-intensive systems, very little work has been undertaken in relation to measures for multimedia systems. A small preliminary empirical study is reviewed as a precursor to a more exploratory investigation of the factors that are considered by industry to be influential in determining development effort. This work incorporates the development and use of a goal-based framework to assist the measure selection process from a literature basis, followed by an industry questionnaire. The results provide a number of preliminary but nevertheless useful insights into contemporary project management practices with respect to multimedia systems.</p> 

<p><a href="papers/dp1999-15.pdf">Download</a> (PDF, 228 KB)</p> 

<hr>

<a name="dp1999-16"></a><h3>99/16: Factors systematically associated with errors in subjective estimates of software development effort: The stability of expert judgment</h3> 

<h4>A. Gray and S. MacDonell and M. Shepperd</h4> 

<p>Software metric-based estimation of project development effort is most often performed by expert judgment rather than by using an empirically derived model (although such may be used by the expert to assist their decision). One question that can be asked about these estimates is how stable are they with respect to characteristics of the development process and product? This stability can be assessed in relation to the degree to which the project has advanced over time, the type of module for which the estimate is being made, and the characteristics of that module. In this paper we examine a set of expert-derived estimates for the effort required to develop a collection of modules from a large health-care system. Statistical tests are used to identify relationships between the type (screen or report) and characteristics of modules and the likelihood of the associated development effort being under-estimated, approximately correct, or over-estimated. Distinct relationships are found that suggest that the estimation process being examined was not unbiased to such characteristics.</p> 

<p><a href="papers/dp1999-16.pdf">Download</a> (PDF, 236 KB)</p> 

<hr>

<a name="dp1999-17"></a><h3>99/17: The NZDIS project: An agent-based distributed information systems architecture</h3> 

<h4>M. K. Purvis and S. Cranefield and G. Bush and D. Carter and B. McKinlay and M. Nowostawski and R. Ward</h4> 

<p>This paper describes an architecture for building distributed information systems from existing information resources, based on distributed object and software agent technologies. This architecture is being developed as part of the New Zealand Distributed Information Systems (NZDIS) project.</p>

<p>An agent-based architecture is used: information sources are encapsulated as information agents that accept messages in an agent communication language (the FIPA ACL). A user agent assists users to browse ontologies appropriate to their domain of interest and to construct queries based on terms from one or more ontologies. One or more query processing agents are then responsible for discovering (from a resource broker agent) which data source agents are relevant to the query, decomposing the query into subqueries suitable for those agents (including the translation of the query into the specific ontologies implemented by those agents), executing the subqueries and translating and combining the subquery results into the desired result set.</p>

<p>Novel features of this system include the use of standards from the object-oriented community such as the Common Object Request Broker Architecture (CORBA) (as a communications infrastructure), the Unified Modeling Language (used as an ontology representation language), the Object Data Management Group&rsquo;s Object Query Language (used for queries) and the Object Management Group&rsquo;s Meta Object Facility (used as the basis for an ontology repository agent). Query results need not be returned within an ACL message, but may instead be represented by a CORBA object reference which may be used to obtain the result set.</p> 

<p><a href="papers/dp1999-17.pdf">Download</a> (PDF, 212 KB)</p> 

<hr>

<a name="dp1999-18"></a><h3>99/18: HTN planning for information processing tasks</h3> 

<h4>S. Cranefield</h4> 

<p>This paper discusses the problem of integrated planning and execution for tasks that involve the consumption, production and alteration of relational information. Unlike information retrieval problems, the information processing domain requires explicit modelling of the changing information state of the domain and how the validity of resources changes as actions are performed. A solution to this problem is presented in the form of a specialised hierarchical task network planning model. A distinction is made between the information processing effects of an action (modelled in terms of constraints relating the domain information before and after the action) and the actions&rsquo; preconditions and effects which are expressed in terms of required, produced and invalidated resources. The information flow between tasks is explicitly represented in methods and plans, including any required information-combining operations such as selection and union.</p>

<p>The paper presents the semantics of this model and discusses implementation issues arising from the extension of an existing HTN planner (SHOP) to support this model of planning.</p> 

<p><strong>Keywords:</strong> HTN planning, information processing, integrated planning and execution</p> 

<p><a href="papers/dp1999-18.pdf">Download</a> (PDF, 188 KB)</p> 

<hr>

<a name="dp1999-19"></a><h3>99/19: Automated scoring of practical tests in an introductory course in information technology</h3> 

<h4>G. Kennedy</h4> 

<p>In an introductory course in information technology at the University of Otago the acquisition of practical skills is considered to be a prime objective. An effective way of assessing the achievement of this objective is by means of a &lsquo;practical test&rsquo;, in which students are required to accomplish simple tasks in a controlled environment. The assessment of such work demands a high level of expertise, is very labour intensive and can suffer from marker inconsistency, particularly with large candidatures.</p>

<p>This paper describes the results of a trial in which the efforts of one thousand students in a practical test of word processing were scored by means of a program written in MediaTalk. Details of the procedure are given, including sampling strategies for the purpose of validation and examples of problems that were encountered.</p>

<p>It was concluded that the approach was useful, and once properly validated gave rise to considerable savings in the time and effort.</p> 

<p><strong>Keywords:</strong> computer-aided learning, automated scoring, computer education, test validation</p> 

<p><a href="papers/dp1999-19.pdf">Download</a> (PDF, 216 KB)</p> 

<hr>

<a name="dp1999-20"></a><h3>99/20: Fuzzy logic for software metric models throughout the development life-cycle</h3> 

<h4>A. Gray and S. MacDonell</h4> 

<p>One problem faced by managers who are using project management models is the elicitation of numerical inputs. Obtaining these with any degree of confidence early in a project is not always feasible. Related to this difficulty is the risk of precisely specified outputs from models leading to overcommitment. These problems can be seen as the collective failure of software measurements to represent the inherent uncertainties in managers&rsquo; knowledge of the development products, resources, and processes. It is proposed that fuzzy logic techniques can help to overcome some of these difficulties by representing the imprecision in inputs and outputs, as well as providing a more expert-knowledge based approach to model building. The use of fuzzy logic for project management however should not be the same throughout the development life cycle. Different levels of available information and desired precision suggest that it can be used differently depending on the current phase, although a single model can be used for consistency.</p> 

<p><a href="papers/dp1999-20.pdf">Download</a> (PDF, 148 KB)</p> 

<hr>

<a name="dp1999-21"></a><h3>99/21: Wayfinding/navigation within a QTVR virtual environment: Preliminary results</h3> 

<h4>B. Norris and D. Rashid and W. Wong</h4> 

<p>This paper reports on an investigation into wayfinding principles, and their effectiveness within a virtual environment. To investigate these principles, a virtual environment of an actual museum was created using QuickTime Virtual Reality. Wayfinding principles used in the real world were identified and used to design the interaction of the virtual environment. The initial findings suggests that real-world navigation principles, such as the use of map and landmark principles, can significantly help navigation within this virtual environment. However, navigation difficulties were discovered through an Activity Theory-based Cognitive Task Analysis.</p> 

<p><strong>Keywords:</strong> wayfinding, navigation, QTVR, virtual environments, activity theory</p> 


<hr>

<a name="dp1999-22"></a><h3>99/22: Predictive modelling of plankton dynamics in freshwater lakes using genetic programming</h3> 

<h4>P. Whigham and F. Recknagel</h4> 

<p>Building predictive time series models for freshwater systems is important both for understanding the dynamics of these natural systems and in the development of decision support and management software. This work describes the application of a machine learning technique, namely genetic programming (GP), to the prediction of chlorophyll-a. The system endeavoured to evolve several mathematical time series equations, based on limnological and climate variables, which could predict the dynamics of chlorophyll-a on unseen data. The predictive accuracy of the genetic programming approach was compared with an artificial neural network and a deterministic algal growth model. The GP system evolved some solutions which were improvements over the neural network and showed that the transparent nature of the solutions may allow inferences about underlying processes to be made. This work demonstrates that non-linear processes in natural systems may be successfully modelled through the use of machine learning techniques. Further, it shows that genetic programming may be used as a tool for exploring the driving processes underlying freshwater system dynamics.</p> 

<p><a href="papers/dp1999-22.pdf">Download</a> (PDF, 264 KB)</p> 

<hr>

<a name="dp1999-23"></a><h3>99/23: Modifications to Smith&rsquo;s method for deriving normalised relations from a functional dependency diagram</h3> 

<h4>N. Stanger</h4> 

<p>Smith&rsquo;s method (Smith, 1985) is a formal technique for deriving a set of normalised relations from a functional dependency diagram (FDD). Smith&rsquo;s original rules for deriving these relations are incomplete, as they do not fully address the issue of determining the foreign key links between relations. In addition, one of the rules for deriving foreign keys can produce incorrect results, while the other rule is difficult to automate. In this paper are described solutions these issues.</p> 

<p><strong>Keywords:</strong> normalisation functional dependencies relational model data model translation</p> 

<p><a href="papers/dp1999-23.pdf">Download</a> (PDF, 184 KB)</p> 

<hr>

<a name="dp1999-24"></a><h3>99/24: The development of an electronic distance learning course in health informatics</h3> 

<h4>D. Parry and S. Cockcroft and A. Breton and D. Abernethy and J. Gillies</h4> 

<p>Since 1997 the authors have been involved in the development of a distance learning course in health informatics. The course is delivered via CD-ROM and the Internet. During this process we have learned valuable lessons about computer-assisted collaboration and cooperative work. In particular we have developed methods of using the software tools available for communication and education. We believe that electronic distance learning offers a realistic means of providing education in health informatics and other fields to students whom for reasons of geography or work commitments would not be able to participate in a conventional course.</p> 

<p><a href="papers/dp1999-24.pdf">Download</a> (PDF, 496 KB)</p> 

<hr>

<a name="dp1999-25"></a><h3>99/25: Infiltrating IT into primary care: A case study</h3> 

<h4>S. Cockcroft and D. Parry and A. Breton and D. Abernethy and J. Gillies</h4> 

<p>Web based approaches to tracking students on placement are receiving much interest in the field of medical education The work presented here describes a web-based solution to the problem of managing data collection from student encounters with patients whilst on placement. The solution has been developed by postgraduate students under the direction of staff of the health informatics diploma. Specifically, the system allows undergraduate students on placement or in the main hospital to access a web-based front end to a database designed to store the data that they are required to gather. The system also has the important effect of providing a rationale for the provision of electronic communication to the undergraduate students within the context of healthcare delivery. We believe that an additional effect will be to expose practicing healthcare providers to electronic information systems, along with the undergraduates who are trained to use them, and increase the skill base of the practitioners.</p> 

<p><a href="papers/dp1999-25.pdf">Download</a> (PDF, 92 KB)</p> 

<hr>

<a name="dp1999-26"></a><h3>99/26: Using rough sets to study expert behaviour in induction of labour</h3> 

<h4>D. Parry and W. K. Yeap and N. Pattison</h4> 

<p>The rate of induction of labour (IOL) is increasing, despite no obvious increase in the incidence of the major indications. However the rate varies widely between different centres and practitioners and this does not seem to be due to variations in patient populations. The IOL decision-making process of six clinicians was recorded and examined using hypothetical scenarios presented on a computer. Several rules were identified from a rough sets analysis of the data. These rules were compared to the actual practise of these clinicians in 1994 Initial tests of these rules show that they may form a suitable set for developing an expert system for the induction of labour.</p> 

<p><strong>Keywords:</strong> rough sets, obstetrics, knowledge acquisition</p> 

<p><a href="papers/dp1999-26.pdf">Download</a> (PDF, 108 KB)</p> 

<hr>

<a name="dp1999-27"></a><h3>99/27: Using the Internet to teach health informatics</h3> 

<h4>D. Parry and A. Breton and D. Abernethy and S. Cockcroft and J. Gillies</h4> 

<p>Since July 1998 we have been teaching an Internet-based distance learning course in health informatics (http://basil.otago.ac.nz:800). The development of this course and the experiences we have had running it are described in this paper. The course was delivered using paper materials, a face-to-face workshop, a CD-ROM and Internet communication tools. We currently have about 30 students around New Zealand, a mixture of physicians, nurses and other health staff. Some teaching methods have worked, some haven&rsquo;t, but in the process we have learned a number of valuable lessons.</p> 

<p><strong>Keywords:</strong> distance learning, healthcare, Internet, CD-ROM</p> 

<p><a href="papers/dp1999-27.pdf">Download</a> (PDF, 80 KB)</p>