EFM - A Model For Educational Game Design Pp509-517
EFM - A Model For Educational Game Design Pp509-517
EFM - A Model For Educational Game Design Pp509-517
Editorial Board
David Hutchison
Lancaster University, UK
Takeo Kanade
Carnegie Mellon University, Pittsburgh, PA, USA
Josef Kittler
University of Surrey, Guildford, UK
Jon M. Kleinberg
Cornell University, Ithaca, NY, USA
Alfred Kobsa
University of California, Irvine, CA, USA
Friedemann Mattern
ETH Zurich, Switzerland
John C. Mitchell
Stanford University, CA, USA
Moni Naor
Weizmann Institute of Science, Rehovot, Israel
Oscar Nierstrasz
University of Bern, Switzerland
C. Pandu Rangan
Indian Institute of Technology, Madras, India
Bernhard Steffen
University of Dortmund, Germany
Madhu Sudan
Massachusetts Institute of Technology, MA, USA
Demetri Terzopoulos
University of California, Los Angeles, CA, USA
Doug Tygar
University of California, Berkeley, CA, USA
Gerhard Weikum
Max-Planck Institute of Computer Science, Saarbruecken, Germany
Zhigeng Pan Xiaopeng Zhang
Abdennour El Rhalibi Woontack Woo
Yi Li (Eds.)
Technologies for
E-Learning and
Digital Entertainment
13
Volume Editors
Zhigeng Pan
Zhejiang University, State Key Lab of CAD&CG
Hangzhou 310027,China
E-mail: zgpan@cad.zju.edu.cn
Xiaopeng Zhang
Chinese Academy of Sciences
Institute of Automation, National Laboratory of Pattern Recognition
Beijing 100080, China
E-mail: xpzhang@nlpr.ia.ac.cn
Abdennour El Rhalibi
Liverpool John Moores University, School of Computing and Mathematical Sciences
Liverpool L3 3AF, UK
E-mail: A.Elrhalibi@ljmu.ac.uk
Woontack Woo
Gwangju Institute of Science and Technology (GIST), U-VR Lab
Gwangju 500-712, South Korea
E-mail: wwoo@gist.ac.kr
Yi Li
Nanjing Normal University, Edu-game Research Center, Nanjing, China
E-mail: yilisd@163.com
ISSN 0302-9743
ISBN-10 3-540-69734-9 Springer Berlin Heidelberg New York
ISBN-13 978-3-540-69734-3 Springer Berlin Heidelberg New York
This work is subject to copyright. All rights are reserved, whether the whole or part of the material is
concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting,
reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication
or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965,
in its current version, and permission for use must always be obtained from Springer. Violations are liable
to prosecution under the German Copyright Law.
Springer is a part of Springer Science+Business Media
springer.com
© Springer-Verlag Berlin Heidelberg 2008
Printed in Germany
Typesetting: Camera-ready by author, data conversion by Scientific Publishing Services, Chennai, India
Printed on acid-free paper SPIN: 12279272 06/3180 543210
Preface
With the widespread interest in digital entertainment and the advances in the techno-
logies of computer graphics, multimedia and virtual reality technologies, a new area––
“Edutainment”––has been accepted as a union of education and computer entertainment.
Edutainment is recognized as an effective way of learning through a medium, such as a
computer, software, games or VR applications, that both educates and entertains.
The Edutainment conference series was established and followed as a special
event for the new interests in e-learning and digital entertainment. The main purpose
of Edutainment conferences is the discussion, presentation, and information exchange
of scientific and technological developments in the new community. The Edutainment
conference series is a very interesting opportunity for researchers, engineers and
graduate students who wish to communicate at these international annual events. The
conference series includes plenary invited talks, workshops, tutorials, paper presenta-
tion tracks and panel discussions. The Edutainment conference series was initiated in
Hangzhou, China in 2006. Following the success of the first event (Edutainment 2006
in Hangzhou, China) and the second one (Edutainment 2007 in Hong Kong, China),
Edutainment 2008 was held June 25–27, 2007 in Nanjing, China.
This year, we received 219 submissions from 26 different countries and regions,
including United Arab Emirates, Canada, Thailand, New Zealand, Austria, Turkey,
Germany, Switzerland, Brazil, Cuba, Australia, Hong Kong (China), Pakistan, Mex-
ico, Czech Republic, USA, Malaysia, Italy, Spain, France, UK, The Netherlands,
Taiwan (China), Japan, South Korea, and China. A total of 83 papers were selected,
after peer review, for this volume. Topics of these papers fall into ten different areas
ranging from fundamental issues in geometric modeling and imaging to virtual reality
systems and their applications in computer entertainment and education. These topics
include E-Learning Platforms and Tools, E-Learning System for Education, Applica-
tion of E-Learning Systems, E-Learning Resource Management, Interaction in Game
and Education, Integration of Game and Education, Game Design and Development,
Virtual Characters, Animation and Navigation, Graphics Rendering and Digital Me-
dia, and Geometric Modeling for Games and Virtual Reality.
We are grateful to the International Program Committee and the reviewers for their
great effort and serious work to get all the papers reviewed in a short period of time.
We are grateful to the Organizing Committee and Executive Committee for their
support of this event. We would also like to thank the authors and participants for
their enthusiasm and contribution to the success of this conference.
The success of Edutainment 2008 was also due to the financial and practical sup-
port of various institutions.
Sponsors
z VR Committee, China Society of Image and Graphics
z Zhejiang University
VI Preface
Co-sponsors
z National Science Foundation of China (NSFC)
z International Journal of Virtual Reality (IJVR)
z International Journal of Computer Games Technology (IJCGT)
z Transactions on Edutainment (ToE)
z Nanjing Normal University, China
z Hohai University, China
z LIAMA-NLPR, Institute of Automation, CAS, China
We would like to thank all of them for offering the opportunity to organize Edu-
tainment 2008 in a way that provided a diversified scientific and social program.
Especially, we would like to thank all members of the International Program Commit-
tee and Organizing Committee for their great job in defining the conference topics,
reviewing the large number of submitted papers, and managing to put all the material
together for this great event.
Committee
Conference Co-chairs
Program Co-chairs
Publicity Co-chairs
Publication Chairs
Financial Co-chairs
Organizing Co-chairs
General Secretary
Executive Committee
Program Committee
Maha Abdallah, France
Sangchul Ahn, Korea
Isabel M. Alexandre, Portugal
Elisabeth Andre, Germany
Dominique Archambault, France
Ruth Aylett, UK
Leandro Balladares, Mexico
Christian Bauckhage, Germany
Rafael Bidarra, The Netherlands
Paul Brna, UK
Gloria Brown-Simmons, USA
Eliya Buyukkaya, France
Yiyu Cai, Singapore
Christophe Chaillou, France
Tak-Wai Chan, Taiwan, China
Maiga Chang, Canada
Yam San Chee, Singapore
Jim X. Chen, USA
Nuno Correia, Portugal
Jinshi Cui, China
Akshay Darbari, India
stephane Donikian, France
Stephen Downes, Canada
Abdennour El Rhalibi, UK
Guangzheng Fei, China
Paul Fergus, UK
Bernd Froehlich, Germany
Marco Furini, Italy
Lisa Gjedde, Denmark
Martin Goebel, Germany
Orgainzation IX
Sparse Key Points Controlled Animation for Individual Face Model . . . . 613
Jian Yao, Yangsheng Wang, and Bin Ding
Abstract. Corrective feedback and error correction are important tasks for
ESL/EFL (English as a Second Language/English as a Foreign Language)
writing instruction. Research findings showed that students’ major difficulty in
error correction lies in their failure to detect errors. Also, researchers proposed
that error analysis can be reinvented in the form of computer-aided error analysis,
a new type of computer corpus annotation. Annotations on digital documents can
be easily shared among groups of people, making them valuable for a wide
variety of tasks, including providing feedback. This study developed a
web-based online corrective feedback and error analysis system called WRITE
(Writing Revision Instrument for Teaching English). With this system, teachers
can make error corrections on digitized documents, on the general web browser
such as Microsoft Internet Explorer, with online annotations in the same way as
the traditional paper-based correction approach. The WRITE system can
feedback correct answers, teachers’ comments, and the grammatical error type
for each error to students. In addition, this system can provide users the
annotation marks subject to different query conditions so that the problem of
cognitive overload can be avoided. For error analysis purposes, this system can
access the database and analyzes students’ errors and displays the results as
requested. Students use WRITE will be able to effectively identify more errors.
Moreover, the ways that the corrective feedback delivered through the online
annotation system can be used by students to develop his/her corrective
strategies.
1 Introduction
Writing processes include tasks such as planning, transcribing, and revising (Ogata et
al., 1999). Revision is often defined as the last stage in and the heart of the writing
process. However, it is not an easy task. To most students, revision means correction
(Lehr, 1995). As cited by Lehr, Adams (1991) proposed: “Merely requiring students to
revise or just spend more time revising will not necessarily produce improved writing.”
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 1–8, 2008.
© Springer-Verlag Berlin Heidelberg 2008
2 J.-J. Lo, Y.-C. Wang, and S.-W. Yeh
context” personal notes and can enable asynchronous collaboration among groups of
users. With annotations, users are no longer limited to viewing content passively on the
web, but are free to add and share commentary and links, thus transforming the web
into an interactive medium.
However, in spite of the advantages mentioned above, the question of how
annotations may help students’ writing has not been sufficiently addressed (Wolfe,
2002). Studies investigating using annotation systems in ESL/EFL error feedback and
analysis are especially needed. Based on the above discussion, this study develops an
online annotation system, called WRITE (Writing Revision Instrument for Teaching
English), which can provide annotation analysis and knowledge sharing, and can be
applied to error correction, error feedback, and error analysis in English writing
instruction.
Document Maker is where students input their documents. As the document is edited,
the system will convert it into the HTML format and save it in Document Database so
that it can be displayed with general web page browsers for error correction marking by
teachers.
Annotation Editor (Fig. 2) is where teachers input their correction markings of the
document. It is implemented on the general web browser such as Microsoft Internet
Explorer. In Annotation Editor, teachers can make correction marks and comments
only, i.e., it is under “read-only” status in that the content of the original document
cannot be changed. Such functionality of making correction marks under “read-only”
status is quite important for students to be able to easily compare their original works
and the corrective feedback.
4 J.-J. Lo, Y.-C. Wang, and S.-W. Yeh
To create a correction and comment, the teacher first highlights the text, named
annotation keywords, to which he/she wants to annotate. Then the teacher assigns an
error code by using two pull-down menus to indicate its major error category and error
type. In WRITE, five major error categories are applied: (1) writing style, (2)
composition structure, (3) sentences, (4) words and phrases, and (5) agreement, tense,
and voice. Under each major error category, there are different numbers of error types
(Yeh et al., 2006). After assigning the error type, the teacher clicks on one of the
annotation tools to activate the corresponding function to place the error correction
mark into the annotation keywords. The annotation tools include “Delete”, “Replace”,
“HighLight”, “Insert-Before”, “Insert-After”, and “Move”. Then the WRITE system
will use JavaScript to automatically insert the <SPAN> tag around the annotation
keywords for showing the effects of annotation marks and store all related annotation
information in Annotation Database. As the teacher moves the cursor over the
annotation mark, related annotation information will be shown and the teacher can
delete the correction mark by clicking the “Delete this annotation” button (Fig. 3).
One of the innovative functionalities of WRITE is the teacher can freely switch
between the annotation mode (see Fig. 2) and the review mode (Fig. 4) to neatly review
the “right” document after correction without showing the correction marks (cf. Fig. 2).
In review mode, the annotation tools are hidden.
WRITE: Writing Revision Instrument for Teaching English 5
2.3 Database
Two database modules are included in the system, Document Database and Annotation
Database. Document Database stores the documents students written in HTML format
through Document Maker. Annotation Database stores the related information of
annotations, such as annotator, annotation type, error type, annotation identification
code, annotation notes, etc. In WRITE, a unique annotation identification code is
assigned to each annotation. It can make dynamic control to the annotation keywords
by regarding each annotation as an object stored in the annotation database. Annotation
Database offers the information for annotation query (manipulated by Composer) and
error analysis (manipulated by Error Analyzer).
6 J.-J. Lo, Y.-C. Wang, and S.-W. Yeh
2.4 Composer
Since a document can be annotated with different annotation tools and error types,
students might be confused due to too much information. Through Composer, with
Annotation Database and the annotation identification code, the WRITE system can
provide users the annotation marks subject to different query contitions so that the
problem of cognitive overload can be avoided. In WRITE, a user can query the
annotations based on the annotation type and the error type. In addition, the WRITE
system can implement full-text search within “Annotation keyword”. “Replaced
words”, and/or “Annotation notes” (Fig. 6).
For error analysis purposes, Error Analyzer accesses the database and analyzes
students’ errors to display the statistical results of student error distributions in bar
charts as requested by the teacher. Four error statistical analysis options are included:
Fig. 7. Illustration of Analysis Result of Single Document for an Individual Student (Analyzed
Result Viewer)
WRITE: Writing Revision Instrument for Teaching English 7
single document for an individual student, all documents for an individual student,
single document for a group of students, and all documents for a group of students
(Fig. 7). Error analysis of single document and all documents for an individual student
is helpful to realize the most severe barrier a student faces in writing a particular
document and the overall most severe barrier a student faces in writing. On the other
hand, error analysis of single document and all documents for a group of students is
helpful to realize the errors most students have made in writing a particular document
and the overall errors most student faces in writing.
2.6 Viewer
Two Viewers are included in the system, Document Viewer and Analyzed Result
Viewer. Document Viewer is where students can view their documents after being
corrected by the teacher. Like Annotation Editor, the student can freely switch between
the annotation mode to view correction marks and the review mode to neatly review the
“right” document without showing the correction marks. It is different from the
Annotation Editor (see Figure 2) in that the annotation tools are hidden for both
annotation mode and review mode. Through Document Viewer, students can know
which parts of their documents are corrected and get detailed error feedback by moving
the cursor over the annotation marks (see Fig. 3). Analyzed Result Viewer displays the
four error analysis options analyzed by Error Analyzer (see Fig. 7).
3 Conclusions
Corrective feedback and error correction are important tasks for ESL/EFL writing
instruction. EFL/ESL learners have great diversities in error correction and feedback
strategies, and a more constructive approach and a more interactive environment for
error feedback and error correction are needed. Research findings showed that
students’ major difficulty in error correction lies in their failure to detect errors. Also,
researchers proposed that error analysis can be reinvented in the form of
computer-aided error analysis, a new type of computer corpus annotation. Annotations
on digital documents are easily shared among groups of people, making them valuable
for a wide variety of tasks, including providing feedback. This study developed a
web-based online corrective feedback and error analysis system called WRITE. With
this system, teachers can make error corrections on digitized documents with online
annotations in the same way as the traditional paper-based correction approach. The
WRITE system can feedback correct answers, teachers’ comments, and the
grammatical error type for each error to students. In addition, this system can provide
users the annotation marks subject to different query conditions so that the problem of
cognitive overload can be avoided. For error analysis purposes, this system can access
the database and analyzes students’ errors and displays the results as requested. Four
error analysis options are included: single document for an individual student, all
documents for an individual student, single document for a group of students, and all
documents for a group of students. Students use WRITE will be able to effectively
identify more errors; moreover, the ways that the corrective feedback delivered through
the online annotation system can be used by the student writer to develop his/her
8 J.-J. Lo, Y.-C. Wang, and S.-W. Yeh
Acknowledgements
This is part of a larger study that has been generously supported by the National
Science Council of Taiwan, R.O.C. (NSC 96-2411-H-033-006-MY3).
References
1. Bargeron, D., Gupta, A., Grudin, J., Sanocki, E.: Annotations for Streaming Video on the
Web: System Design and Usage Studies. WWW 8, Toronto, Canada, 61–75 (1999)
2. Brush, A.: Annotating Digital Documents for Asynchronous Collaboration. Technical
report (2002)
3. Ellis, R.: Corrective Feedback in Theory, Research and Practice. In: The 5th International
Conference on ELT in China & the 1st Congress of Chinese Applied Linguistics, May
17-20, Beijing Foreign Language Studies University, Beijing, China (2007), Retrieved
October 23, 2007 from the World Wide Web:
http://www.celea.org.cn/2007/edefault.asp
4. Lee, I.: L2 writing teachers’ perspectives, practices and problems regarding error feedback.
Assessing Writing 8, 216–237 (2003)
5. Lehr, F.: Revision in the Writing Process. Accessed online (2007/11/6):
http://www.readingrockets.org/article/270
6. Ogata, H., Feng, C., Hada, Y., Yano, Y.: Computer Supported Proofreading Exercise in a
Networked Writing Classroom. In: ICCE 1999 (1999)
7. Skehan, P.: A Cognitive Approach to Language Learning. Oxford University Press, Oxford
(1998)
8. Ware, P.D., Warschauer, M.: Electronic feedback and second language writing. In: Hyland,
K., Hyland, F. (eds.) Feedback in Second Language Writing: Contexts and Issues, pp.
118–122. Cambridge University Press, New York (2006)
9. Wolfe, J.: Annotation technologies: A software and research review. Computers and
Composition 19, 471–497 (2002)
10. Yeh, S.-W., Lo, J.-J., Huang, J.-J.: The Development of an Online Annotation System for
EFL Writing with Error Feedback and Error Analysis. In: ED-MEDIA 2006, Orlando,
Florida, USA, pp. 2480–2485 (2006)
u-Teacher:
Ubiquitous Learning Approach
1 Introduction
Mobile devices are part of our everyday environment and consequently part
of our educational landscape [9]. The current mobile trends in education have
demonstrated that learning no longer needs to be classroom. Current trends
suggest that the following three areas are likely to lead the mobile movement:
m-learning, e-learning and u-learning. There are estimated to be 2.5 billion mo-
bile phones in the world today. This means that this is more than four times
the number of personal computers (PCs), and todays most sophisticated phones
have the processing power of a mid-1990s PC. Even, in a special way, many
educators are already using iPod in their curricula with great results. They are
integrating audio and video content including speeches, interviews, artwork, mu-
sic, and photos to bring lessons to life. Many current developments, just as ours,
incorporate multimedia applications. We allow to educators and students can
create their own content. Therefore, its a great way for educators to create, or-
ganize, and distribute content. In the late 1980’s, a researcher at Xerox PARC
named Mark Weiser [8], coined the term “Ubiquitous Computing”. It refers to
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 9–20, 2008.
c Springer-Verlag Berlin Heidelberg 2008
10 F.F. Zacarı́as et al.
the process of seamlessly integrating computers into the physical world. Ubiqui-
tous computing includes computer technology found in microprocessors, mobile
phones, digital cameras and other devices. All of which add new and exciting
dimensions to learning.
CDC. The Connected, Limited Device Configuration (CLDC) is for small wire-
less devices with intermittent network connections, like mobile phones, and per-
sonal digital assistants (PDAs). The Mobile Information Device Profile (MIDP),
which is based on CLDC, was the first finished profile and thus the first finished
Java ME application environment. MIDP-compliant devices are widely available
worldwide.
On the other hand, The Connected Device Configuration (CDC) is for larger
devices (in terms of memory and processing power) with robust network con-
nections, such as set-top boxes, Internet appliances, and embedded servers. The
NetBeans IDE provides a wizard that enables you to quickly create a MIDP
project. When creating the project, you can choose to develop your application
in the Visual Mobile Designer (VMD) or in the Source Code Editor. The visual
mobile designer’s use allows you the ability to graphically plan out the flow of
the application and design the screens the application will use. The designer
automatically creates the code for the application.
Finally, our application has an intelligent agent whose tasks are: to evaluate
the answers of the test either received through the web or, through messages
via cell phone. As second task this the supervision of the correct update of new
lessons, exams, homeworks, etc. The integration of all these features make our
methodology doesnt just work, it works well! This ensures the efficiency of our
system in the ubiquitous learning, as shown in the rest of paper.
4 u-Teacher Model
Our proposal is the combination of mobile technologies and web technologies. In
Fig. 3 we can appreciate the main interface our application and it is important
to point out that all interfaces developed for the web are also available for the
cell phone. Even, we can assert that this is the feature that has motivated and
increased the level of learning in our students.
14 F.F. Zacarı́as et al.
– Read a book in English for an hour every day, analyzing the grammar in
sentences and looking up words in an English dictionary.
– Listen to an audio-book or other recording in English, stopping it frequently,
trying to understand what is being said, and trying to imitate the speaker’s
pronunciation.
– Spend your afternoon practicing the pronunciation of the English ”r” sound.
– Carefully write an e-mail message in English, using a dictionary or a Web
search every 20 seconds to make sure every word is correct, and taking 5
minutes to write one sentence.
– Think about an English sentence you’ve read, wondering if it could say ”a”
instead of ”the” in the sentence, and trying to find similar sentences on the
Web to find out the answer.
– Walk down the street and build simple English sentences in your head (talk-
ing to yourself in English about the things you see around you).
By virtue of these actions we have implemented lessons that cover these as-
pects (Fig. 4 and Fig. 5). These characteristics have proven to give excellent
results in the use of our tool. these results obey to that learners could make use
of his course of English in anywhere and anytime. Our tool has been considered
as part of a university initiative to encourage creative uses of technology in ed-
ucation and campus life. The problem with learning and teaching English as a
foreign language is that all English learners want to speak English well; however,
most learners don’t want to spend time on learning English on their own. (Why
they sign up for English classes and hope their teacher will force knowledge into
their heads?.) This lack of motivation means that learners basically don’t spend
16 F.F. Zacarı́as et al.
their own time on learning English, and if they do, they don’t do it regularly. For
example, a typical learner might study English phrasal verbs for 12 hours before
an English exam. However, he will not read a book in English for 30 minutes
every day. He just doesn’t feel that learning English is pleasant enough, so he
will only do it if he has to. The problem is that a huge one-time effort gives you
nothing, while small, everyday activities will give you a lot. If you are one of
those learners and don’t feel like practicing the pronunciation of the ”r” sound
or thinking about English sentences every day, we have news for you: You’re
going to have to make yourself want to do these things. In other words, you’ll
have to work on your motivation. Fortunately, there are proven techniques to
help you with that.
Some other basic things we must remember to take into account when learning
another language are:
– No mistakes: Avoid mistakes. Try to use the correct form of the language
from the beginning.
– Pronunciation: Learn to pronounce that language sounds. Learn to under-
stand phonetic transcription and the phonetic alphabet.
– Input: Get the language into your head by reading and listening to lots of
sentences in that language, you could read or watch movies.
Language is primarily a spoken form of communication. We learn our native
language as children by hearing the spoken language and then imitating it. This is
something often overlooked. I believe that the most successful language learning
methods are more audio-based than otherwise. In fact, I learned English that
way. We can’t ignore the importance of reading, but clearly the most fundamental
aspect of communicating in a language is speaking and listening.
Videos are another important part that supplements the actions that should
be carried out (see Fig. 5). While more we listen conversations and sentences in
that language more quick will advance in our learning. Videos, movies and music
stimulate the auditory sense (such as in Fig. 4 and Fig. 5); this allows making
an imitation of the pronunciation more clear and correct. This portal tends to
improve the users knowledge of the English language throughout practicing with
exercises that will be asked to complete. If the user has little or null knowledge
of this language, he shall obtain the bases to be able to begin to speak it and to
understand it. In particular, we have incorporated the reproduction of videos to
our mobile devices with the objective of offering to learner the opportunity to
practice in anytime and anywhere taking advantage of the mobile technology.
With respect to the test, we have implemented an agent whose task is the
one of validating that the answers are correct. Here, the learner answers the
test offline (anytime or anywhere) and then he chooses the upload option. The
answers are sent to the portal through a file (text message) so that they are
evaluated by our agent.
The result of the test evaluation is sent to the apprentice through a text
message. With respect to Chat, this it is available through the interactive portal
only. With this characteristic we provide to learner with a mobile tool that
offers bigger educational opportunities. u-Teacher contains an amusing game that
absorbs the learner in the subconscious learning of such an extensive vocabulary
as it is desired. It is important to point out that in this first version we have
added only the hangman game. However, due to the high index of interest shown
by u-Teacher’s users regarding this game, we are developing new games that the
same as it happened with this they capture the attention of learners.
In [6] the authors describe five properties of handheld computers that produce
unique educational affordances:
– Portability - can take the computer to different sites and move around within
a location
– Social Interactivity - can exchange data and collaborate with other people
face to face
– Context Sensitivity - can gather data unique to the current location, envi-
ronment, and time, including both real and simulated data
18 F.F. Zacarı́as et al.
– Convenience for both faculty and students of portable digital course content,
and reduced dependence on physical materials.
– Flexible location-independent access to digital multimedia course materials,
including reduced dependence on lab or library locations and hours.
– Effective and easy-to-use tool for digital recording of interviews, field notes,
small group discussions, and self-recording of oral assignments.
– Greater student engagement and interest in class discussions, labs, practices
outside of class, and bigger concentration.
– Enhanced support for individual learning preferences and needs.
7 Conclusions
In this paper we have presented the impact that our called tool u-Teacher has
had in the teaching of the English language. u-Teacher uses ubiquitous tech-
nology and the concept of ubiquitous learning. This is part of ongoing research
and development being undertaken at Autonomous University of Puebla. The
research team is extensively involved in ubiquitous technology and communica-
tions. Innovation and experimentation for academic u-Teacher use was widely
reported as well, with 75% of firt-year students reporting having used at least
one u-Teacher feutured in a class or for independent support of their studies. In
u-Teacher: Ubiquitous Learning Approach 19
addition to the findings outlined above regarding academic u-Teacher use, the
evaluation also identified some significant institutional impacts of the project:
Acknowledgments
Thank you very much to the Autonomous University of Puebla for their financial
support. This work was supported under project VIEP register number 15968.
Also, we thank the support of the academic body: Sistemas de Informacin.
References
1. Chen, Y.S., Kao, T.C., Sheu, J.P., Chiang, C.Y.: A Mobile Scaffolding-Aid-Based
Bird -Watching Learning System. In: Proceedings of IEEE International Workshop
on Wireless and Mobile Technologies in Education (WMTE 2002), pp. 15–22. IEEE
Computer Society Press, Los Alamitos (2002)
2. Curtis, M., Luchini, K., Bobrowsky, W., Quintana, C., Soloway, E.: Handheld Use
in K-12: A Descriptive Account. In: Proceedings of IEEE International Workshop
on Wireless and Mobile Technologies in Education (WMTE 2002), pp. 23–30. IEEE
Computer Society Press, Los Alamitos (2002)
3. Fischer, G.: User Modeling in Human-Computer Interaction. Journal of User Mod-
eling and User-Adapted Interaction (UMUAI) 11(1/2), 65–86 (2001)
4. Jacobs, M.: Situated Cognition: Learning and Knowledge Relates to Situated Cog-
nition (1999) [verified 31 Oct 2004]
http://www.gsu.edu/∼ mstswh/courses/it7000/papers/situated.htm
5. Jones, V., Jo, J.H.: Ubiquitous learning environment: An adaptive teaching sys-
tem using ubiquitous technology. In: Atkinson, R., McBeath, C., Jonas-Dwyer, D.,
Phillips, R. (eds.) Beyond the comfort zone: Proceedings of the 21st ASCILITE
Conference, Perth, December 5-8, pp. 468–474 (2004),
http://www.ascilite.org.au/conferences/perth04/procs/jones.html
6. Klopfer, Squire, Holland, Jenkins: Environmental Detectives-the development of an
augmented reality platform for environmental simulations, Educational Technology
Research and Development. Springer, Heidelberg (2002)
20 F.F. Zacarı́as et al.
7. Zacarias, F.F., Lozano, T.F., Cuapa, C.R., Vazquez, F.A.: English’s Teaching Based
On New Technologies. The International Journal of Technology, Knowledge & So-
ciety, Northeastern University in Boston, Massachussetts, USA, Common Ground
Publishing, USA (2008)
8. Weiser, M.: The computer for the twenty-first century, pp. 94–104. Scientific Amer-
ican (September 1991)
9. Zacarı́as, F., Sánchez, A., Zacarı́as, D., Méndez, A., Cuapa, R.: Financial Mobile
System Based On Intelligent Agents in the Austrian Computer Society book series,
Austria (2006)
A Model for Knowledge Innovation in Online
Learning Community
Qinglong Zhan
1 Introduction
With the development of e-learning, online learning community (OLC) plays a
more important role in e-learning. OLC is not only the inevitable result of de-
velopment of e-learning, but a basic component of it as well. Through utilizing
modern information technology such as Internet that offers learning environ-
ments of whole new communication mechanisms and abundant resources, OLC
is an important environment of knowledge innovation which can realize pro-
duction, sharing, application and innovation of knowledge. Since 1990s, for the
research on OLC, many researchers centralized on concept, mechanism of for-
mation and growth, knowledge construction/building etc., but the research on
knowledge innovation is an omitted paradigm.Therefore, this paper focuses on
introducing mechanism of knowledge innovation into OLC and constructing a
model which can facilitate innovative knowledge and develop online learner’s
ability of knowledge innovation.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 21–31, 2008.
c Springer-Verlag Berlin Heidelberg 2008
22 Q. Zhan
2 Conceptual Background
2.1 OLC
network, facilitates the learner acquiring and sharing knowledge from other ex-
perienced learners by informal learning. It is an ideal environment of innovative
process which is networking, interacting and knowledge-driving. So, knowledge
sharing and innovation are two key activities in OLC. Knowledge innovation is
promoted by knowledge sharing in which learners can obtain new elicitation,
thinking or inspiration that provides the possibility of creating new knowledge.
Internalization
Social interaction
ZPD
Knowledge
sharing
Knowledge
Knowledge
integration
comparison
Combination Socialization
Knowledge Knowledge
negotiation creation
Knowledgebase
Cognition connection
Externalization
information for making it more personal, or to build and structure nodes and
links, thereby forming a network of ideas in the knowledge base”(Jonassen, 2000).
that will be accumulated, thus each learner could share more knowledge. But-
ler(2001) noted that “the knowledge sharing activity is an important construct
for explaining the dynamics of a virtual community since without some forms of
community knowledge sharing activity, any virtual community will fail to sur-
vive”. Reinforcing OLC will offer the ideal chance for the learner to put into
knowledge sharing, to keep OLC knowledge being in the state of activating, to
develop tacit and explicit knowledge so as to transmit other learners. Cummings
(2003) identifies five primary contexts that can affect such successful knowledge
sharing implementations, including: the relationship between the source and the
recipient,the form and location of the knowledge, the recipient’s learning predis-
position, the source’s knowledge sharing capability, and the broader environment
in which the sharing occurs. Generally, there are several ways of sharing knowl-
edge as following:(1)utilizing the forum, chat-room, electronic meeting, E-mail
to exchange;(2)question-leading tacit knowledge sharing, namely OLC learners
share tacit knowledge on the particular issue and artifacts or communicate their
thought processes; and (3) Sharing the produced new opinion after learning from
course content and knowledgebase.
various fields, occurs when the learner gives up or revises initial beliefs and cor-
rects one’s own view based on other learners’ contributions. An indication for
integration-oriented consensus building is that“participants show a willingness
to actively revise or change their own views in response to persuasive argu-
ments”(Keefer, Zeitz and Resnick, 2000). Conflict-oriented consensus building
is that learners need to more closely operate on the reasoning of others instead
of simple acceptance of their contributions, and need to pinpoint out specific
aspects of other learners’ contributions and modify them or present alternatives.
5 Conclusions
References
1. Afhuah, A.: Redefining firm boundaries in the face of the Internet: Are firms really
shrinking? Academy of Management Review 28, 34–53 (2003)
2. Butler, B.S.: Membership Size, Community Activity, and Sustainability: A
Resource-based Model of Online Social Structures. Information Systems Re-
search 12, 346–362 (2001)
3. Consway, B., Whittingham, V.: Managing Knowledge and Learning at Unipart.
Knowledge Management Review 4, 14–17 (2001)
4. Cummings, J.: Knowledge Sharing: A Review of a Literature (2003),
http://lnweb18.worldbank.org/oed/knowledge eval literature review.pdf
5. Fahey, L., Prusak, L.: The Eleven Deadliest Sins of Knowledge Management. Cal-
ifornia Management Review 40, 265–276 (1998)
6. Gray, J.H., Densten, I.L.: Towards an Integrative Model of Organizational Culture
and Knowledge Management. International Journal of Organizational Behavior 9,
594–603 (2005)
7. Jonassen, D.H.: Computers as Mind Tools for Schools: Engaging Critical Thinking,
2nd edn. Prentice Hall, New Jersey (2000)
8. Kearsley, G., Lynch, W.: Structural Issues in Distance Education. Journal of Ed-
ucation for Business 71, 191–196 (1996)
9. Keefer, M.W., Zeitz, C.M., Resnick, L.B.: Judging the Quality of Peer-led Student
Dialogues. Cognition and Instruction 18, 53–81 (2000)
10. Kreijns, K., Kirschner, P.A.: The Social Affordances of Computer Supported Col-
laborative Learning Environments. In: 31th ASEE/IEEE Frontiers in Education
Conference, Reno, Nevada, USA (2001)
11. Nonaka, I., Toyama, R., Konno, N.: SECI, Ba and Leadership: a Unified Model of
Dynamic Knowledge Creation. Long Range Planning 33, 5–34 (2000)
12. Puntambekar, S., Nagel, K., Hbsher, R., et al.: Intra-group and Intergroup: An
Exploration of Learning with Complementary Collaboration Tools. In: Hall, R.,
Miyake, N., Enyedy, N. (eds.) Proceedings of Computer-supported Collaborative
Learning, Toronto, Canada, pp. 207–214 (1997)
13. Scardamalia, M., Bereiter, C.: Student Communities for the Advancement of
Knowledge. Communications of the ACM 39, 36–37 (1996)
14. Schellens, T., Valcke, M.: Collaborative learning in asynchronous discussion groups:
What about the impact on cognitive processing? Computers in Human Behavior 21,
957–975 (2005)
15. Schwen, T.M., Kalman, H.K., Hara, N., et al.: Potential Knowledge Management
Contributions to Human Performance Technology Research and Practice. Educa-
tional Technology Research and Development 46, 73–89 (1998)
16. Seufert, S.: Design and Management of Online Learning Communities (2002),
http://www.aib.ws.tum.de/euram/seufert paper.pdf
17. Sorensen, C., Lundh-Snis, U.: Innovation through Knowledge Codification. Journal
of Information Technology 16, 83–97 (2001)
18. Starbuck, W.H.: Learning by Knowledge Intensive Firms. Journal of Management
Studies 29, 713–740 (1992)
19. Trentin, G.: From Formal Training to Communities of Practice via Network based
Learning. Educational Technology 41, 5–14 (2001)
20. Weinberger, A., Fischer, F.: A Framework to Analyze Argumentative Knowledge
Construction in Computer-Supported Collaborative Learning. Computers & Edu-
cation 46, 71–95 (2006)
The Design of Software Architecture for E-Learning
Platforms
1 Introduction
E-Learning is just-in-time education integrated with high velocity value chains[7]. It is
the delivery of individualized, comprehensive, dynamic learning content in real time,
aiding the development of knowledge communities, linking learners and practitioners
with experts. Generally, e-learning improves the flexibility and quality of the education
by[6] :
- providing access to a range of multimedia resources, such as graphics, sounds,
animations and videos;
- supporting the reuse of high quality and expensive resources;
- supporting increased communications between instructors and students and
between students;
- enabling instructors to provide different materials to the students from different
backgrounds;
- encouraging students to choose materials according to their own interests and to
study at their own pace;
- encouraging students to take responsibilities for their own studies.
So far, a number of research teams have implemented different kinds of Web-based
education platforms to support learner-centered, interactive and active learning. The
Web is used not only as a delivery medium but also to foster free exploration of
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 32–40, 2008.
© Springer-Verlag Berlin Heidelberg 2008
The Design of Software Architecture for E-Learning Platforms 33
learning materials and to allow the learners to interact with materials, instructors and
other learners. Almost all the systems fall into three categories: Courseware developing
System(such as The Geometers Sketchpad, and Authorware), teaching supporting
system(such as Blackboard[1]) and learning supporting system(such as Web CT[2]) ,
and web-based educational resources system or portal (such as NGFL). Although those
e-learning systems provide versatile functions to support e-learning and are widely
used by schools, universities. They, however, have the following shortcomings:
- Current e-learning systems are teacher-centered. They are designed to facilitate
teaching activities but not consider the learners’ diverse learning goals and demands.
The learning process is not well-supported.
- Most e-learning systems are not developed for customization. They cannot meet
all requirements of instructors and learners in their teaching and learning activities.
Meanwhile, most instructors of K-12 are not good at developing software, and
cannot build a complicated software system on their own.
- Most e-learning systems are not structured in a perfect architecture, and their
elements are tightly coupled. So there are a lot of problems such as redundant
development, difficulties to integrate with others and maintenance difficulties
among e-learning software products.
Because of these shortcomings, current e-learning systems have limited applications
in supporting real world teaching & learning. In our e-learning platform, we are
engaged in improving the performance of e-learning through building a flexible and
hierarchical reusable dynamical application architecture and present a service-oriented
approach of software components integration. Section 2 introduces the architecture
design of our Platform. Section 3 specifies the method of integrating different software
components. Section 4 specifies an application of our software architecture for
e-learning platform to develop a web-based special topic learning website. Finally, the
paper is concluded in Section 5.
(1) Product line layer: It mainly includes five SPL, which are e-teaching supporting
SPL, e-learning supporting SPL, education information managing SPL, education
resource development SPL and education resources management and service SPL. This
layer is responsible to provide reusable domain solution of e-learning, describe the
functional and non-functional requirements of components which composed the
various product line architectures and the associate relation between these components,
and realize the building of specific software application system. This layer is the core of
The Design of Software Architecture for E-Learning Platforms 35
the platform software architecture. Each product line is designed for a specific goal of
domain application and can support a group of buildings of similar application systems.
Through selecting and assembling different frameworks and components, Instructors
can quickly instantiate a customized software system.
(2) Application Layer: This layer is a specific group of e-learning software systems
which are built via different product lines.
(3) Application framework layer: The layer is the collection of the public
frameworks and domain specific frameworks. Each framework is composed of a series
of related components, component associated relations and restrictions. It is the
collection of associated components which solute certain sub problem of the product
line. The public framework mainly includes UI management, access control
management, user management, workflow engineer, system management, all of which
are common frameworks to all the product architecture. Meanwhile, the domain
specific frameworks are designed according to different teaching&learning activities
design. It covers teaching&leaning plan design, and contents presentation, and exercise
management, and homework management, and examination management, and
question&answer management, and assessment management, and learning resource
development, and learning resource management, and educational administration
information management. By introducing the application framework layer, we reduce
the complexity of the architecture of product line and increase the reusable efficiency
greatly.
(4) Component library layer: This layer is the foundation of the platform software
architecture. It is a reusable collection of software units which is already validated by
other projects. It mainly includes the component library of educational resources
development, and the component library of educational resources management, and the
component library of building teaching&learning environment, and the component
library of educational information management, and the component library of
cooperation, and the common component libraries for software system management.
Moreover, the component library layer is extensible and distributed. As long as the
third-party components meet with the certain agreement of application framework, they
can be added to the component library layer.
build the object models for data persistent from XML structures. Based on XML data
bus, each component (encapsulated into web service) only read and write the XML
data to accomplish its business logic without the need of communicating with others
component. The XML data bus is controlled by a data bus controller, which is in
charge of Marshalling and Un-Marshalling. Marshalling refers to the process that
takes an object or a tree of objects, and creates the XML representation that will
record its state. It’s a kind of serialization.Un-marshalling is just the opposite, it takes
the XML representation and builds an object model from it for object persistent. Once
this data bus is in place, it is easy and quite straightforward to integrate the different
domain components and weave them into a black-box framework, which is
dynamical and compatible and supporting of thermal plugging. Moreover, on
inheriting the characteristics of Web Services, the purpose of loosely coupled and
independence of programming languages and operating system also be reached. Thus
users can only concentrate on the business logic and all the other issues in their
domain project without having to concern themselves with the technical details of
application development. Figure 2 shows the principle of component integration and
framework weaving based on XML data bus.
Website Information
Management
Select Website
Presenting template Knowledge Tree Edit
Application to Construct
A Special Topic Learning Contents Plan Learning Strategies Plan
Learning Website
(Select web-based
domain frameworks)) Website Generation
Learning Contents Edit
Website Publishing
Fig. 5. Learning Contents Presentation Page of Success English Special Topic Learning Website
5 Conclusion
In this paper, flexible and hierarchical reusable software architecture for e-learning
platforms is introduced. It consists of four layers: the application layer, the product line
architecture layer, application framework layer, and component library layer. These are
responsible for building customized e-learning systems just like building block
40 D. Zhou et al.
References
[1] Blackboard (2005), http://www.blackboard.com/
[2] WebCT (2005), http://www.webct.com/
[3] Mattson, M.: Evolution and Composition of Object-Oriented Frameworks, printed in
Sweden, KaserntryckerietAB, Karlskrona (2000)
[4] Wenhui, H., Wen, Z., ShiKun, Z.: Study of Application Framework Meta-Model Based on
Component Technology. Journal of Software 15(1) (2004)
[5] Long, W., Zhong, S., Zhou, D.: A Distance Education System Based On Web Services.
Journal of Computational Information Systems 2(1), 139–144 (2006)
[6] Zhou, D., Cheng, X., He, X.: The Development of a Customized E- Learning System.
Journal of Computational Information Systems 2(1), 211–216 (2006)
[7] Ducker, P.: Need to Know: Integrating e-Learning with High Velocity Value Chains, A
Delphi Group White Paper (2000), http://www.delphigroup.com/pubs/
whitepapers/20001213-e-learning-wp.pdf
[8] Anastasopoulos, M., Atkinson, C., Muting, D.: A Concrete Method for Developing and
Applying Product Line Architectures. In: Aksit, M., Mezini, M., Unland, R. (eds.) NODe
2002. LNCS, vol. 2591, pp. 294–312. Springer, Heidelberg (2003)
An Educational Component-Based Digital TV
Middleware for the Brazilian’s System
Abstract. One of the major problems faced by the Brazilian population is the
low level of the fundamental schools. Television is the most popular source of
entertainment and information of the Brazilian population being present in
approximately 54 million families all over the country. These families watch
television for more than 8 hours daily. Moreover, at this moment, the Brazilian
TV system is moving from analog to digital. That means not only that image
and sound will be delivered with much better quality but also that it will be
possible to send interactive multimedia programs, creating a brand new way of
watching TV. That is in fact the main novelty of the digital system it will be
possible to offer personal interactive services such as banking, games and most
importantly educational programs. This work introduces a software framework
called “Extended Middleware for Digital TV (EMTV)” which is suitable for the
generation of interactive applications executed over digital television systems.
Its concept was developed focusing on the Brazilian technological options for
Digital TV. Technically, EMTV is a procedural GEM compliant application
which, from the programmer’s point of view, acts as a declarative middleware
extension. The framework was developed to be component-based in order to
minimize the need for programming knowledge to deploy the digital TV
applications using EMTV. The main goal of the platform is to facilitate the
construction of interactive multimedia educational applications, a crucial field
for the Brazilian population. The concept is tested and validated by the
construction of a Quiz application presented at the end of the paper.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 41–51, 2008.
© Springer-Verlag Berlin Heidelberg 2008
42 J.R. Costa and V.F. de Lucena Junior
It is only in 1998 that Brazil started to research DTV technology and initially
decided to develop its own standard which used to be called SBTV (Brazilian TV
System) and whose main characteristic is the use of an OFDM (Orthogonal
Frequency-Division Multiplexing) modulator system equipped with an artificial
intelligence module on the reception side to make the receiver multipath-noise robust.
Despite the good results achieved the a few years ago, the Brazilian government
showed clearly that it was giving in to the pressure of the TV content providers when
it decided to adopt the Japanese ISDB (Integrated Services Digital Broadcasting)
standard. At that time, the Brazilian committee renamed the standard to ISDTV [1]
(International System for Digital TV) on account of some local contributions. As a
matter of fact this model, already in use in Brazil since last December 2007,
consolidates important aspects of the Japanese technology such as the use of the BST-
OFDM (Band Side Transmission – Orthogonal Frequency-Division Multiplexing)
modulation which is very effective against multipath noise even with fast-moving
mobile receivers1. The Brazilian contributions [2] relate to the use of the MPEG-4
AVC standard also known as MPEG-4 “Part 10” or H.264 which is an evolution of
the MPEG-2 standard primarily used in ISDB because it achieves compressed audio
and video rates 40% to 70% 2 higher than those of the MPEG-2. With such
characteristics, and considering that the Japanese standard committee is aggregating
the proposed changes, the Brazilian model was recently defined to be referred to
internationally as International ISDB.
Another Brazilian contribution to ISDB refers to the development of its DTV
middleware specification called GINGA [3]. Just like other DTV middleware it has a
procedural part, the GINGA-J (Ginga – Java) [1], and a declarative part known as
GINGA-NCL [1] (Ginga – Nested Context Language).
1
HDTV (1920x1080i 16:9) in fixed terrestrial TV receivers. LDTV (320x240 4:3) in fast
moving mobile receivers.
2
Depending on the MPEG-4 profile and on the nature of the images.
An Educational Component-Based Digital TV Middleware 43
fact that GINGA-J has not yet been officially released it will probably be a GEM
implementation. GINGA-NCL is the declarative part of GINGA and is considered to
be very powerful and flexible as it not only controls appearance and the positioning of
media objects but also considers the temporal relationships between them [1].
Terrestrial open DTV transmissions started in Brazil without any middleware
support. According to the local industry, this was due to the need to the middleware to
support the H.264 standard and due to the wait for official GINGA specifications.
Anyway, even without the support to any type of middleware, the innovations used by
the Brazilian system made both the signal generation equipment and the set-up boxes
very expensive, a fact likely to be a problem for the popularization of the technology
in the country.
Nevertheless, the new interactive features of the digital system are expected to
have great impacts on the population, the most significant being the use of this
technology as a tool to contribute to educational processes. This paper introduces a
component-based framework, named “Extended Middleware for Digital TV
(EMTV)”, which was developed at a time when the GINGA was not even available
for public download and whose main target is to help TV-content providers with no
advanced programming skills to deploy DTV applications for educational purposes
more specifically. The EMTV is free for any use and does not require any expensive
tool since it thereby hopes to contribute in democratizing DTV technology in Brazil.
Hardware
Framework STB
interactive applications. The EMTV framework uses an external file which format is
defined in the framework. The handling of this file demands the use of any of the
third-part text editors referred to as authoring editors in Figure 2.
Interactive applications have already been used in several countries where DTV
technology has been implemented. As the range of applications increases, new terms
like T-commerce, T-government and T-learning and others have been created to
classify it according to its main purpose. The term “T-learning” [6] refers to the use of
DTV technology in educational processes, one of the main purposes of EMTV.
T-learning is viewed as the convergence between DTV technology and E-learning
3
Starting from MHP 1.0.3 version the XLet applications can be loaded directly from the return
channel connected to a TCP/IP network if available.
An Educational Component-Based Digital TV Middleware 45
T-Learning T-Learning
FORMAL INF
INFORMAL
T-Learning QUIZ T-Learning
T-L
Non-Formal EDUTAINMENT
E
EDUTAI
Non T-learning
arning applications
which basically refers to the use of computational technology for training or any other
educational activity. It comes from “E-learning” the popularization of Quizzes as a
very effective educational tool.
Walldén and Soronen [10] wrote a paper about relationship between e-learning and
t-learning which classifies education processes in four classes: Formal learning (leads
to acknowledge diplomas), Non-formal learning (education from formal institutions
but that does not provide official diplomas), Informal learning (education which
comes from social activities) and Accidental learning also known as edutainment
(education + entertainment: the knowledge or skill is acquired not intentionally). A
Quiz is a kind of test containing a series of questions along with some alternatives
which the users can select according to their skills, knowledge or personal opinions.
This principle allows Quizzes to be used not only in any educational process,
typically in edutainment and informal processes, but also in other non T-Learning
applications.
The need of an external file is already part of the declarative approach. It allows
EMTV to be easily reused, as different configuration files generate different
applications with a different content and behaviour. The EMTV configuration file is
based on the XML format which furthers three main advantages at least:
• XML is easily read and written by both humans and algorithms.
• XML can be validated through DTD files.
• Several free XML editors can be used for edition and validation.
Another advantage of the use of the XML format for the configuration file is that it
is very convenient to describe the software components properties. Software
components are artifacts constituted by one or more instances of classes, independent
enough to provide some interest functionality such as visual, behavioural or both. The
behaviour and characteristics of a software component are defined by the properties it
makes available through public methods which constitute the interface of the
component. There are many good software-engineering proven reasons [11] to design
components-based software. EMTV is component-based mainly because it
contributes to an easy system reuse, maintainance and expansion.
46 J.R. Costa and V.F. de Lucena Junior
External
XML Config file EMTV
x Interface components +
description. Image 1
GEM
x Content Image N middleware
x Positioning
(STB)
Fig. 4. The EMTV configuration files provides enough information to generate the applications
3 EMTV Components
The first version of EMTV provides only the minimal set of components to allow the easy
deployment of TVD Quiz applications with the possibility to use or not a return channel.
Those components were defined from a study of similar systems, such as MOODLE [8]
which has a module to generate Quiz applications typically for the Internet.
Application Screen
Application Text
Fig. 5. Quiz application interface that helps to identify graphical components for EMTV
4
EMTV uses nanoXML library to be able to read the XML file attributes. This library is a very
small and efficient XML parser for Java and can be found at http://nanoxml.cyberelf.be/.
An Educational Component-Based Digital TV Middleware 47
allows the programmer to control the size, position and appearance of the application.
One of its attributes indicates if the screen background should be filled with an
external picture, giving to the application an enhanced appearance.
It is a graphical component that is used to represent static images over the screen. It is
made up by HAVI, DVB and JMF libraries but mainly by the instantiation of the Java
Image Class. The attributes allow the programmer to indicate an external image and
controls its position on the screen. The loading method of the image is done by JMF
through the classes java.awt.MediaTracker and java.awt.Toolkit. Just like the
Application Text, there is a field to control the visibility of the component. The
programmer can place as many Application-Image components as needed.
This is a singleton component which enables EMTV to build Quiz applications. Its
fields provide all the information to create a multi-page Quiz. Each page contains one
question, represented by an HText Havi instance, and multiple alternatives for this
question. The programmer can define the position, the foreground colour, the
background colour of the question whose text is updated every time the user changes
the question page through the remote control left and right arrows. The programmer
can indicate a different number of alternatives for each question and if each question
has a single or multiple choices answer. Fields control the initial position for the first
alternative of all questions as well as the relative increment on the x and y axis for the
48 J.R. Costa and V.F. de Lucena Junior
next available alternatives. Other fields indicate the foreground colour for the
alternatives, the foreground colour used as the user navigates through the alternatives,
using the remote control up and down arrows, as well as a field to indicate an external
picture which will be placed close to the chosen alternatives by pressing on the
ENTER key of the remote control. Each alternative has special attributes which
allows the programmer to define if the alternative is in fact an edit box where the user
can input alpha-numeric, numeric or password characters when the alternative, by
default represented by an org.havi.ui.HText is replaced by an org.havi.ui.
HSinglelineEntry instance. This last situation is useful for the Quiz to get specific
user information, like the username to be registered on an external server if an
Application Communication component is used. In this case, an alternative can be
converted into a text instruction for the edit box if the selectable attribute for an
alternative is indicated as false. The component always inserts an additional
configurable question which allows the user to indicate he/she has given the answers.
This event disable the possibility of user change its responses and also starts the
sending of the collected information in case of the presence of an Application
Communication component. After this event the component changes the background
color of all alternatives marked with the “isAnswer” attribute allowing the user to
check his score.
EMTV was tested and validated at the Ceteli DTV laboratory of UFAM as shown in
Figure 7 and 8. The process used an external Apache server connected to the same
TCP/IP network as an interactive profile MHP Set-top box (STB). The server is used
to provide the EMTV packet application as requested by the STB. Once loaded in the
STB, the EMTV requests the server the configuration file and all necessary picture
files to generate the Quiz to be presented on the HDTV monitor. The user interacts
with the application though the STB remote control and once he/she has finished, the
information gathered is sent to an external Apache/PHP server. In the case of this test,
the server used to provide the application was also used to register the users’
responses.
An Educational Component-Based Digital TV Middleware 49
HDTV Monitor
remote
QUIZ control
Apache + PHP Server Application
Return
channel EMTV.jar+ configuration
TCP/IP
STB
(provides EMTV.jar and MHP Interactive Profile
register the users responses) STB requests:
http://serverIP/EMTV.ait + configuration + EMTV
file + pictures files
http://serverIP/quizresp.php
The same results could be achieved loading the EMTV application directly from
the DTV data channel. To validate EMTV on this case would be necessary, besides
the proper hardware equipment to generate the MPEG-2 Transport Stream signal and
the DSMCC (Digital Storage Media Command and Control) system, to correctly
setup the PMT (Program Map Table) and the PAT (Program Association Table) so
the STB can detect and load EMTV. This test would also require a minimal change on
the software so EMTV would be able to download any external files through the
DSMCC synchronously or asynchronously.
50 J.R. Costa and V.F. de Lucena Junior
Fig. 8. Pictures of a Quiz application generated by EMTV running in a real DTV system
5 Conclusion
Since last December the Brazilian TV broadcasters start to transmit their signals in
digital form. Many good things have been written about this fact and the local
population is anxious for taking advantage of this new technological facet of a very
popular entertainment vehicle. More than a better image and sound it is expected to
make good use of the interactive possibilities available with the new digital system.
Knowing the big deficiency in education all over the country and the cultural
relationship with TV we can assure that the new digital TV system will certainly play
a very important role in the generations to come. Certainly there will be lots of
multimedia interactive applications available for usage in a near future. That is
exactly the main contribution of the EMTV middleware for interactive digital TV
systems presented in this work.
EMTV is a very simple to use and already capable to generate a series of useful
multimedia applications for educational purposes. It has the advantage of neither
requiring any broad knowledge on DTV systems, nor the purchase of expensive
proprietary tools, and thereby encourages the popularization of a generation and use
of interactive DTV software.
EMTV is available for free, for any use, by sending an e-mail request to the
authors. The following steps of this work will include the development of a graphical
authoring tool, as well as the development and improvement of new graphical
components.
An Educational Component-Based Digital TV Middleware 51
Acknowledgements
We would like to thank you very much the students involved in the development of
this work. Many thanks for the whole CETELI/UFAM staff for giving the laboratorial
support.
References
1. Soares, L.F.G., Souza, G.L.: Interactive Television in Brazil: System Software and the
Digital Divide. In: EuroiTV (2007)
2. Souza, G.L.: Standard 06 - ISDTV-T Data Codification and Transmission Specifications
for Digital Broadcasting, Volume 4 – GINGA-J: Environment for the execution of
procedural applications. São Paulo, Brazil. ISDTV-T Forum (2006)
3. Middleware GINGA Web Site (2008), http://www.ginga.org.br/
4. Interactive TV Web Site (2008), http://www.interactivetvweb.org
5. César, P.: A Graphics Software Architecture for High-End Interactive TV Terminals,
Helsinki University of Technology (2005) ISBN 951-22-7888-X
6. Bates, P., Atwere, D.: Interactive TV: A learning platform with potential. Learning and
Skills Development Agency (2003)
7. John, J.: DVB/MHP Java TVTMData Transports Mechanisms (2002)
8. Moodle Course Management System Home Page (2008), http://moodle.org/
9. Walldén, S., Soronen, A.: Edutainment. From Television and Computers to Digital TV
(2004)
10. Johnson, K., Hall, T., O’Keeffe, D.: Generation of Quiz Objects (QO) with Quiz Engine
Developer (QED) (2005) ISBN: 0-7695-2385-4
11. Heineman, G.T., Councill, W.T.: Book: Component-Based Software Engineering, 1st edn.
Addison-Wesley, Reading (2001)
12. MHP Organization Web Site (2008), http://www.mhp.org
13. Java TVTM Web Site (2008), http://java.sun.com/products/javatv/index.html
Designing and Developing Process-Oriented Network
Courseware: IMS Learning Design Approach
Abstract. How to design and develop high quality and reusable network
courseware has got more and more attention. The existing standard has obtained
preliminary achievement. However, IMS Learning Design (LD) can become a
new kind of approach. It can transfer the emphasis of network courseware from
object-oriented to process-oriented. Based on the introduction of IMS Learning
Design, this paper proposed the idea of using LD to design process-oriented
network courseware, then a concrete model which use IMS LD tool to develop
courseware was provided.
1 Instruction
As the web-based learning and distant education become popular, the network
courseware has substituted for traditional stand-alone courseware gradually; it has
become the future trend. At present, many computer technology personnel and
educators devote to the development of network courseware, but it is still unable to
meet teacher's needs. What’s the reason? On the one hand, the well-developed
courseware is unable to suit for different kinds of instructional strategies; On the other
hand, the low level repetition development phenomenon is not avoided.
In order to reduce the waste of resources, at present many research facilities
overseas, like IMS, ADL, IEEE, AICC etc, are all devoting to establish the correlative
standards. Now some specifications, such as ADL SCORM, IMS Content Packaging,
have become mature and obtained preliminary application, but all these specifications
pay more attention to content interoperability and re-use, they frequently neglects the
instruction process and strategy which are manifested in the courseware.
∗
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 52–59, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Designing and Developing Process-Oriented Network Courseware 53
In allusion to the current status, we conceived if there is a method to simple the process
of courseware development. According to this method, the courseware can be
disassembled into discreteness; the different parts can be managed as object, then
54 Y.-l. Zhou and J. Zhao
teacher can cooperate with computer professional and other authors, which guarantees
the quality of courseware. Like this, the courseware is not merely content stack and
technology display. It emphasizes the instruction process and strategy.
The characteristic of LD may realize this conception to design and develop the
process-oriented network courseware. First, comparing with other specifications, LD
proposed a relatively high-level frame, which guarantees the courseware have general
description document and structure. It make the courseware content have a general,
simple, easy description and packing method, enables it be applied independently as a
logical unit[4]. Secondly, it can provide a run environment to realize courseware frame
and content packing. It can develop corresponding application software, achieving the
goal of courseware resources sharing and re-use [5]. Finally, it also increased some new
characteristics to expand the scope of technical supported instruction method,
highlighted the process-oriented characteristic.
The implementation of LD may divide into three steps: creation, production and
transmission [6]. In the process, it needs LD tools. They may divide into Editor and
Player according to their function. At present more than 20 tools have been
developed, such as Reload, Coppercore. Each element of IMS LD can be used to
design and develop majority of elements in courseware, its corresponding relations
are available in table 2. In the process of developing, LD tool can provide an external
concrete environment; moreover the thought behind LD is possible to provide theory
guidance.
Based on the thought, the author proposed a development model, showed in Fig1.This
model is divided into three modules: the courseware development module,
management platform, as well as import and output module. These three modules
correspond separately with LD implementation three steps, namely creation,
production and transmission.
Designing and Developing Process-Oriented Network Courseware 55
This paper chose the Science course in seventh grade (Zhejiang Education
Publishing House, China) as an example to expatiate the model. We selected “the
science in calendar” of Section 4, Chapter 4, which contains three activities: different
calendars, leap year, and the exploration of the twenty-four solar terms.
This is the core module in the development process. It provides a dynamic editing
environment for teachers to design and develop. The network courseware development
module needs to use LD editing tool, like Reload, Collage and so on. . It mainly
includes four depositories:
a) Resource Pool
The main function of this pool is providing courseware materials; it corresponds to the
learning object (LO) in LD, including many instruction resources and multimedia
materials. LD itself doesn’t provide tools to develop LO. According to the
characteristic of instruction content, need and teachers’ own skills, they may chooses
many kinds of development tools, like Frontpage, Dreamweaver, PowerPoint. In this
example, we used Dreamweaver to develop the activities in the form of Homepage,
then put them to LD tool (in this example adopts Reload) to packing.
b) Strategy Pool
It is a feature pool of this model that highlights the dynamic developing process.
Activity is the center of LD; the learning activity forms the different study flow.
56 Y.-l. Zhou and J. Zhao
Because the content and learning flow is separate in LD, therefore we also
correspondingly separates the resources and teaching strategy in courseware.
A wide variety of pedagogical approaches can be represented by IMS LD, such as
problem based learning, game based learning, WebQuest and so on [7]. It will obtain
different instructional model and effect by quoting different instructional strategies for
the same resources. The utilization of the “activity structure” element may carry on the
、
arrangement quotation and reference of different activities, which forms their own
learning flow. Simultaneously, using the “condition” to choose and decide the next step
learning program, thus forms the different learning method and path.
In “the science in calendar”, the student's task is mainly to study the three activities,
then carry on group cooperation, upload their work; the teacher is responsible to the
work: providing feedback, monitoring and managing the entire learning process.
According to the three activities, this example adopted the way of sequence. Only
completing the preceding activity, can students enter the next step. For example, the
teacher provides feedback to students and monitor the learning process, the whole
process can be described as Fig2.
c) UOL pool
The courseware material and UOL are both the components of courseware. The
material may not become courseware alone, while the UOL can be an independent
small courseware, which can be reused.UOL includes the characteristics of courseware,
but the courseware is an exceptional case because of some kinds of instruction needs.
UOL adopted networking technology, so it can be recombined, which make the
teaching resources be shared among different local. The teachers can not only reuse the
resources or flow of the courseware, but also the unit module, which enhanced the
compatibility and reusability of courseware well.
The specification must enable a learning design to be abstracted in such a way that
repeated execution, in different settings and by different persons, is possible. It is
necessary to use each element of LD comprehensive to develop UOL that conforms to
the request, showed as Table 2. The metadata is used to describe each kind of learning
content structure, making it pack as an independent learning object. Each small unit has
provided the mutual connection for teachers to retrieval and assembly.
Designing and Developing Process-Oriented Network Courseware 57
d) Database
The database can save other information besides resource and strategy, like the
evaluation information, file name, the media type etc. It aims to enable teachers quote
the corresponding resources conveniently in the knowledge library. The information in
database is temporary, the content renews unceasingly. The attribute of “property” is a
variable that can establish the variable and memory in database. It is the essential part
of monitoring, personalization, evaluation, user interaction.
In “the science in calendar”, we defined three attribute variables: the File variable
used for students to upload their work, the Integer variable used to deposit student's
learning result, the Boolean variable is defined to decided if enter the next step.
The main objective of this module is to carry on management and maintenance to the
overall system. According to the classification of discipline, the nature of curriculum
and other specific demands, the developer complete each operation to pools:
increasing, deleting and inquiry, providing support for the courseware manufacture.
The courseware is deposited in the form of XML documents; simultaneously it needs to
examine the consistency, redundancy and validity of database.
LD check and manage the network courseware by using LD player, such as
Coppercore,Sled plug.Coppercore is used to check the well-developed courseware,
checking its correctness and redundancy, however, the sled plug manage the checked
courseware: create, delete and add roles (teacher, student and so on).
This module aims to realize the share and reusability among different system,
providing the function of exporting and importing. It mainly includes two processes:
package and parse, simultaneously providing the friendly surface for the user.
1) Package
LD platform employs XML language to edit. If all LD systems are expected to share
developed courseware, we must pack the courseware. The main usage of LD is to
contain learning design inside the content package, thus establishes a learning unit. It
did not limit the concrete content and form of resources which packed, but through
defining a unified structure to pack different digital resources. Finally, the package is
exported in the form of wraps (zip) to server.
5 Conclusion
Learning design is one promising new area of both theoretical and technical
development. Although LD is not a special specification to develop resource, but it
makes the resources development sense, the evaluation and feedback to resource to be
more promptly and accurately, enhanced the interaction of resources. At the same time,
the work flow of UOL itself is available to share and re-use, which expanded the
concept of reusability greatly. In brief, it is a brand-new thought and method to employ
LD to design and develop process-oriented network courseware. It causes teacher and
the computer technology personnel cooperate well to develop more courseware of high
quality, managing and monitoring the learning process well.
However, at present, the resources conform to LD is little. It will take a long time to
build a rich resources storehouse. As the LD obtained widespread acceptance
gradually, as the technology break through, many current difficulties will be solved as
well, more people and organizations will pay attach importance to LD, LD will
certainly to receive more welcome.
References
4. Yue-liang, Z., Miao-miao, Z., Gen-qiu, Y.: Developing Educational Paradigm of ICT
Application for Supporting Learning Processes and Activities: Learning Design Approach.
Mmodern Educational Technology (5), 5–8 (2006)
5. Longkefeng, S.: IMS learning design specification and instruction design (2002),
http://www.etc.edu.cn/articledigest30/imsxuexi.htm
6. IMS Learning Design Information Model. Version 1.0 Final Specification (2003),
http://www.imsglobal.org/learningdesign/ldv1p0/imsld_infov1p0.html
7. Koper, R., Tattersall, C.: Learning Design: A Handbook on Modeling and Delivering
Networked Education and Training. Springer, Heidelberg (2005)
Design and Implementation of Game-Based Learning
Environment for Scientific Inquiry
Abstract. Scientific Inquiry refers to the study that learners conduct with
similar or simulated scientific research. It is a problem-oriented learning
activity with quite rich contents and forms driven by curiosity. It is also a
method by taking the initiative to acquire scientific knowledge and understand
science. The game-based learning environment for scientific inquiry (GBLESI)
can effectively expand the scope for learners. This article takes the NMR
experiments as an example to introduce its Design and Implementation.
1 Introduction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 60–69, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Design and Implementation of Game-Based Learning Environment 61
2 Relative Study
The Bee Dance is developed by Illinois University, and the subjects are pupils of the
Abraham Lincoln Elementary School. In the fictitious environment, the pupils have
no other task except observing and understanding phenomena. In the game, two
learners represent two bees who share the same space, and they can put forward
questions and think to study the function of bee dance, during which teachers can give
appropriate hints.
The Virtual Solar System Project is developed by the education college of Indiana
University and the educational technology department of Georgia University.
Learners can study the model of solar system themselves in the CAVE environment,
and study the movement of the earth and moon through recording trails of them.
62 R. Yun, M. Wang, and Y. Li
Fig. 3. Learners in the CAVE environment to study the models of solar system
Scientific inquiry, to say simply, refers to the activity that individuals recognize and
explain nature and reveal the essence of nature through systemic investigation and
study. To understand scientific inquiry exactly, we should know the following several
aspects:
Firstly, the objective of scientific inquiry is the nature, and scientific inquiry is to
investigate and study the natural phenomena or problems. Through it, individuals can
find and reveal the essence and relationships of objects and master the regularity of
natural development.
Secondly, as a cognitive activity, scientific inquiry has some procedures and
periods. Although science includes many different subjects with different methods
and strategies of studies, every study has to go through such similar procedures or
periods as formulating problems, hypothesizing, making study scheme, checking
hypothesis, making conclusions, and so on.
Thirdly, in scientific inquiry, in order to find and reveal the essence of natural
phenomena or objects, individuals have to use a series of scientific methods, such as
observation, comparison, classification, measurement, communication, forecast,
hypothesis, experiment, making models, and so on. Through these ways, they can find
answers to problems about the nature, and know the natural world more deeply. So
scientific methods are the soul of scientific inquiry. Of course, different methods can
be used in different fields for different problems, and there is no settled or uniform
pattern.
Fourthly, as an exploring activity, scientific inquiry has double meanings. It can
refer not only to the inquiry in the scientific field, namely, the activities that scientists
Design and Implementation of Game-Based Learning Environment 63
do to study the nature, but also to the learning activity of learners, which is similar to
the scientific study.
Firstly, we must stimulate the interests of learners by some circumstances related with
issues, which are created by the games. The goal of creating circumstances associated
with issues is to expand the thought of the learners, make them put forward their own
questions and finally decide what they need to find through the following work. The
appropriate game scene for circumstances associated with questions should have the
following characteristics:
• Obstacles bringing conflicts, imbalance and the intelligence challenge;
• Interests providing interests, arousing learners to reflect;
• Open solving the problems through wide and various thought, which have no
definite answers or solutions.
• Diversity be propitious to learners of all levels to find answers gradually
• Practices searching for solutions of problems by exploration activities. Both
individual work and team work is ok.
Secondly, we speculate on solutions of the problems in the game scene. It’s a key
step for resolving problems and we should remember to consider the learners’ age and
life experience.
Thirdly, we organize learners to discuss the problems and make plans. This is the
preparatory stage for learners to implement activities and also the key points of
effective Inquiry. It can be said that the detailed design and feasibility of this step is
directly related to the success of the next activities.
Fourthly, we carry out the plan to see if the solution is right. Organizing Learners
to discuss and observe the places of the game scenes that we should pay attention to,
or introducing them the rules for use and operation, both are important and
meaningful for them if they use the equipment for the first time. We also should give
learners adequate time to explore, insure that they can finish all the prearranged
activities by experimental plans.
Fifthly, here comes the last step to gather everyone together and have their ideas.
Learners will communicate with each other and finally educe some common
conclusions. Sometimes, work of collecting learners’ ideas is carried out in the fourth
step also. It is a process of wiping off pseudo things for truth. The results may be
positive for our assumptions and may also be negative. Whatever, only when we
collect the learners’ ideas well and sorted out them effectively, we can make the
whole work significative and the final conclusion reflect the truth.
The National Research Council (NRC) of USA published a monograph named
Inquiry and The National Education Standards: A Guide for Teaching and Learning.
In this book, they expatiate on 5 essential features of classroom inquiry. We can see
them in the followed table 1:
64 R. Yun, M. Wang, and Y. Li
Features Expatiation
Question Learners are engaged by scientifically oriented
questions.
Evidence Learners give priority to evidence, which allow them to
develop and evaluate explanations that address
scientifically oriented questions.
Explanation Learners formulate explanations from evidence to
address scientifically oriented questions.
Evaluation Learners evaluate their explanations in light of
alternative explanations, particularly those reflecting
scientific understanding.
Communication Learners communicate and justify their proposed
explanations.
The virtual environment is designed for single person with 3 main parts As is (
)
showed in Fig4 below : Show pictures of NMR CT, Scan human body with the
NMR CT System, and Reveal essence of The NMR.
The first part is used to show pictures made by NMR equipment ( e.g. NMR CT
pictures of the head, NMR CT pictures of the backbone, NMR CT pictures of the
chest and so on. ). In this part, we can give out both the healthy organ pictures and
bad ones at the same time, so we can see their differences. This part is used to arouse
the interests of learners and make it a proper time to introduce the NMR theory.
Teachers will give some questions to learners in this part.
The second part is used to simulate NMR CT System. In the virtual environment,
we create several persons and apparatus to simulate the real activities in the hospital.
Learners interact with the models and get proper response: movies, pictures, words
for explanation and so on. By doing research in the environment, learners will
gradually summarize the condition for NMR phenomenon. For the whole job of
creating such a virtual learning environment, the second part plays the most important
role. It must offer learners enough chance to interact with the environment and get
plenty of proper response. In order to keep the learners’ interests, we have to make the
environment somewhat like a game, such as setting up an area to display the learners
“Scanning points” which is decided by their activity in some scenes: configure the
apparatus, change the water content in human body, scan more places and answer
more questions .
The last part will make learners close to the essence of NMR theory in forms of
movies, pictures and text. After gathering the learners to discuss their research and the
phenomenon in the second part, teachers can use this part to make them understand
the theory better.
Design and Implementation of Game-Based Learning Environment 65
scan the human body. Now the Scanning Simulating Scene will be loaded into the
environment.
Step5: Inside the scanning device, we use some digital and laser light effect to
simulate the scanning process of the magnetic field and the radio frequency signals.
Step6: Both inside and outside of the scanning device we set displayers to display the
result of learners’ work, that usually in forms of pictures, movies and score.
Step7: Learners press buttons to change the magnetic field and the radio signals ,
they scan many places of human body and get different feedback. By doing research
work in the virtual environment, learners will gradually find there own idea about
NMR CT System’s working condition.
Step8: At the top left corner of the scene, we set an area to show the learners’
“scanning score”, which is decided by how they set up the device, how many places
they scanned and how they answered the questions. This part is the embodiment of
the environment’s challenge and fun as a “game based ” learning environment.
(3) After most learners have finished their research work, teachers gather them
together to talk about the questions. Teachers should listen to every speaker’s words
carefully and make proper comments. At the same time, teachers should better use the
environment again when they make comments or explain some theory.
(4) At last, teachers use the third part to explain the formal and in-depth theory of
NMR. It is a nice choice to show the learners some movies here.
models. This small tool will not destroy the model’s material quality and details while
simplifying its surface. After being optimized, models become smaller and easier for
explorer to load and display. The more simple models are, the more smoothly for
explorer to change the scene. We choose VRML2.0 to create the virtual environment
world and organize the models. Some simple models (such as wireframe, box and
sphere) and animations (such as translation and rotation) are also created by VRML
using its inner node.
We use the inner node of VRML2.0 to create the most static scene and its
transform. In order to improve code efficiency and reusability, we need to create
many new nodes, which have their own fields and functions. Here the object-oriented
thinking is necessary , the new nodes should be encapsulated and inheritable. So we
can create nodes based on the proto mechanism.
Proto mechanism includes three parts: declaration, implementation and instance.
The keyword proto is used to declare the node type defined by developers. The
definition of a node type contains field, exposedField, eventIn, eventOut and
nodebody. Proto can be defined as:
PROTO<node type name>[<interface list >)]{<node body>}
LOD is a system feature for optimizing the amount of details rendered in a scene.
VRML provides an LOD node, which can explicitly change different leveled detailed
versions of model. We employ this way to optimize VRML files so as to accelerate
rendering speed in real time walkthrough.
Script program can read and write the Fields defined by users in script nodes
through their field names, and the method getField() provide a way to refer interface
fields.
5 Conclusion
Compared with the scientific exploration in real life, game based learning
environment for scientific inquiry avoid the constraints of space-time, security and
resources. It can lower the cost to provide learners relatively rich learning content.
Incorporate appropriate games, problem –oriented virtual environment can stimulate
the interest of learners well and provide them with joy and rich learning activities.
There are varieties of ways to implement a game based learning environment for
scientific inquiry, In this article, we introduce a new way to implement the virtual
environment for NMR experiment with VRML2.0, JavaScript and 3DS Max
modeling tools. Through practice test, we have achieved good results and hope it
useful for others’ design or implementation.
References
1. Downes, S.M.: Truth, Selection and scientific inquiry. Biology and Philosophy 15(3),
425–442 (2000)
2. Davis, K.S., Falba, C.J.: Integrating Technology in Elementary Preservice Preservice
Teacher Education. Orchestrating scientific inquiry in Meaningful Ways 13(4), 303–329
(2002)
3. Ieronutti, L., Chittaro, L.: Employing virtual humans for education and training in
X3D/VRML worlds. Computers & Education 49(1), 93–109 (2007)
4. National Research Council of USA: Inquiry and the National Science Education
Standards: A Guide for Teaching and Learning, National Academies Press (2000)
5. Murano, P., Mackey, D.: Usefulness of VRML building models in a direction finding
context. Interacting with Computers 19(3), 305–313 (2007)
6. Liu, Q., Sourin, A.: Function-based shape modelling extension of the Virtual Reality
Modelling Language. Computers & Graphics 39(4), 629–645 (2006)
7. Koročsec, D., Holobar, A., Divjak, M., Zazula, D.: Building interactive virtual
environments for simulated training in medicine using VRML and Java/JavaScript.
Computer Methods and Programs in Biomedicine 80, 61–70 (2005)
8. Moore, K., Dykes, J., Wood, J.: Using Java to interact with geo-referenced VRML within
a virtual field course. Computers & Geosciences 25(10), 1125–1136 (1999)
9. Whiteman, P.: How the bananas got their pyjamas: A study of the metamorphosis of
preschoolers’ spontaneous singing as viewed through Vygotsky’s Zone of Proximal
Development. A Thesis for Doctor Degree of Philosophy, the University of New South
Wales (2001)
10. Yun, R., Pan, Z., et al.: An Educational Virtual Environment for Studying Physics Concept
in High Schools. In: Lau, R.W.H., Li, Q., Cheung, R., Liu, W. (eds.) ICWL 2005. LNCS,
vol. 3583, pp. 326–331. Springer, Heidelberg (2005)
11. Davison, A.: Pro JavaTM 6 3D Game Development, pp.3–12. Apress (2007)
12. Dave, A., Kevin, D.H.: Beginning OpenGL Game Programming, Course PTR (2004)
13. Multigen-Paradigm Incorporated (online2006.01.20), http://www.multigen.com
14. James, G., Bill, J., et al.: The Java Language Specification (1996)(online2006.01),
http://java.sun.com/docs/books/jls/html/index.html
Research and Implementation of Web-Based E-Learning
Course Auto-generating Platform
1 Introduction
Web-based course as the main approach to carry out teaching activities of E-Learning
plays an important role in E-Learning, thus courses of high quality are the important
guarantee to develop E-Learning. However, the current status of developing web-based
E-Learning courses is not optimistic and the courses that can be used in practice are not
the majority. There are several problems: first, most of persons who develop web-based
E-Learning courses are members of computer profession, but teachers who truly
understand characters of students and subject and teaching regulations of E-Learning
have serious difficulty in developing web-based courses. So the effect of teaching
activities is discounted and teaching design can’t be realized. Second, although the
forms of presenting teaching contents and carrying out teaching activities are various, it
is difficult to update teaching contents and activities according to characters of students
and subject and teaching regulations of E-Learning [1]; Third, because of the
differences of running environments and the independence of development, the modes
of running and maintaining are not united [2]. The problems hereinbefore influence
the quality of web-based E-Learning courses severely. On a long view, low-
quality web-based E-Learning courses not only waste time, money and human resource
of developers, but also lead the result that students gain little and teachers loose
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 70–76, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Research and Implementation of Web-Based E-Learning Course 71
interests [3]. Therefore, the advantages of E-Learning are weakened and the
development of E-Learning is baffled. In order to settle the problems above, the
research developed a scientific and valuable platform of auto-generating web-based
E-Learning courses. After being trained, teachers who truly understand characters of
students and subject and teaching regulations of E-Learning can develop web-based
E-Learning courses without the help of members of computer profession. The courses
include instructional design philosophy of teachers and can be maintained and
readjusted anytime and easily.
2 Design of Platform
The module is the core of the platform. The flow (Fig. 2) of course development is:
after applying a new course, user chooses template of course’s style and instructional
strategy to confirm function modules of course (user can design function modules if
he/she don’t choose instructional strategy) and enter the interface of course making. In
the interface, user can edit function modules, page contents and style of the course. The
course could run as an independent web-based E-Learning course after published.
The platform used ASP.NET2.0 as its key technology. A series of new features of
ASP.NET2.0 such as Master Page, Website Navigation, User Management, Profile and
Theme/Skin were applied to enhance the efficiency and quality of the platform
development. The software development environment of the platform was Visual
Studio 2005 and its background database was SQL Server 2005.
Research and Implementation of Web-Based E-Learning Course 75
According to the idea of software engineering, the architecture of the platform was
composed of four layers so as to achieve code reusablity, maintainability and
expansibility [9]. Fig.3 shows the four layers architecture of the platform.
Presentation Layer: Presentation layer was composed of ASP.NET pages that share
a common page layout.
Business Logic Layer (BLL): Business Logic Layer enforced custom business rules.
Data Access Layer (DAL): Data Access Layer served as bridge between BLL and
Database Layer.
Database Layer: Database Layer was used for storing data.
In four layers architecture each layer only interacts with its neighbor layers. Using
well defined interface, the inner implementation of each layer was irrelevant to others
and switch of heterogeneous database can also be achieved easily by updating
configuration files. Thus, the efficiency and quality of the platform development can be
enhanced.
4 Conclusion
The purpose of the research is to develop a platform of auto-generating web-based
E-Learning course, in order to make convenient for teachers whose professions are
irrelevant to computer to develop web-based courses that follow characters of students
and subject and teaching regulation of E-Learning. The platform applied teaching
philosophy to function modules and used guide pattern to instruct teachers to develop
courses, so the quality of courses can be guaranteed. The platform is using in our
university now. It is approbated by teachers and students. Teachers and
students’enthusiasm for E-Learning is greatly enhanced and the teaching quality of
web-based E-Learning courses is improved. The platform has promoted the further
development of E-Learning to some extent.
References
1. Luo, H.: Actuality and Countermeasure of Network Courses Development. Modern Distance
Education 4, 33–34 (2003)
2. He, K.: Modern Educational Technology and Design and Development of High-quality
Network Courses. China Audiovisual Education 6, 5–310 (2004)
76 Z. Wang, X. Wang, and X. Wang
3. Chen, Y.: Research and Improvement of Course Development Tools Based on Web, pp. 3–8.
Beijing Normal University (2002)
4. Schewe, K.-D.: A Conceptual View of Web-Based E-Learning Systems. Education and
Information Technologies, 81–108 (2005)
5. Niemi, H., Nevgi, A., Virtanen, P.: Towards self-regulation in Web-based learning. The
annual meeting of the American Educational Research Association (2003)
6. He, K., Li, K., Xie, Y.: Theory Fundament of Guide-Participant Teaching Mode. E-Education
Research 2, 3–9 (2000)
7. Cui, M., Zhang, R.: Application & Research for BBS in Online Education. Journal of Chang
Chun Teachers College 23(3), 43–45 (2004)
8. Beck, I.L., McKeown, M.G.: Inviting Students into the Pursuit of Meaning. Educational
Psychology Review 13(3), 225–239 (2001)
9. Cazzulino, D.: Beginning C# Web Applications with Visual Studio.NET, pp. 223–470.
Tsinghua University Press (2003)
A Humanized Mandarin e-Learning System
Based on Pervasive Computing
1 Introduction
With Chinese opening and fast economy development in recent years, the
communication between China and the world becomes more and more important in a
wide range. So Mandarin, the important communication tool and culture carrier that
lets the foreign countries know China, attracts more and more governments,
educational organizations and corporations. Consequently, hundreds of Mandarin
learning platforms, tools, electronic learning materials, and so on, have been
developed, such as Go to China Mandarin Platform Advance in 2004 [1], StepByStep
by HeiLongjiang University in 2001 [2], Chinese Master by Yuxing Soft LTD. in
2004 [3], etc. Although they are all good tools for learning Chinese and have been
designed to help students to learn a large number of Chinese words as quickly as
possible, they only shift the problem to the dependency on an available and
appropriate disc drive or software.
The spectacular development of Internet provides an untraditional and wide way of
education. As a result, e-Learning is becoming a fixed part of most people’s life, as
they are forced to life-long learning, no less than our Mandarin Studying. The e-
Learning system can also afford communications between the foreign learners and the
teaching server. There are some typical Mandarin e-Learning systems, for example
Chinese Horizon Mandarin Training by Yahoo in 2004 [4], The Ottawa Mandarin
School’s curriculum by Taipei language Institute in 2001 [5], EASE Mandarin by
Mandarin House Language Institute in 2001 [6], etc. However, our investigation
indicates that the teaching systems of e-Learning are lack of intelligence. They cannot
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 77–87, 2008.
© Springer-Verlag Berlin Heidelberg 2008
78 Y. Ming et al.
interact with learners, cannot adjust the curriculum contents based on their learning
situation, and most importantly cannot make the learning anywhere and anytime.
Pervasive Computing emerges as the times requires and gradually penetrates our
daily life. When we talk about the main advantages of pervasive computing, we
usually think of anywhere, anytime, any format, and any device [7]. This means:
·Anywhere: global accessibility, with regard to various kinds of communication
networks.
·Anytime: twenty-four hours, but also independent of other services or persons .
·Any format: email, public services, inter- and intranet, various data formats.
·Any device: (Table-) PC, Personal Digital Assistant (PDA), cell phone, etc.
Up to now, the computing scientists have developed a wide range of applications in
pervasive computing, for instance, pervasive retailing information systems [8], wide-
area e-health monitoring [9], domestic ubiquitous computing applications [10], etc.
But pervasive computing based language e-Learning systems, especially Mandarin e-
Learning, have not seen so far. Thus we can take full advantage of pervasive
computing technologies to design our Mandarin e-learning system where the learner
can put forward his personalized learning requests according to his knowledge
structure and learning plan. Then, our system will analyze the learner’s learning
history and demands, adjusting the curriculum contents, which stand for his learning
requests and preferences anytime and anywhere without any restrictions.
Moreover, in order to facilitate the foreign learners and bring more and more web-
based Mandarin e-learning resources and personalized learning guides, we design
pronunciation component as a major function of our system where Mandarin language
learners can practice their pronunciations including tones. In a word, our system is a
pool of Mandarin language place where you can interact with people from all around
the world to share your interests and concerns about Mandarin.
The structure of this paper is as follows: our e-learning system is described briefly
in section 2; in section 3 and section 4, the application server and the enterprise server
are presented separately: then we describe the core function of our system—
pronunciation assessment in part 5; finally, the design and implementation of our
Mandarin e-learning system as a whole is introduced in details.
2 System Description
In terms of the concept of pervasive computing discussed above, we design our
pervasive Computing Mandarin e-Learning System as in Fig.1.
To implement this system, Java is used due to the isomerism of platform, and JSP
used to design the web page. We put the web server and IP-to-PSTN gateway in two
computers. Apache Tomcat is used for the web server and Asterisk used to provide a
central switching in the gateway [11].
The Web Server in Fig.1 is further illustrated in Fig.2, as it is the core of our
humanizing e-Learning system. It is composed of two main parts—application server
and enterprise server. The Teacher Agent in Fig.1 provides a major e-Learning
function of the system — pronunciation evaluation. They will be described in
following sections in details.
A Humanized Mandarin e-Learning System Based on Pervasive Computing 79
4 Enterprise Server–Database
Register Success
Register User Register User Login
Ⅱ . Tone Evaluation
Mandarin is tonal language. The Mandarin four different tones include a lot of
important information. So we process the tone recognition separately. The pitch
contrail can distinguish four Mandarin tones efficiently, and then an event detection
pitch detector based on the dyadic wavelet transform which can detect the catastrophe
point of speech signal when people speak [15]. We use cured-fitting technology to
realize the classification of the different tones.
Observing the pitch contrails, we find the slope of contrail can distinguish the
different tones easily when we regard the information about the relative phoneme of
tone as in Fig.6. So we decide to extract the slope of pitch contrails to recognize the
Mandarin four tones.
Then we can calculate the tone accuracy of pronunciation by using the result of
tone recognition. We set the average of these radian Ai is A , the absolute value of
A is | A | .
n
| A |=| (∑ Ai ) / n | (3)
i =1
When the tone is tone one, the accuracy of tone one is | A | /(π /18) × 100% .
When the tone is tone two, if | A |≥ π / 3 , the accuracy of tone two is 100% , if
| A |< π / 3 , the accuracy of tone two is (| A | −π /18) /(π / 3 − π /18) × 100% .
When the tone is tone four, the method of compute accuracy is the same as tone two.
When the tone is tone three, the method of compute accuracy became complex. We
need to find the lowest sample B , divide the Ai sequence into two sub-sequence
86 Y. Ming et al.
Ai1 = A1 ,..., B and Ai2 = B,..., An , then compute the | A1 | and | A2 | . Set variable
R = π − | A1 | − | A2 | , if 0 < R < π / 3 , the accuracy of tone three is 100% , else the
accuracy of tone three is ( R − π / 3) /(π − π / 3) × 100%
·Soft Phone: If the learner meets some difficulties during the course of language
studying and clicks this button, a soft phone is started up via IP-to-PSTN gateway,
and then he can communicate with a virtual teacher about his questions.
·Voice Mail: Java Mail used to design this button. Information of learners (time,
space, the result of studying, etc.) can send to both the teacher and learners via email
to make Mandarin learning conveniently and efficiently.
The Mandarin E-learning system also provides many smart servers. He can see his
learning information, which is collected automatically for his teacher making the
teacher know his learning effects. One important feature is that he can test his
Mandarin pronunciation accuracy by the PRACTISE button as is shown in the Fig.7.
A Humanized Mandarin e-Learning System Based on Pervasive Computing 87
7 Conclusion
This paper describes a pervasive computing system design and implementation for
Mandarin e-Learning. We analyze the foreign learners’ requirements and discuss the
implementation in pervasive environment. Learners will mainly use the platform for
learning, because within the platform not only the learning material is presented, but
also communication and interaction takes place. This gives us bright hope in the
success of our scheme and we are convinced that such a scheme will indeed become
practical and scalable for its deployment over Internet for Mandarin e-Learning. It’s a
place where foreign learners can find the solutions they are looking for.
References
1. http://www.go2china.net
2. http://www.cowob.com
3. http://www.learnchinese.cn/agents.html
4. http://www.chinesehorizon.com/
5. http://www.ottawa-mandarin-school.ca
6. http://www.easemandarin.com/
7. Loidl, S.: Towards pervasive learning: WeLearn.Mobile. A CPS package viewer for
handhelds. Journal of Network and Computer Applications 29, 277–293 (2006)
8. Kourouthanassis, P.E., Giaglis, G.M., Vrechopoulos, A.P.: Enhancing user experience
through pervasive information systems: The case of pervasive retailing. International
Journal of Information Management 27(5), 319–335 (2007)
9. Su, C.J.: Mobile multi-agent based, distributed information platform (MADIP) for wide-
area e-health monitoring. Computers in Industry, Corrected Proof (Available online
September 17, 2007) (in press)
10. Schmidt, A., Terrenghi, L., Holleis, P.: Methods and guidelines for the design and
development of domestic ubiquitous computing applications. Pervasive and Mobile
Computing (Available online July 22, 2007) (in Press)
11. Ming, Y., Miao, Z.: Humanied Mandarin E-Learning Based on Pervasive Computing. In:
I3E 2007: The IFIP Conference on e-Business/Commerce, e-Services, e-Society,
Huazhong Normal University, Wuhan, China, October 10-12 (2007)
12. Thriskos, P., Zintaras, E., Germenis, A.: DHLAS: A web-based information systemfor
statistical genertic analysis of HLA population data. Computer Methods and Programs in
Biomedicine 85, 267–272 (2007)
13. Johnson, R.: Expert One-on-One J2EE Design and Development, Wrox-2003 (Chapter 9:
Practical Data Access)
14. Young, S.: Large Vocabulary Continuous Speech Recognition: a Review, Cambridge
University Engineering Department, pp. 1–23
15. Ming, Y., Miao, Z., Su, W.: Tone Analysis for a Pervasive Computing Mandarin e-
Learning System. In: The Second International Symposium on Pervasive Computing and
Applications, Birmingham, UK (July 2007)
An Interactive Simulator for Information
Communication Models
Mohamed Hamada
Abstract. Information theory is the science which deals with the concept
‘information’, its measurement and its applications. In common practice
information is used in terms of a Communication Model in which the emphasis
lies on the transport of information, as generated by a source, to a destination.
The communication system should transmit the information generated by the
source to the destination as fast and accurately as possible. To achieve this goal
several coding techniques were developed based on mathematical concepts.
Due to this mathematical nature, information theory course is used to be taught
by a traditional lecture-driven style. Studies showed that lecture-driven style is
not effective with computer engineering students due to their active learning
preferences. In this paper we introduce an interactive communication model
simulator to facilitate teaching and learning of the basic concepts of information
theory course. We also show the effectiveness of using the simulator in
classroom.
1 Introduction
Information theory is the science which deals with the concept ‘information’, its
measurement and its applications. In common practice information is used in terms
of a Communication Model in which the emphasis lies on the transport of
information, as generated by a source, to a destination. It addresses questions such as:
How to transmit and store information as compactly as possible? What is the
maximum quantity of information that can be transmitted through a channel? How
can security be arranged? Etcetera [1]. In answers for these questions several coding
algorithms (such as Huffman code, Shannon code, Fano code, etc.) and several
concepts (such as the concept of information entropy, and the concept of channel
capacity) were developed.
Information theory concepts are abstract in nature and hence used to be taught by a
traditional lecture-driven style which is suitable for learners with reflective
preferences. Since computer engineering learners tend to have strong active
preferences (Rosati [14]), a lecture-driven teaching style is less motivating for them.
Our communication model simulator (CMS) is designed to tackle this issue and meet
the active learning preferences for computer engineering learners. CMS can be used as a
supporting web-based tool for active learning not only for information theory course,
but also for several other courses such as courses in telecommunications, error
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 88–98, 2008.
© Springer-Verlag Berlin Heidelberg 2008
An Interactive Simulator for Information Communication Models 89
correcting codes, image processing, and other related fields. Such courses cover a
variety of topics including coding techniques, communication channels, information
source, error-detection and correction, information entropy, mutual information, in
addition to basic concepts of probability. We cover such important topics in our CMS
environment. CMS is written in Java as an applet using Java2D technology of Sun
Microsystems [6]. This implies that our CMS environment is portable, machine
independent and web-based enabled which makes it useful tool as interactive learning
environment for CS and CE learners.
In designing our CMS learning tools we considered the active construction
learning model [3, 16] that has some basic design principles including the following.
1. Teachers act as facilitators not as knowledge transmitters. This means
knowledge must be actively constructed by learners, not passively transmitted by
teachers.
2. Assessment procedures should be embedded in the learning process and should
consider learners’ individual orientations.
To show the effectiveness of our CMS environment as a model of interactive
learning tool, several classroom experiments were carried out. The preliminary results
of these experiments showed that using our tools not only improved the learners’
performance but also improved their motivation to actively participate in the learning
process of the related subjects and think beyond the scope of the class.
The paper is organized as follows. Following this introduction, section two briefly
explains the topics covered in our CMS including Huffman, Shannon, Fano and
Arithmetic coding techniques, channel capacity, information source, and the basic
communication model. In section three we introduce the communication model
simulator. The performance evaluation of the simulator will be given in section four.
Finally, we conclude the paper and discuss the results and possible future extensions
in section five.
2 Information Theory
Information theory is the science which deals with the concept ‘information’, its
measurement and its applications. Its main goal is how information can be
transmitted, from source to destination, as reliable as possible. The basic
communication model, shown in Fig. 1, has two basic components; transmitter and
receiver. The transmitter is responsible for the formatting and sending of the data
from the source through the channel. Then the receiver gets and re-formats the data to
its original form and passes it to the final destination.
The transmitter component has several units. The data reduction unit is responsible
for removing unnecessary data from the source. The source encoding unit formats the
reduced data to become as compact as possible for faster transmission. To prevent
improper use of the transmitted data, the encipherment unit is used. In practice, the
data channel is subject to external noise which may cause data distortion, to help
detect these data errors, the channel encoding unit is used. The receiver component
has units with opposite functions to the transmitter component. The channel decoding
unit uses the channel encoding information to detect and correct any possible errors.
90 M. Hamada
TRANSMITTER
(Modulation)
Noise Channel
(Demodulation)
RECIEVER
The encrypted data is then decrypted by the decipherment unit. To return the
compacted data into its original format, the source decoding unit is used. Finally, the
data reconstruction unit puts the data into its final form that is suitable for the
destination.
Coding techniques are essential for reliable and efficient data transmission. Our
CMS covers the basic coding techniques such as: Huffman, Fano, Shannon, Ziv-
Lempel, and Arithmetic algorithms. Since we consider the noisy channel (which is
more common in practice), the transmitted data are subject to distortion. To deal with
this distortion we integrated the Hamming error-detection and correction techniques
into our CMS environment.
Entropy is one of the most deep and fascinating concepts in mathematics. It was
first introduced as a measure of disorder in physical systems but for information
theory it will be most important in a dual role as representing average information and
degree of uncertainty. We cover the concept of entropy and related concepts such as
channel capacity and mutual information. Due to the lack of space we can’t explain
more about these concepts here. For more details we refer to [9].
By clicking the “Message” button in the ctrl panel, learners can input the desired
source message they want to send (Fig. 5). The “Send/Stop” button in the ctrl panel is
used to start the transmission process of the inputted message. The transmission
process is then visually seen on the CMS window in a step-by-step manner where
users can pause and resume the transmission at any time to see which change had
happened to the message at each point of the transmission.
The “Show” menu in the ctrl panel is used to view the details of many concepts
such as the message transmission details (Fig. 6) and the coding/decoding processes
details (Figs. 7 and 8). In addition to other concepts such as entropy, mutual
information, channel capacity, and so on.
Figs. 4 to 8 show a detailed example. In this example the inputted source alphabet
is h, e, l, o, - (for space), w, r, and d with probability distribution 0.1, 0.1, 0.25, 0.15,
0.1, 0.1, 0.1, and 0.1 respectively (Fig. 4). The inputted source message is “Hello
World” (Fig. 5). The transmission details of this message are shown in Fig. 6. During
the transmission process of this message the Huffman coding technique was selected
by the learner. Figs. 7 and 8 show the details of the Huffman coding/decoding
processes (of the source alphabet h, e, l, o, -, w, r, d) respectively.
An Interactive Simulator for Information Communication Models 93
In addition to viewing the details, learners can try the concepts themselves through
the interactive “Test” tool. From the ctrl panel interface learners can click on the
“Test” button to view the test menu (Fig. 9). From the test menu learners can select a
topic and start the test. For example, in Fig. 9, the Huffman coding test was selected.
An Interactive Simulator for Information Communication Models 95
Fig. 11. The coding test detailed steps (Huffman code example)
From the Huffman coding test interfaces (Figs. 10 and 11) learners can try to find the
correct Huffman code for the inputted source alphabet (h, e, l, o, -, w, r, d). They can
repeat the trail (with some possible hints) until finding the correct code. Fig. 11 shows
an animated binary tree that represents the Huffman code. Learners can also try to
construct the tree to find the correct code.
4 Evaluation
We carried out two experiments and an opinion poll in order to evaluate the
effectiveness of our CMS tools on the learning process of engineering students. The
first experiment evaluates the learning preferences of the students according to
the “Learning Style Index” of Felder-Soloman [15]. The second experiment evaluates
the effectiveness of using the tools on the students’ performance. Finally an opinion
poll was carried out in order to observe the students’ feedback about the tools.
To help learners find their learning preferences, the Felder-Soloman Index of
Learning Style [15] was introduced. Fig. 12 shows a summary of the learning style
quiz results from the author's evaluation as well as the data found at the University of
Western Ontario by Rosati [14] where he surveyed 800 students and found that
engineering students typically have preferences toward active, visual, and sensing
learning preferences. It also contains a similar data by J. Masters et. al. at San Jose
University [14]. It is clear that our data (on Japanese students) and Masters's data (on
American students) support the data collected on Canadian students by Rosati [14].
96 M. Hamada
90
80
70
60 University of Aizu
50 University of Western
40 Ontario
San Jose University
30
20
10
0
Active Sequential Sensing Visual
A preliminary study shows that using the CMS can improve the learning process of
computer engineering students who study the information theory course and related
courses. Last semester, 100 students in the information theory class were divided into
four groups, each group containing 25 students. A set of 40 randomly selected
exercises was distributed among the groups, 10 for each group. Each group members
could collaborate inside their group but not with any other group members. No group
could see the exercises of other group. Two groups were asked to answer their
assigned exercises using the CMS and the other two groups without using it. An equal
time period was provided to all the groups. The result of the answers showed a better
performance for the two groups using the CMS. Then, the experiment was repeated
by redistributing the exercises among the four groups. Again, the two groups with the
CMS showed better performance in answering the assigned exercises.
In addition to the experiments an opinion poll among the learners was carried out.
The result of the poll is shown in Table 1. Among the 100 learners in the class, 95 had
completed the poll as shown in Table 1(a). Among the 95 responses 79% preferred
using the CMS tools as shown in table 1(b). Most questions on the opinion poll were
Likert-type questions that made a clearly negative or positive statement about the
CMS tools and allowed the learners to strongly agree, agree, be uncertain, disagree, or
strongly disagree. Scores for the CMS tools were generated based upon the learner
responses. The scores could fall between 0 (worst) and 50 (best) and were divided
into five ranges: greatly disliked the tools (score: 0-10), disliked the tools (score: 11-
20), uncertain in preferences for the tools (score: 21-30), liked the tools (score: 31-
40), and greatly liked the tools (score: 41-50). Table 1(c) shows the average score for
the CMS tools lies on the far end of the "liked the tools" range.
An Interactive Simulator for Information Communication Models 97
Table 1(d) and (e) show the responses to other important questions. These results
show that the majority of learners found that the CMS tools helped clarify important
concepts and encouraged them to think about concepts outside of class. The latter is a
significant accomplishment that could lead learners to seek more knowledge and
information on their own.
5 Conclusion
With the vast advance in technology, the traditional lecture-driven classroom is giving
way to a new and more active environment, where students have access to a variety of
multimedia and interactive course materials. Such interactive course materials have
already been introduced for several topics in engineering courses; see for example
[2, 4, 5, 7, 8, 10, 11, 12, 13].
In this paper, we followed the same path and introduced a communication model
simulator to support active learning in the information theory course. It can also be
used in other related courses such as telecommunication, error correcting codes,
image processing, and other similar courses. Our CMS environment is web-based,
easy-to-use, and stand-alone which make it a useful tool of e-learning. Through the
results of our experiments, we also showed that our CMS tools can enhance learners’
motivation and performance. In addition an opinion poll showed a positive feedback
on the CMS tools from the students. In future work, we plan to enhance our CMS
tools by adding more features and more visual examples, and by performing more
performance evaluation experiments.
References
1. der Lubbe, J.: Information Theory. Cambridge University Press, Cambridge (1997)
2. Hadjerrouit, S.: Learner-centered Web-based Instruction in Software Engineering. IEEE
Transactions on Education 48(1), 99–104 (2005)
3. Hadjerrouit, S.: Toward a constructivist approach to e-learning in software engineering. In:
Proc. E-Learn-World Conf. E-Learning Corporate, Government, Healthcare, Higher
Education, Phoenix, AZ, pp. 507–514 (2003)
4. Hamada, M.: Visual Tools and Examples to Support Active E-Learning and Motivation
with Performance Evaluation. In: Pan, Z., Aylett, R.S., Diener, H., Jin, X., Göbel, S., Li,
L. (eds.) Edutainment 2006. LNCS, vol. 3942, pp. 147–155. Springer, Heidelberg (2006)
5. Head, E.: ASSIST: A Simple Simulator for State Transitions. Master Thesis. State
Univesity of New York at Binghamton (1998),
http://www.cs.binghamton.edu/~software/
98 M. Hamada
1 Introduction
Edutainment has long been recognized as a combination of entertainment and
education. To most people, edutainment assumes the forms of TV programs, movies,
video games and computer games. The educational aspects of edutainment programs
are mostly based on “braingames” which aim to achieve certain intellectual learning
objectives within specific curriculum. The learning outcomes of such “braingames”
normally focus on academic goals such as improving literacy, numeracy and problem
solving skills. On the other hand, physical education is another important aspect in the
development of a student from early childhood to late adolescence [1]. Within a
typical curriculum framework, dance plays a role in providing education and training
in locomotion, balance, support, rotation, social and team skills. In the context of
Thailand, traditional Thai dances form an important part of the national and cultural
heritage which should be taught and preserved.
Development of dance robots has improved greatly due to rapid advancement of
electronics, computer, mechatronics and control technologies. However, most
humanoid robots such as Honda ASIMO, Fujita HOAP-3 and Sony QRIO are
extremely expensive and are beyond the reach of the public or educational
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 99–106, 2008.
© Springer-Verlag Berlin Heidelberg 2008
100 C.C. Fung, T. Nandhabiwat, and A. Depickere
organizations such as schools. On the other hand, low cost toy humanoid robots have
been gradually gaining acceptance in the consumer markets. In addition, such robots
have also been expanding their functionalities with each new version being produced.
The most important fact is that these robots cost only a fraction of the more advanced
research humanoid robots. It is therefore hypothesized in this study that low cost
robots could be used as an edutainment tool to motivate and encourage school
children to participate in the dance sessions of the physical education curriculum. It is
proposed that this category of robots could form an integral tool for entertainment and
for the teaching of Thai dances in Thailand.
In this paper, a low cost toy humanoid robot, RS Media [2], is used to teach Thai
dances and to entertain in public exhibitions. In order to establish a unique identity
and character, the robot is dressed in traditional Thai custom and was given the name
“iThaiSTAR”. The name stands for “intelligent Thai Sanook Training-Assist Robot”.
Sanook in Thai means fun and entertaining. iThaiSTAR has performed in schools and
danced with primary school children. It has also taken part in a major Exhibition at
Bangkok – the ICT Expo [3], and was invited to appear at TV shows. It has also been
reported in articles on the Web and in printed newspapers. Informal feedbacks and
comments from the public mainly concerned the acceptance of such robots and the
feasibility of using them for the purpose of teaching and entertainment. A high degree
of acceptance and keen interest from the public was demonstrated. This could lead to
the potential of using low cost robots for training, entertainment and edutainment
purposes in the future.
WowWee Limited is a privately owned company and it is best known for its line of
Robo-branded consumer robotic products. WowWee's Robosapien was the first line
of robotic products that use biomorphic motion technology and is programmable to
perform a variety of functions. The latest Robosapien models are the V2 and RS
Media. Over 50 million RS units have been sold since its release in 2004. The
Robosapien V2 is larger robot with an expanded list of English verbalizations. RS V2
also introduced basic color recognition sensors, grip sensors, sonic sensors and a
wider variety of possible movements [12]. The RS Media has a body very similar to
RS V2, but an entirely new processing unit based on a Linux kernel. RS Media is
equipped to be a media center and has the ability to record and playback audio,
pictures and video.
While many mass-produced humanoid robot kits or humanoid robots developed so
far are inarguably flexible and capable to perform movements that are human alike,
many of them are small in size, or do not possess the ability to produce audio
independently. As it is intended in this study that the robot used in this research
should be accessible by the public and be able to perform the Thai dance
autonomously without the need of a computer or external speakers. Thus, this makes a
robot selection process currently limited to Robosapien series, in particular the RS
Media. In 2007, a special Robot Extension SDK written by Sun Microsystems was
102 C.C. Fung, T. Nandhabiwat, and A. Depickere
released and it was bundled with the RS Media. 200 were available for sale at the Java
One conference. In order to take the full programming capability of the robot, the RS
Media Java SDK was acquired for this research.
Thai dance, or “Ram Thai” in Thai language, is the main dramatic art form of
Thailand and it is considered as one of countless worldwide dance types in existence.
Thai dance on its own, likes many forms of traditional Asian dance, can be divided
into two major categories that correspond to the “high art” or classical dance, and,
“low art” or folk dance [14]. Thai classical dance includes main dance forms like
“Fawn Thai” accompanied by folk music and it varies according to the style of the
particular region. “Khon” is the most stylish form of Thai dance performed by group
of non-speaking dancers while a story is being told by a chorus at the side of the
stage. “Lakhon” has costumes which are identical to Khon, but Lakhon dance
iThaiSTAR – A Low Cost Humanoid Robot 103
movements are more graceful, sensual, and fluid. The upper torso and hands are
particularly expressive with movements portraying specific emotions; etc. Thai folk-
dance includes main dance forms such as “Ram”, which are originated from numerous
regional dances. “Likay”, contains elements of pantomime, comic folk opera, and
social satire. They are generally performed against a simply painted backdrop during
temple fairs. “Ram Muay” is the ritualized dance that takes place before Southeast
Asian kickboxing matches such as Muay Thai. “Wai Khru” is a ritualized form of
dance meant to pay respect to, or homage to the “khru” or teacher. It is performed
annually by Thai classical dance institutions; etc.
As there are many varieties of Thai dances, it is vital that a careful selection of the
one to be used in this research is properly consulted with a Thai dance professional.
Thai dance ranges from a simple movement to complex movements that if an
inappropriate one is chosen, the robot might not be able to perform due to its physical
limitations.
After a careful study, iThaiSTAR is currently capable to demonstrate the Thai folk-
dance performance called “Ram Wong”. The art of Ram Wong is originally adapted
from “Ram Tone”, where it uniquely specifies that dancers must follow the rhythm of
the tone drum which is especially made for the dance. Ram Wong is one of Thailand’s
most popular folk dances. It has been popular among Thai people in some regions of
Thailand. This Thai dance used to be played with the performance of Thai traditional
music instruments consists of “Ching” (a kind of Thai important percussion
instrument made of metal), “Krab” (a kind of Thai important percussion instrument
made of wood), and “Tone” (Thai drum made of carved wood or baked clay). In
1940, Ram Wong dance pattern influence spreads to the other regions in Thailand. It
has effectively become very popular among the people in every region of the country
and has created the rhythmic dialogue to sing together with the performance of Thai
music. Basically, the dialogues are about persuasion, teasing, praising, and parting.
Ram Wong has been very popular among the people in central region of Thailand
during the World War 2 (1941-1945). Owing to the support of the government, Ram
Wong has been reformed by the Fine Arts Department of Thailand in 1944. At that
time four new rhythmic dialogues had been created. The songs and music instrument
had been adapted to be more contemporary. Some movements such as “Tar Sod Soi
Ma La”, “Tar Ram Sai”, etc., had been settled as standard patterns of Ram Wong. The
name Ram Tone (the tone dance) had been changed into Rum Wong (the circle dance)
because of its movement which people was often dancing around like making a
circular movement. Later, Premier Piboonsongkram created six more new rhythmic
dialogues introducing the Ram Wong as a modern Thai dance. Finally, the Ram Wong
has ten songs with specific movement patterns where the dancers moving round in a
circle. The song lyrics refer to the goodness of Thai culture and the ability and daring
of Thai warriors. After World War 2, Ram Wong has been kept active among people
until now. Ram Wong is widely performed not only by Thai people, but also
foreigners in dancing ball. Many Ram Wong songs have been created by following the
ten specific standard movement patterns.
104 C.C. Fung, T. Nandhabiwat, and A. Depickere
The two Ram Wong songs that iThaiSTAR could perform, each acquired from the
ten specific preserved standard movement patterns. The two songs are “Ngam Sang
Duen”, which obtained the standard movement called “Tar Sod Soi Ma La” (imitated
from the actual actions of the local people making a flower garland by having one of
the hands holding a cotton string and another hand pulling a flower on the string
outwards from the body towards the side). The song “Ram Si Ma Ram” synchronizes
with the standard movement called “Tar Ram Sai” (imitated from the actual actions of
people trying to persuade one another by stretching both arms almost parallel to the
ground and twisting both hands up and down opposite one another) This is shown in
Fig. 1.
Fig. 1. Main movements of “Tar Sod Soi Ma La [15]” and “Tar Ram Sai [16]”
used, the servos for the legs will not be turned on thereby making the robot could only
move its hands, head and upper body.
5 Conclusion
This paper has reported the initial phase of the development of iThaiSTAR, a low cost
humanoid robot for the purpose of entertainment and teaching of Thai dances. The
research work so far has been the investigation, design and development of using an
off-the-shelf toy robot, WowWee’s RS Media for the purpose of this study. The
public appearances of iThaiSTAR have attracted much attention from the audience
and the media at Thailand. Informal comments have been positive and expert
comments on the dance performed by the robot are good. It is expected that the
research will be embarked into its second phase with further development of content
and objective assessment of the effectiveness of iThaiSTAR. Another related project,
R4RE, is currently under development. The project will use similar robots for applied
edutainment programs at rural communities in Thailand and Australia.
106 C.C. Fung, T. Nandhabiwat, and A. Depickere
References
[1] Curriculum Council of Western Australia, Curriculum Guides and Elaborated Guides
[Accessed: December 11, 2007], http://www.curriculum.wa.edu.au/pages/
curric_guides/index.html
[2] WowWee [Accessed: December 11, 2007], http://www.wowwee.com/
[3] Bangkok ICT Expo [Accessed: December 11, 2007],
[4] http://www.bangkokictexpo.com/
[5] Michalowski, M.P., Sabanovic, S., Kozima, H.: A Dancing Robot for Rhythmic Social
Interaction. In: Proceedings of the 2nd Annual Conference on Human-Robot Interaction
(HRI 2007), pp. 89–96. ACM, New York (2007)
[6] LERN: Leading Edge Robotics News, Keepon Keeps Dancing On, Robot, issue 8, p. 15
(Fall 2007)
[7] Michalowski, M.P., Sabanovic, S., Michel, P.: Roillo: Creating a social robot for
playrooms. In: Proceedings of the 15th IEEE International Symposium on Robot and
Human Interactive Communication (RO-MAN 2006), pp. 587–592. University of
Hertfordshire, United Kingdom (2006)
[8] Shinozaki, K., Iwatani, A., Nakatsu, R.: Concept and Construction of a Dance Robot
System. In: 2nd International Conference on Digital Interactive Media in Entertainment
and Arts (DIMEA 2007), Perth, Australia (2007)
[9] Shinozaki, K., Iwatani, A., Nakatsu, R.: op. cit.
[10] Kosuge & Wang Laboratory, “MS DanceR,” Research [Accessed: October 29, 2007],
http://www.irs.mech.tohoku.ac.jp/research/RobotPartner/
dance.html
[11] Randerson, J.: Japanese teach robot to dance. Guardian Unlimited (August 8, 2007)
[Accessed: November 28, 2007],
[12] http://www.guardian.co.uk/technology/2007/aug/08/robots.japan
[13] Tanaka, F., Cicourel, A., Movellan, J.R.: Socialization between Toddlers and Robots at an
Early Childhood Education Center. Proceedings of the National Academy of Sciences of
the USA (PNAS) 104(46), 17954–17958 (2007)
[14] Hosein, K.J.: Review of the RS Media. RoboCommunity (January 20, 2007) [Accessed:
September 22, 2007], http://www.robocommunity.com/article/10627/
Review-of-the-RS-Media/
[15] RoboGuide, RS Media, RS Media RoboGuide (February 5, 2007) [Accessed: September
22, 2007], http://www.evosapien.com/robosapien-hack/nocturnal2/RSMedia
[16] Mahidol University, The Rituals and Traditions of Thai Classical Dance, Mahidol
University [Accessed: September 23, 2007],
[17] http://www.mahidol.ac.th/thailand/classical-dance.html
[18] Banramthai.com, Ngam Sang Duen, Banramthai [Accessed: November 3, 2007],
http://www.banramthai.com/html/rw_kham.html
[19] Banramthai.com, Ram Si Ma Ram, Banramthai. [Accessed: November 3, 2007],
[20] http://www.banramthai.com/html/rw_ramsi.html
The Study on Visualization Systems for
Computer-Supported Collaborative Learning
1 Introduction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 107–113, 2008.
© Springer-Verlag Berlin Heidelberg 2008
108 S. Kim, H. Kim, and S. Han
interactions graphically, so that a teacher can monitor and evaluate their collaborative
learning activities in CSCL environment. For our experiments, we use a web-based
discussion group board and the visualization tool to evaluate students’ interaction
activities. We also propose the learning steps that describe how to use the visualization
tool effectively to promote interactions and feedbacks. Our experiment results show
that the visualization tool is clearly useful for a teacher to understand students’
collaborative interaction and thus learning activities.
grasp interactions among learners is possible it. The feedback engages with learner’s
discussion activities, gives a proper reinforcement and promotes interactions among
learners more briskly.
In this study, a system that we want to develop will follow a collaborative discussion
learning model step. So we first decide to construct web site having a discussion board.
Fig. 2. The discussion board for CSCL(Korean version) and visual page
The visualization page only will be offered to the instructors like as Figure 3, whenever
they click view button. It will show relationship of interactions as visual form by
bringing the learner’s log information. Human’s shape and names as shown in the
bottom are learners who took part in the discussion. And we let the size of spot cor-
responded proportionately to amount of interactions. Because interactions among
learners are described to a curve, learners can be connected between sender and
The Study on Visualization Systems for Computer-Supported Collaborative Learning 111
receiver. And the more amount of message, the more line color becomes dark. If an
instructor put a mouse on the spot, he views learner’s name. If they put a mouse on the
curve, a message that a curve means appears in the part of right bottom. An instructor
can give the feedback by grasping discussion active degree and conditions of
participation.
4 Experiment
In this study, we progressed web discussion collaborative learning with 112 students in
Gyeong-in National University of Education, who were juniors. The participants in
subject discussion were made up 74 students 2 classes, them in debate discussion were
made up 38 students of a class. The first subject of subject discussion is ‘Would
artificial intelligence lead our life to be comfortable?’ The second subject is ‘How to
protect my pc safely?’ Finally, the subject of debate discussion is ‘In ubiquitous age, do
we need school that is physical space?’ when discussing, we gave 40 minutes. And in
each class, there was a student who played teacher’s role and were 3-4 assistant
teachers. As analysis of visual tools, we focused on convenience, effectiveness and
easy when using. Because this visual tool is given not to learners but to teachers
according to learning steop, we surveyed and interviewed with 13 students who played
teacher’s role.
Figure 4 shows answers that 13users playing real teachers gave in an aspect of effec-
tiveness, 84% of them gave positive answers. Also in an aspect of convenience 100%,
and in that of convenient using method, 84% of them gave. Therefore, we realized that
this monitoring tool was useful to observe learners and promote a discussion. Also,
according to interview of users, they were thinking that this tool is effective, as follows.
112 S. Kim, H. Kim, and S. Han
Acknowledgement
This work was supported by the Korea Research Foundation Grant funded by the
Korean Government (MOE) (KRF-2007-B00082).
References
1. Bastiaens, T.J., Martens, R.L.: Conditions for web-based learning with real events. In:
Abbey, B. (ed.) Instructional and cognitive impacts of web-based education, pp. 1–31. Idea
Group, Hershey (2000)
2. Davices, D.: Learning network design:coordinating group interactions in formal learning
environments over time and distance. In: O’Malley, C. (ed.) Computer supported
collaborative learning, pp. 101–124. Springer, London (1994)
3. Gauthier, G., Frasson, C., VanLehn, K.: In Intelligent Tutoring Systems. In: Proceedings of
the 5th International Conference, pp. 325–333 (2000)
4. Jang, H.W., Suh, H.J., Moon, K.A.: Analysis of Collaborative Learning Model and
Collaboration Tools in e-Learning Environment (2005)
5. Lee, I.-S., Leem, J.H., Jin, S., Sung, E.M., Moon, K.A., Seo, H.J.: Analysis of collaborative
learning behaviors and the roles of collaborative-learning agent. In: Proceedings of e-Learn
2004. Association for the Advancement of Computing in Education, pp. 2748–2754 (2004)
6. Marzo, J.L., Peña, C.I., Aguilar, M., Palencia, X., Alemany, A., Vallès, M., Johè, A.:
Adaptive multiagent system for a web-based tutoring environment - final report March
(2003)
7. Lin, X., Hmelo, C., Kinzer, K., Secules, J.: Designing Technology to Support Reflection.
Educational Technology Research & Development 47(3), 43–62 (1999)
8. Kim, M.-Y., et al.: New Flash Masters Of Korea 2005. The electronic Times (2005)
9. Gwon, S.-J., Kim, D.-S.: Development of Computer Supported Collaborative Learning
Platform Prototype. Korean Association for Educational Information and Broadcasting
2001 7(1), 119–145 (2001)
Computer-Assisted Paper Wrapping with Visualization
1 Introduction
There are many ways to wrap not only box-types of goods but also various shapes of
goods. The paper wrapping is performed by experts who have a lot of wrapping experi-
ences generally. It is difficult for beginners to perform the paper wrapping because the
paper wrapping is performed with various shapes of goods and a sheet of paper with
the various sizes; additionally, the wrapping process is not unique and there are many
wrapping processes. Therefore, the instruction books about paper wrapping are utilized
for beginners as a tool for paper wrapping support. However, the instruction books do
not show enough information about paper wrapping as assistant means for beginners:
for example, only 2-D images about wrapping process may be shown or one of the
examples is illustrated. In order to support paper wrapping activities, the method for
representing paper wrapping processes effectively is necessary.
A lot of studies related to the paper folding, which is called Origami, have been
conducted until now. Most of these studies are carried out by mathematicians: they at-
tempt to elucidate the geometrical properties of Origami by use of mathematical meth-
ods [2,3]. The study which proposed the method for generating the folded Origami
model from the information of crease lines on crease pattern has been conducted [5]. In
this study, the folding processes are not so considered although the properties of crease
lines are elucidated. As a study representing a folding process of Origami in 3-D vir-
tual space, Miyazaki, et al. [4] developed a virtual interactive manipulation system for
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 114–125, 2008.
c Springer-Verlag Berlin Heidelberg 2008
Computer-Assisted Paper Wrapping with Visualization 115
simple folding operations. A user folds a piece of papers by operating a mouse, and
then the system constructs folded paper model. The user can get the objective model by
determining all folding operations.
The most studies about Origami support are mainly utilized for Origami designers.
In these studies, it is necessary to design folding operations, crease pattern and oth-
ers exactly. On the other hand, in the paper wrapping, folding operation is constructed
by stages, based on the information about the shape of target goods and the size of
wrapping paper. Each folding operation depends on a condition of wrapping paper.
Therefore, from a viewpoint of paper wrapping support, the method for managing these
wrapping processes effectively is necessary. In addition, the way of searching effective
wrapping processes which satisfy the demand of user is necessary to fulfill the require-
ment of beginners as assistant means of paper wrapping.
In this paper, we propose an approach to support the paper wrapping activity by
managing the processes of wrapping a target object with a piece of papers. In order
to support paper wrapping, it is necessary to represent, generate and compare various
wrapping processes. Therefore, in our approach the wrapping processes are managed
hierarchically with tree structure. In this way, we can also delete non-effective wrapping
processes easily and it is easy for users to check out various wrapping conditions. Fold-
ing operations are constituted by simulating in virtual space, called an internal model,
that describes both information of an object and deformation processes of wrapping pa-
per, and then the tree structure is constructed by generating folding operations over and
over. In addition, we propose a method for searching effective and feasible wrapping
processes based on the demand of user. The purpose of this study is to develop a system
that supports the paper wrapping visually by CG animation in wrapping processes.
2 Framework
2.1 Wrapping Process
Even when a target object and a piece of wrapping papers are given, the wrapping
process is not decided uniquely. In fact, individuals wrap materials with the different
ways in different applications even if they used the same object and the same wrapping
paper. So, the required wrapping processes depend on the demands of users.
In order to support the paper wrapping activity, it is necessary to design an effective
wrapping process which satisfies the following conditions: there are no exposed object
faces and visible backsides of a wrapping paper, and the process is physically foldable.
Thus, it is necessary to analyze whether designed wrapping processes are effective or
not and well manage various wrapping processes. To deal with these issues, a powerful
method for specifying the deformation processes of wrapping paper and managing the
effective knowledge is necessary.
The conditions which are used to judge whether the given crease pattern is feasi-
ble and/or foldable in three-dimensional space were argued in Origami mathematical
fields[1]. The conditions are called fold-ability, and are dependent on the position and
folding angle of the crease lines on crease pattern. Although only the necessary condi-
tion for fold-ability has been provided, the sufficient condition has been not so. Thus,
116 K. Matsushima, H. Shimanuki, and T. Watanabe
there are a lot of crease patterns which cannot be folded although they satisfy these
conditions.
We construct a structural state space which represents wrapping processes on a tree
structure, called stage tree to manage the effectiveness of wrapping processes for users.
Each node in the tree represents one stage in each wrapping process. Applying the
folding operation, which corresponds to an edge of tree, to each stage, a new stage is
generated. Wrapping processes are designed by constructing a stage tree while generat-
ing folding operations. In this way, many effective or non-effective wrapping processes
are managed hierarchically and users can view the wrapping condition which each stage
attached with. In addition, it is easy to cut out the stages which are not necessary for
users. So, it is allowed to judge fold-ability by actually generating folding operations in
a virtual space.
paper wrapping, which Miyazaki, et al. [4] have already proposed. This model consists
of vertices V, edges E and faces F as the basic data elements. An edge ei, j ⊆ E has in-
formation about its coordinates and the angle θi, j between two faces fi , f j . If θi, j = ±π
then the edge(crease) means valley/mountain folding. In addition, the model has a data
structure which can represent the overlapping faces on the same plane. An example of
this structure is shown in Fig. 2. The structure groups the faces on the same plane and
holds the order of overlapping by a face list ( f2 → f3 ). A wrapping operation is applied
to faces on the same plane. Each face is stored at a face list in the order of the normal
vector calculated from the object face on the same plane.
f 2 f3 Face list
f2 f1
f3 f1 Face stack
Stage Tree. When the information about an object and a wrapping paper is given,
there are a lot of possible wrapping processes. Moreover, searching the wrapping pro-
cess which satisfies the demand of user from many processes is required. In Fig. 3, a
tree structure (stage tree) in which nodes represent wrapping stages and edges represent
folding operations is shown. The tree structure obtains the relationships among wrap-
ping stages and a record of wrapping processes. The route that arrives from the root to
the leaf shows one sequence of wrapping process. A new stage is generated by applying
any folding operation to an existing stage.
S1
S
2 S 3 S
4
S 5 S 6 S7 S 8 S9
S k ・・・ Sn
predict the final wrapping condition from the incomplete wrapping stage. Namely, pre-
dicting the final wrapping condition is equivalent to calculating the wrapping process
completely. In addition, there are various demands from users for wrapping support:
e.g. appearance, degree of difficulty, minimum required size of a wrapping paper, etc.
There are no wrapping processes which satisfy various demands at once.
In our method, a user determines evaluation criterions for generating a stage tree. In
other words, the wrapping processes not required are cut off from a stage tree. Some
wrapping processes which satisfy a demand of user should be provided, even if the most
effective wrapping process is not specified. Additionally, as a depth of stage tree is deeper,
a stage on the way exerts less effect on final wrapping condition. Therefore, a user can de-
termine the depth of sub-stage tree which the folding operations are applied to all foldable
faces. All stages which belong to the sub-tree are never cut off. In this way, the accuracy
of tree search is determined and the most effective wrapping process does not have to be
searched exactly. In Section 3.4, our pruning algorithm is described in detail.
Start
Generation of
folding operation
no
Renewal internal model
yes
End
Fig. 4. Flowchart of generating the feasible stages from a certain stage in one wrap-folding
operation
First, a target face for a folding operation is decided in each stage. All target faces
should be foldable at least for one face of object. There is a plural number in the faces
that can become a target face in each stage, and a stage tree branches off by depending
Computer-Assisted Paper Wrapping with Visualization 119
on how to choose these faces. Second, the folding operation is actually generated. The
folding operation is constituted by generating crease lines, dividing faces and moving
the divided faces. We show the method for generating crease lines in detail in Sec-
tion 3.3. Finally, the face list in our internal model is updated. Whenever a folding
operation is generated, a face list is updated.
As a wrapping stage becomes more complex, the number of the faces stored in a face
list increases. In this case, each folding operation is applied to all stored faces in a face
list in which target face is stored in the same way. These processings are repeated as far
as there are the faces in which folding operations are calculated.
f 3′
S1
f 2′ f4 S 2 S 3 S 4 S
5
f1′
f5 ( f 4 , f1′) ( f 4 , f 3′) ( f 5 , f1′) ( f 5 , f 2′) ( f 5 , f 3′)
folded face f p and wrapped face f0 , each stage is generated by a folding operation. The
following equation shows all the combinations in an example of Fig. 5.
( f p , fo ) = {( f4 , f1 ), ( f4 , f3 )}, {( f5 , f1 ), ( f5 , f2 ), ( f5 , f3 )}
In this case, the number of combinations between f p and fo is 5. In the stage tree of
Fig. 5, the grayed node S2 is calculated, and five child nodes are generated from S2 .
Start
input : fp :folding target face
f o′ :wrapping target face
Generation of a crease line output : a stage applied by folding operation
on based on reference axis
fp
corresponding to the crease. Otherwise, by adding other folding operations, the colli-
sion among faces is evaded like Fig. 1(b) and Fig. 1(c). In this section, we show the
processing in detail.
Rotation axis
f o′
Confliction point
fp eb Unification
Confliction face
edges connected from eb sequentially around a confliction point. Furthermore, βi is the
plane angle between ei and ei+1 , and satisfies:
N−1
∑ βi < 2π (1)
i=0
1. eb is defined as the base line and the crease line e0 having the folding angle θ0
provided from eb is generated.
2. The following procedures are repeated for all i s (0 ≤ i ≤ N − 1).
(a) The crease line ei+1 is generated at a position of the angle γi from a base line
calculated by the following equation:
i
γi = ∑ βj (2)
j=0
.
(b) θi+1 ← θi+1
3. The crease line eN having the folding angle π is generated at a position of the angle
α from a base line provided by the following equation:
2π − γN−2
α= (3)
2
4. A line between f p and confliction face are divided at the confliction point, and then
unification edge and crease line eN+1 are generated.
An example of generating crease lines is shown in Fig. 8. First, the base line is specified
by the extended line of eb , and the crease line e0 having the folding angle θ0 is gener-
ated. Second, the crease line e1 having the folding angle θ1 is generated at a position
of the angle β0 from the base line. The angle θ1 equals to the angle of e1 . In the same
way, the crease line e2 having the folding angle π − θ2 which depends on the folding
angle θ2 of e2 is generated at a position of the angle β0 from the base line. Finally, it
is necessary to generate the crease line in order to maintain the consistency of folding
ability. This crease line is generated at a position of the angle α from the base line by
using Equation (3). The folding angle is π . Then, the folding angle of the unification
edge is made to be 0 by dividing a line between f p and confliction face. We show the
state of wrapping before and after generating crease lines in Fig. 8. The unification edge
is abbreviated in Fig.8(b).
Some wrapping processes are generated based on the demand of user. The appropriate
numerical wrapping processes which satisfy the demand are provided to users. The hill-
climbing method is used as the basic tree constructing method, and this processing is
divided into two phases. First, sub-stage tree whose depth is lower than d which user
determines is generated. Next, each generated leaf stage is viewed as an initial wrapping
state, and the construction for these stages is performed using hill-crimbing method.
122 K. Matsushima, H. Shimanuki, and T. Watanabe
e4
e′1
e′2
φ1
φ0 α
φ2
θ 0 e0 e′0
β0 e3 π
β1
e1 e2 α
θ1 θ2
Unification Confliction
edge point
(a) Before folding (b) After generating crease lines
S1
S2 S3 S4
S5 S6 S7 S8 S9 S10
: cut nodes
・ ・ ・ ・ ・
Consequently, the number of generated wrapping processes satisfies the demand of user.
Fig.9 illustrates the distribution of search area if the depth of sub-stage tree is 2.
We use two variables Ecomp and Eback to evaluate the difficulty of a sequence of fold-
ing operations and the appearance of complete wrapping states. Since the evaluation
with these criteria is equivalent to the evaluation for edges of stage tree, it is possible
for incomplete wrapping stages to be evaluated. The evaluated value of a stage E is
α Ecomp β Eback
E= + (α + β = 1) (4)
2 N
Ecomp is the number of confliction points when wrap-folding is applied to the previous
stage. Eback is the number of face-groups which have viewable backside in generated
stage and N is the total number of face-groups. These values are added to the evaluation
value of the parent stage, and the value is viewed as the evaluation value of the target
stage. Because not only the final wrapping condition but also the wrapping condition on
the way is important as for the paper wrapping, it is proper to search effective wrapping
processes with these end-points. Each coefficient in Equation (4) is determined by a
user in order to make the generated wrapping processes effective.
Computer-Assisted Paper Wrapping with Visualization 123
4 Experimental Result
Our approach for designing a sequence of wrapping processes was tested to validate
the evidence of availability. We performed our experimentation from a viewpoint of
construction of stage tree and the number of generated stages.
18000 35
16000
30
14000
se
ss
esg 12000 25 ec
tas or
p All generated stages
ngi
de10000 20
ta pp
re 8000 15 ar Effective wrapping
ne w processes
g 6000 ev
l it
A 10 ce
4000 ffE
5
2000
0 0
1 2 3 4 5 6 7 8 9 10 11 12 13
The depthof sub stage tree
when the depth d is less than 9. The different number between all generated stages and
effective wrapping processes are plotted in Fig. 11.
4.3 Discussion
We proved that many unnecessary stages are reduced with our proposed pruning algo-
rithm through experiments. On the other hand, effective wrapping processes are also
reduced. Effective wrapping processes were not generated in this experiment when the
depth of sub stage tree is less than 9. Depending on the value that the user determined,
there is the case that an effective wrapping process is not generated at all. So, we ar-
rive at our conclusion that we need a more effective evaluation method which makes it
possible to predict whether the wrapping process is complete.
In addition, the wrapping processes are regarded as different processes when the or-
der of wrapped face is different, even if generated last wrapping state is equivalent.
Therefore, the difference between generated wrapping processes was small, and many
wrapping processes whose last wrapping condition was equivalent were generated dif-
ferently in this experiment. From a viewpoint of paper wrapping support, most effective
wrapping process does not have to be specified closely. All we need to do provide the
guidance principle of wrapping processes which satisfy the demand of user. It seems
that it is effective to use the information about the crease pattern on a wrapping paper
which is mainly used in research field about Origami. Because the crease pattern has
only the crease line information of the wrapping paper, and the information about over-
lapping faces is not represented. In other words, it means that the crease pattern has
only the wrapping essential information. We can treat only the essential property of the
wrapping processes which are necessary point for paper wrapping support by using the
information about crease pattern as a basic evaluation criterion.
5 Conclusion
We have proposed an approach to represent and design wrapping processes in paper
wrapping as support. As the main result, the method for managing multiple wrapping
Computer-Assisted Paper Wrapping with Visualization 125
processes with a tree structure and formalization of folding operations for paper wrap-
ping were proposed. The proposed method made it possible to design various foldable
wrapping processes and represent a sequence of wrapping processes by CG.
In addition, the pruning algorithm for a stage tree is explained. The evaluation cri-
terions are determined by user and unnecessary wrapping processes are cut off. Many
wrapping processes which were unnecessary for user were reduced, but some effec-
tively generated wrapping processes were also reduced. The result of our experiment
clearly shows that the method for predicting the final wrapping condition is necessary.
In this paper, we have considered the wrapping process by focusing on the faces of
both the object and wrapping paper in generating folding operations. Although our ap-
proach makes it possible to design multiple wrapping processes, generated stage tree
has many useless information and we does not consider the initial position of wrapping
paper and object. From the viewpoint of paper wrapping support, we need the method
for analyzing guidance principle of generating stage tree so to decrease useless infor-
mation and make generated wrapping processes effective. A foreseeable extension of
this research would be to design wrapping processes using information of crease pat-
terns. To use information about crease pattern is also useful to estimate optimally initial
position of wrapping paper and object.
References
1. Belcastro, S., Hull, T.: Modeling the Folding of Paper into Three Dimensions Using Affine
Transformations. Linear Algebra and its Applications, 466–476 (1999)
2. Hull, T.: On the Mathematics of Flat Origamis. Congressus Numerantium (100), 215–224
(1994)
3. Kawasaki, T.: R( )=1. In: Proc. of the 2nd International Meeting of Origami Science and
Scientific Origami, vol. 3, pp. 31–40 (1994)
4. Miyazaki, S., Yasuda, T., Yokoi, S., Toriwaki, J.: An ORIGAMI Playing Simulator in the
Virtual Space. The Journal of Visualization and Computer Animation 7, 25–42 (1996)
5. Shimanuki, H., Kato, J., Watanabe, T.: Constituting Feasible Folding Operations Using In-
complete Crease Information. In: Proc. of IAPR Workshop on Machine Vision Applications,
pp. 68–71 (2002)
Hangeul Learning System
Abstract. This system is not only for a foreigner but also for everyone in
Korea who doesn’t know Hangeul (Korean language). It is difficult to study
Hangeul themselves without any helper. This paper presents the AR based
system that helps people to learn basic Hangeul letters and pronunciations in
their home without any helper by applying the characteristics of consonant
and vowel. We also suggest the Word Studying Methods using the proposed
system. At this time, it is developed based on the pattern matching function
of ARToolKit, we plan to improve the system by applying the character
recognition function.
1 Introduction
The language learning environment has been sustained with text–oriented environment
for many years. It has been changed slowly to match the today’s digital era. The recent
technology trend is not simple information transmission but overall information
transmission (e.g. ask for the sight, the sound, mutual communication). That is the
latest trend of acquisition of knowledge in the education field. Although some
traditional instructions use multimedia, but it has limited users’ participation. It did
nothing else than keyboard typing or mouse click. In this paper, to overcome these
limits and to improve the current language learning systems, we propose a new
approach to assist learning the Hangeul language using Augmented Reality.
Korean language ‘Hangeul’ is one of the most interesting languages. Hangeul
creates words by combining the basic buildingblocks, Jamo. At least two and often
three of Jamo are placed on the pattern to form a word. Depending on the pattern and
Jamo, users can create various words. Since this type of arrangement is not usually used
in other language, foreigners may have difficulty to learn Hangeul. We propose the
Hangeul Learning System using AR to overcome this problem
This is followed by related works in section 2 and details of the main components in
Sections 3, and results illustrating for several different scenarios are presented in
Section 4. Section 5 is Conclusion.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 126–134, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Hangeul Learning System 127
2 Related Works
The most well known system for learning languages using AR is the system that can
learn Japanese Kanji [5]. After this system was introduced, few systems that could
assist users to learn languages have been developed. There are also few education
systems using AR. There are systems to teach molecular structure in the field of
chemistry or atmosphere circulation in earth science. They are designed to use in
school and use visual information to increase interests of users. In this section, we
present the Japanese Kanji learning system and two science education systems.
It was made for helping users to learn Japanese Kanji [5]. Users learn Kanji characters
by playing the matching games developed using handheld Augmented Reality
technologies. A user supposes to find the corresponding answer when the question is
given on the screen.
Many people may think ARbased education systems are only for children. But the
presented system in this section is not for children but for welleducated people. Users
can see the 3D structure of the molecular using the system [1].
The system is formed with browser, tagtoggle, cleaner and molecular marker. Users
can see molecular structure and atom they want. Traditional text based chemistry
education can not see as 3D structure of a molecular, but the system shows 3D structure
of the molecular and helps users to understand the molecular structure better.
(a) (b)
Fig. 4. The chemistry education system (a) GUI overview with menu button and main menu and
molecule structure (b) 3D visualization of the molecular structure
Hangeul Learning System 129
This system shows principles that form rain and cloud as shown in Figure 5 [4]. The
system is good for children. It can simulate the water circulation through scenarios.
The system contains auditory part and visual part. The system provides a rain sound
to improve the learning experience. The system also induces user’s participation in sc
enarios. Users can modify the position of a marker to interact with the system.
(a) (b)
Fig. 6. (a) Geological features; (b) Using Augmented Reality for Teaching EarthSun Relation
ships to undergraduate Geography Students
Another similar system is the Solar-System and Orbit Learning system [2]. Using
this system, users can understand why earthquake occurs and learn volcano topography
and name of nature of soil. Users can also learn the relationships between Earth and
Sun by using the system shown in Figure 6 (b) [3]. The relationships are represented in
3D, so users can view at various positions and orientations.
130 J.w. Jung and J.w. Lee
3.1 Approach
The Hangeul Learning System is able to recognize and to express variety of characters
using the characteristics of Hangeul. When a person uses this system, the marker
representing each character must be registered. For example, the word ‘가’ can be
recognized by the system when the corresponding marker is registered. However, the
user put two markers, ‘ㄱ’ marker and ‘ㅏ’ marker together, to create the word as
shown in Figure 7.
The same two markers can be combined to create different words such as ‘고’‘나’ ,
‘거’ ‘ㄱ’
and . Since marker could rotate, it would be used for words requiring the
‘ㄴ’
marker ‘ㅏ’ . Likewise, marker could be used for words requiring ‘ㅓ’ or
ㅗ’ markers. This approach can reduce the number of markers needed in the system.
Learners can learn many words using few markers. Figure 8 and Figure 9 describe the
approach by using characters of English and Japanese to help readers understand the
concept.
3.2 Functionalities
The Hangeul Learning System consists of five areas as shown in Figure 10. When a
user put markers on the area 1, the system recognizes the word and shows the word by
displaying the corresponding text, image or 3D object at the area 2. The system
recognizes words by registering corresponding markers in the system and by using the
pattern matching part of ARToolKit . When the user touches the marker on the area 4,
the corresponding sound is produced. When the user touches the marker on the area 5,
the phonetic sign is displayed on the area 3.
Hangeul Learning System 131
4 Experiments
ㄴ ㅏ
Figure 11 (a) is initialization screen. A user put ‘ ’ and ‘ ’ to create ‘ ’ 나
ㄴ ㄱ
(Figure 11 (b)). The marker ‘ ’ can be also used for ‘ ’ by rotating it as shown in
132 J.w. Jung and J.w. Lee
Fig. 13. An image and a 3D model used to represent the given word
5 Conclusion
In this paper we described the Hangeul Learning System that was developed based on
Augmented Reality. The system has two main characteristics. The first one is creating
the words by using fewer number of markers than the number of constants and vowels
of Hangeul. The second is providing visual and auditory outputs to users. For visual
outputs, 2D images and 3D models are used for the current system. We will plan to add
the functionality that can provide related animation to users. As a result Hangeul
Learning System can offer learners better ways to learn basic characters of Hangeul.
The current system also has limitations. The current system only deals with words with
one character. We will plan to provide users to learn words with more than one
character. To reduce the number of markers registered, we plan to use the character
recognition approach instead of the pattern matching approach. Since we create words
with the given marker, the character recognition approach could be reliable.
Acknowledgments
This work was sponsored and funded by Korea Game Development & Promotion
Institute as Korean government project (Ministry of Culture and Tourism).
References
[1] Almgren, J., Carlsson, R., Erkkonen, H., Fredriksson, J., Møller, S., Rydgård, H., Österberg,
M., Fjeld, M.: Tangible User Interface for Chemistry Education: Portability, Database, and
Visualization. In: SIGRAD 2005, pp. 19–24. Linköping University Electronic Press (2005)
[2] Woods, E., Billinghurst, M., Looser, J., Aldridge, G., Brown, D., Garrie, B., Nelles, C.:
Augmenting the science centre and Museum Experience. In: Proceedings of the 2nd
international conference on Computer graphics and interactive techniques in Australasia
and South East Asia (GRAPHITE 2004), pp. 230–236. ACM Press, Singapore (2004)
134 J.w. Jung and J.w. Lee
[3] Shelton, B.E., Hedley, N.R.: Using Augmented Reality for Teaching Earth-Sun
Relationships to Undergraduate Geography Students. In: Augmented Reality Toolkit, The
First IEEE International Workshop, pp. 8–16. IEEE Press, Germany (2002)
[4] Kim, J.H.: Develop experience learning contents and research the application of a field
based Augmented Reality. In: Korea education & research information service (keris), seoul
(2005)
[5] Wagner, D., Barakonyi, I.: Augmented Reality Kanji Learning. In: Proceedings. The
Second IEEE and ACM International Symposium on Mixed and Augmented Reality, pp.
335–336. IEEE Press, Los Alamitos (2003)
An Ajax-Based Terminology System for E-Learning 2.0
College of Info. Tech. & Comm., Qufu Normal University, Rizhao 276826, P.R. China
{cxcsd,whqet,czhhn}@126.com
1 Introduction
A terminology system is to provide timely help for a learner when he encounters
unfamiliar professional new. This is becoming more and more indispensable with the
development of the nonlinear organized web-based learning environment. There are
several terminology systems available such as Chinese Computer Terminology
System [1] and the basketball dictionary in website hoopchina[2], they have done a
lot of fruitful work, yet e-learning based on the system may be hindered by the
inherent disadvantages as follows.
1. Poor interactivity. Based on the traditional network design model [3], if a user
queries a professional term, he has to stare at a blank screen for several seconds in
a poor web condition because these systems must refresh the full page, which
tremendously decreases his learning interest and interrupts his learning process.
2. Low efficiency. Most of current terminology systems are independent ones, that is,
they are not committed to any web-based course. So, when a user encounters a new
term in an article, he has to open a new page of a terminology interpreting system
for help. That is time-consuming; furthermore, he may be misled by several such
operations.
3. Low revisability of term database. The model of modifying term database of most
current terminology systems is manual. So that it is a hard work to update the term
database. Beside time-consuming, it is also error-importing. This leads to low
revisability.
E-learning 2.0 (as coined by Stephen Downes[4]) takes a 'small pieces, loosely
joined' approach that combines the use of discrete but complementary tools and web
∗
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 135–146, 2008.
© Springer-Verlag Berlin Heidelberg 2008
136 X. Cui, H. Wang, and Z. Cao
services - such as blogs, wikis, and other social software - to support the creation of
ad-hoc learning communities[5]. This paper makes use of Ajax technology to create
an E-learning 2.0 community. By introducing Ajax technology, the proposed
terminology system outgoes others by better interactivity, higher efficiency and
timely modification. Specifically, the following approaches are introduced to improve
the system, 1) automatically add explanation to term, 2) search suggest, 3) save
automatically, 4) preload page.
The rest of this paper is structured as follows. We initially present basic of
terminology system, including shaping education by e-learning 2.0 and basic of Ajax.
Next, the architecture of terminology interpreting system is presented. Afterwards,
this paper provides the implement of the system. Finally, some concluding remarks
and our vision for the next steps are presented.
There are some very interesting changes going on in the world of e-learning. The
changes in e-learning are being driven by two primary forces; the first force is a
steady increase in the pace of business and information creation, the other is the
advent of Web 2.0.
There are three Generations of E-Learning, E-Learning 1.0, E-Learning 1.3, and E-
Learning 2.0.(see Table 1) E-Learning 2.0 is based on tools that combine ease of
content creation, web delivery, and integrated collaboration. Creation of content can
occur by anyone as part of their day-to-day life [6]. In essence, the expectation of E-
Learning 2.0 is that sharing and learning becomes an organic action that is directed
and driven by the learners. Learners are starting to explore the potential of blogs,
media-sharing services and other social software - which, although not designed
specifically for e-learning, can be used to empower students and create exciting new
learning opportunities [7].
The proposed terminology system can provide learners good experience. Learners
not only can query and see the definition of the term but also can modify or add the
definition when they need it. Learners become their learn process’s drivers and
content creators. So, the system is an e-learning 2.0 app.
Asynchronous JavaScript and XML (Ajax), Ajax isn’t a technology [8]. It’s really
several technologies, each flourishing in its own right, coming together in powerful
new ways. Ajax incorporates:
1. Standards-based presentation using XHTML and CSS;
2. Dynamic display and interaction using the Document Object Model;
3. Data interchange and manipulation using XML and XSLT;
4. Asynchronous data retrieval using XMLHttpRequest;
5. JavaScript binding everything together.
Google Suggest and Google Maps are two examples of using Ajax to close the gap
between desktop and web application. They take a great leap forth towards the
richness of standard desktop applications. No longer are you forced to wait five
seconds for the page to reload every time you click on something. Ajax applications
change in real time. They can let you drag boxes around, they can refresh themselves
with new information, and they can completely re-arrange the page without clearing
it. And there's no special plug-in required. Ajax is just a style of design, one that milks
all the features of modern browsers to produce something that feels fewer webs and
more desktop.
The core idea behind Ajax is to make the communication with the server
asynchronous, so that data is transferred and processed in the background. As a result
the user can continue working on the other parts of the page without interruption. In
an Ajax-enabled application only the relevant page elements are updated, only when
this is necessary [9] [10].
Some of the characteristics of Ajax applications include:
1. Continuous Feel: Traditional web applications force you to submit a form, wait a
few seconds, watch the page redraw, and then add some more info. Forgot to enter
the area code in a phone number? Start all over again. Sometimes, you feel like
you're in the middle of a traffic jam: go 20 meters, stop a minute, go 20 meters,
stop a minute ... How many users couldn’t endure too many error message and
gave up the battle? Ajax offers a smooth ride all the way. There's no page reloads
here - you're just doing stuff and the browser is responding.
2. Real-Time Updates: As part of the continuous feel, Ajax applications can update
the page in real-time. Currently, news services on the web redraw the entire page at
intervals, e.g. once every 15 minutes. In contrast, it's feasible for a browser running
an Ajax application to poll the server every few seconds, so it's capable of updating
any information directly on the parts of the page that need changing. The rest of the
page is unaffected.
3. Graphical Interaction: Flashy backdrops are abundant on the web, but the basic
mode of interaction has nevertheless mimicked the 1970s-style form-based data
entry systems. Ajax represents a transition into the world of GUI controls visible
on present-day desktops. Thus, you will encounter animations such as fading text
to tell you something's just been saved, you will be able to drag items around, you
will see some static text suddenly turn into an edit field as you hover over it.
4. Language Neutrality: Ajax strives to be equally usable with all the popular
languages rather than be tied to one language. Past GUI attempts such as VB, Tk,
138 X. Cui, H. Wang, and Z. Cao
and Swing tended to be married to one specific programming language. Ajax has
learned from the past and rejects this notion. To help facilitate this, XML is often
used as a declarative interface language.
5. User first: Ajax was born with the idea to make user much more comfortable when
he using web application by providing user with close-to-instantaneous
performance, rich interfaces and tremendously improved user experience. This is
of great importance to what just mentioned above.
Based on the characteristics of Ajax applications, we chose Ajax to improve user
experience. In web-based learning environment, students’ learning is highly
dependent on learning environment, and vulnerable to outside interference. The
traditional web application often leads students give up their learning because of its
shortcoming for support student’s learning, such as poor interactivity,
unresponsiveness, simplistic interfaces, low usability. Fortunately, Ajax can
overcome this.
Client Server
Initial HTTP Request
Term display Tomcat
HTML, CSS, JavaScript AJAX engine
Term add
JDBC
JDBC
Ajax engine
Js calls XmlHttpRequest
Term modify
• Search suggest: The same as Google suggest, the interface is simply prominently
features a text box to enter search terms. Everything appears to be the same until
you start typing in the textbox. As you type, search suggest requests suggestions
from the server, show you a drop-down list of search terms that you may be
interested in. Each suggestion is displayed with a number of results available for
the given term to help you decide.(see Fig. 3)The function may be very helpful
for students when they type the “term” in the text box to search the mean of term,
for he maybe don’t know the full spell of the term.
• Save automatically: Supposed that submission is failed after you changed the
explanation of the term in term modify module. Don’t be depressed! The system
,
has saved it automatically for you you can find it in the drafts.
140 X. Cui, H. Wang, and Z. Cao
• Preload page: You click the link to the next part of the article after you have read
the first part of the article, you’ll have to stare a blank page for five minutes
sometimes even much longer in the traditional web system, and you’ll lose your
interest at the same time. The terminology system takes into account this point for
you. When you enjoy the first part, the system guesses you’ll read the next part and
load it for you in background. So, when you click the link, the second part of the
article is OK for you!
As depicted in Fig. 1, the client side includes four main modules, term display
module, term add module, term modify module, term query module. Detailed
functional descriptions of the four modules are as follows:
• Term display module: There are two ways to display the interpretation of the
terms. Firstly, the system automatically adds explanation to term, every term in
users’ reading articles will be added links to show explanations as long as the term
is in database, as mentioned above. In addition, the style of the link is different
with the others links in the system and decorated with dotted underline to help
student to identify the link, which can be seen in Fig. 2. So, the student can better
enjoy the article and gain knowledge in the reading. Secondly, learners can visit
main page of terminology system (see Fig. 4), there are the list of new terms, the
catalogs of terms and top 10 of contributors. Learners can see the explanations of
terms, modify them, add them and query them.
• Term query module: If a learner finds a new term in reading and wants to know
the mean of it, he may click a link added automatically by system to see the
explanation; this is the result of query module. In another case, a learner is reading
an article about XSL, he can type “XSL” in the below of user interface to query the
explanation, he also can go to main page of system to get the answer. The system
gives “search suggest” to help user to query the term.
• Term modify module: If learner find some areas of improvement for explanation
of term when he/she see explanation, he can click the modifying link nearby to
modify the explanation. Save automatically function can help learner save the page
edited by learner. In order to guarantee the accuracy of explanation of term, student
user’s modifications need to be audited by teacher users. The explanations of terms
in system are all audited.
An Ajax-Based Terminology System for E-Learning 2.0 141
• Term add module: Learner also can add term in system if find a term need to be
made clear. Save automatically function also is used to prevent mishaps. Student
user’s modifications also need to be audited by teacher users.
At server side, we choose Java and Mysql as functional implementing tools, because
they are efficient and open. When the Sever receives a request form a client, it
processes the task by searching the database. When the work is done, it transmits the
answer to the client or tells it a failing message.
142 X. Cui, H. Wang, and Z. Cao
Input Transmit
key Words return date
Call searchSuggest() to send
Send XMLHTTP Request XMLHTTPRequest to server
Return date
Transmit request
Call handleSearchSuggest() to
transmit return data from server
Sever side
˄a˅ ˄b˅
Fig. 5. (a) Data flow of search suggest module. (b) Flowchart of transmitting data in client.
and use the function setSearch() to put the item choosed to the search text box. The
code of handleSearchSuggest is as follows.
function handleSearchSuggest() {
if (searchReq.readyState == 4) {
var ss = document.getElementById('search_suggest')
ss.innerHTML = '';
var str = searchReq.responseText.split("\n");
for(i=0; i < str.length - 1; i++) {
var suggest = '<div
onmouseover="javascript:suggestOver(this);" ';
suggest += 'onmouseout="javascript:suggestOut(this);" ';
suggest +=
'onclick="javascript:setSearch(this.innerHTML);" ';
suggest += 'class="suggest_link">' + str[i] + '</div>';
ss.innerHTML += suggest;}}}
when he encounters a new term. In order to reach this, the system was designed
simply and conveniently to use.
Take an example, when Tom is reading an article about java servlet in the network
course, although he learned the term java before, the mean of it is always not clear.
Learning process may be interrupted if Tom doesn’t know the mean of it. Fortunately,
in this system, Tom can move mouse to the link; the system will jump out a suggested
box (Fig. 2). Tom can get the simple mean of “java” from the box or click the “detail”
link to get more, he also can click the link “modify” to modify the mean of term if he
thinks it can be improved. The related terms of “java”, such as jsp, Ajax, javascript,
are also displayed in the box; Tom can click the corresponding link to get the mean of
the term which he is interested in.
During his reading, he may want to know the mean of asp, which is occurred to
him. He can use the search input box in the right side of the bottom of page to search
it (Fig. 2). During his inputting, he can get the help of “search suggest” (Fig. 3) and
get a drop-down list of search terms that he may be interested in.
He also can go to main page (Fig. 4) of the system to search the term or modify the
term. He also can add term in system if he finds a term needed to be made clear.
The collection of the users’ feedback and their elaboration will lead to an integrated,
from both pedagogical as well as technological aspect, terminology system. The
evaluation conducted was mainly focused on the usability and acceptability of the
system. The aim was to evaluate the current functionality as well as the interface
usability, in order to obtain results for future enhancements on the system.
The main questions asked were the following: (1) Is the system easy to use? (2) Is
the system efficient to use? (3) Is the system subjectively pleasing? (4) Are you
frequently modify or add the term? (5) Are you frequently interact with others? In
order to survey the results, we set the answer is A. Strongly fit; B. mostly fit; C. fit; D.
not fit.
40 S
Strongly fit
rse 30 Mostly fit
ws Fit
na Not fit
fo 20
m
u 10
N
The Fig. 7 illustrates the result of the evaluation. We got 100 answers. For the first
question, 86% of the users stated that the terminology system is easy to use, others
thought the system is not easy to use. Concerning the efficiency of the system, 82% of
the users (18% strongly fit, 29% mostly fit, 35% fit are included) thought that the
system can effectively help their learning. When were questioned subjectively
pleasing, only 9% stated that they are not willing to use the system, which indicates
success of the system. However, 28% of the users rarely contribute to the terminology
system, including modifying or adding term. This shows the limitation of arousing
participation interest. And the last question indicates the interaction among the users
should improve in future.
Acknowledgements
This work is partially supported by the Natural Science Foundation of China under
grant Number 10771120, Key Project of Culture Department of Shandong Province
and Instructional Innovation Project of Education Department of Shandong Province
under Grant NO.B05042; Project of Outstanding backbone University teacher
international cooperative Cultivation of Shandong Province, P.R.China.
The authors would like to give their sincere thanks to Prof. Shumin Kang, Prof.
Siwei Bi, Dr. Yuxia Lei, Dr. Zili Zhou, Dr. Xiulan Li and other colleagues and friends
for the pleasant cooperation in Project of Instructional Innovation of Education
Department of Shandong Province and also for their constructive proposals.
References
1. Chinese Computer Terminology System, http://ccts.cs.cuhk.edu.hk/
2. The basketball dictionary of the hoopchina, http://www.hoopchina.com/ext/index.php/
3. Horton, W., Taylor, L., Ignacio, A., Hoft, N.L.: The Web page Design Cookbook. John
Wiley, New York (1996)
146 X. Cui, H. Wang, and Z. Cao
Abstract. This article introduces the concept and style of E-Learning, analyzes
the Blog’s characteristic while being applied in teaching, then designs an
E-Learning platform model based on Blog. At last, it also discusses the model’s
function, characteristic and existing practice problems.
E-Learning is a new type of learning based on the Web. Blog is regarded as a new web
culture, and the application in education and instruction is a hot point in recent research.
We have done some research and practice on how to merge Blog into E-Learning.
Blog (Weblog) is the serial accounts or diaries on the Web. It is a personal website with
its content given by the owner, listed from the latest. It has a social function. It appears
in the form of text and super-media. The users simply publish their individual knowl-
edge, ideas, thoughts and news as posts to share on the web. Blogs integrate articles,
accounts and communication together to make them into a web-page, with the infor-
mation much larger than traditional diaries. The blogs’information is filed by date,
classification and tag.
Blogs can be used as individual or course resource, and as exhibitive center and
communication center in the process of education and instruction with the features as
follows:
(1) In the process of instruction, blogs can be organized in the form of “basic
blog”and “group blog”. Basic blog is the simplest form with teaches or students sup-
plying certain topic and relative resources for discussion or comment. Group blog is the
transformation of basic blogs. It consists of groups of students to complete the blog
diaries together. The authors can edit not only their own diaries but also others’in the
same group, which permit them to discuss on the same topic and even complete a
collaborate project.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 147–152, 2008.
© Springer-Verlag Berlin Heidelberg 2008
148 Y. Chen and L. Wu
(2) Blog is a constructive learning tool suitable for the students’cognitive psy-
chology. It can organically integrate “situation”, “collaboration”, “dialogue”and
“meaning construction”. The value of blog in instruction can be seen in the following
aspects:
• expanding and extending the space for teaches and students to communicate;
• fostering the interest of self-seeking, and promoting collaborative level; re-
cording the whole learning process, and systematically realizing knowledge
management;
• accounting and evaluating together, and recording the feeling, discussion, re-
flection in the learning process to make a all-round evaluation.
(3) Blog, as a new form of communication on the web, will help the students to
enhance their abilities in receiving, analyzing, and processing information when it is
used in instruction.
Blog is now being dynamically growing up. We can see easily that blog changes
knowledge, news, and learning. It has great potentiality in the process of changing the
world.
E-Learning, in its broad sense, refers to the learning through electronic media with three
general forms: Learning through satellites’TV system, Learning through audio-video
meeting system, and Learning through computer web system. The narrow sense of
E-Learning refers to the learning on-line or web based learning. It is simple the learning
through computer network with the web learning environment consisting of multimedia
web learning resources, on-line learning community, and web technical platform.
E-Learning has synchronous and asynchronous forms.
(1) Synchronous E-Learning refers to the learning in virtual classrooms. The in-
structional information is shown as text or projection on the presentation device.
Sometimes the recording of the teaching can be presented directly on the line. Powerful
device and technical support is needed in this form of learning, to guarantee the fluent
communication between teachers-students and students-students. The advantage is that
the teachers can provide simultaneous feedback to students and the simultaneous
communication can make a real classroom environment, which gives the students a
feeling of family-ship. The disadvantage is that a formal timetable is needed, which
makes it not flexible enough.
(2) In asynchronous E-Learning, learners can make their own pace of web course
learning. They can check their learning record at any time and learn by their own
learning and thinking habit. They can also get the personal feedback from the instructor.
Students can communicate with their classmates and teachers on-line, exchange
e-mails, have class or group discussion through the web. This form of learning is
flexible and easily self-controlled.
Idea and Practice for Paperless Education 149
(1) The full advantage of the space on the web is used for effective E-Learning. Firstly,
the teachers and students can choose and register their free blogs at many web sites
150 Y. Chen and L. Wu
providing blog spaces. Secondly, Each part on the platform is equal at the inter- or
multi-communication.
The teacher is the center of the model. He/She adds the URL of the students to
his/her blog in the linkage of the classified catalogue to examine, supervise and
evaluate the learning process of the students. The students, at the same time, will an-
swer the questions or hand in their assignment in accordance to the teacher’s require-
ment in the blog. The students can choose to read the blogs and give comments as a
passer-by. They can record their acquisition and experience in their blogs. The blogs
digitalize the files and notes of learning and record the behavior of the students, such as
their feedback to the instruction, their report to social practice, and the collection of
their learning resources for investigative learning. The platform, on the other hand,
aims mainly at the process not the results of the learning, where the views of the
teachers or the books are presented. The students can publish their understanding to
them here, and the teachers can comment or provide a guidance to students’questions
and make the process into a dynamic interaction between teachers and students.
(2) The learning resource is digitally managed. Learning resource is the sustainable
guarantee for the instructional activities on the platform. The management of the
learning resources is an important task in the model of the platform. The model sup-
ports the upload and download of various multi-media files, which can add and update
the learning resource constantly. The content in the blogs includes reference materials,
courseware, teacher plan, experiment instructions, works of students, discussion on
important and difficult points, and exercises. The research of the teachers to the course
and production, the papers, and the honor the teachers have received can also be in-
cluded. Figure 3 shows the functional framework of resource management.
Idea and Practice for Paperless Education 151
Fig. 3. The functional framework of resource management of the Blog. RIM : Resource Issue and
Management; RSB: Resource Search and Browse; SM: System Management; RM: Resource
Management.
(3) All-round evaluation can be implemented to the learning process. The evaluation
of the instructional effect can be implemented with three sections ---individual,
group( in-or inter group) and teachers. Process and qualitative evaluation are recog-
nized in the platform with importance attached to comprehensive evaluation to the
learners. The evaluation can be done simultaneously on the web by the reflection to
detect the redeeming feature and the deficiency, and to find new problems arousing
new elicitation. The teacher occupies the dominant position in the instructional process.
He/She organizes the instructional activities, handles the feedbacks and evaluate the
learning. The students communicate and feed back in the blog when the blog record
their learning developments. The record in the blog shows the behavior and contribu-
tion of the students, supplying reference for the evaluation.
(4) Self-determined, collaborative and individualized learning can be achieved. The
platform based on the web provides a effective means for the learners in different
places. They can learn synchronously and asynchronously in a virtual classroom with
the technology of the web, database and artificial intelligence.
(1) The model is public and incompact. It can act only as a complement tool but not
a typical criterion for education and instruction management.
(2) The ideas of liberalization and individualization may drop a hint for the students
to publish irresponsible comments or do improper operation.
(3) The resource management has only one-level classification, which may cause
indistinct navigation. Only full-text search can be done with no classified search, and
may be inconvenient for searching the resources. The resources are appended with texts,
and the labels of attribution are not criterion.
4 Conclusion
Blog is an instructional tool or model based on the web. It is individualized, interactive
and convenient for knowledge management. E-Learning is a fashion for learning
152 Y. Chen and L. Wu
on-line, which makes the learning automatic, interactive, collaborative and individu-
alized. The research and practice we have made aim at the amalgamation of the two to
compose a model and provide a platform for the learning and communication of
teachers and students.
References
1. Cuifeng, Z.: E-Learning:Product of Information Age. Vista of Global Education (5), 52–53
(2002)
2. Lianghai, W.: Blog’s Appliction in Learning Source Management. China Science and Tech-
nology Information (21), 248–249 (2007)
3. Ming, Z., Youfu, D.: Construct the Cooperation Learning Platform with Blog. Journal of
Educational Technology (4), 4–5 (2006)
4. Qing, W.: Construct the Remote Education Teaching Mode Integrating Blog with Webquest.
Journal of Nanjing Radio & Television University (1), 13–16 (2006)
SyTroN: Virtual Desk for Collaborative, Tele-operated
and Tele-learning System with Real Devices
1 Introduction
The use of virtual reality has been strongly facilitated during the last ten years: tech-
niques are mature, costs have been reduced and computers and devices are enough
powerful for real-time interactions with realistic environments. Industrialists have also
validated huge prototyping, simulating and training systems: we can cite the VTT
project, which aims to train on technical gestures while manipulating a virtual milling
machine[1]. We can also cite the GVT project[2], with individual and collaborative [3]
learning for industrial maintenance procedure on military equipment. The use of virtual
reality for large public is more recent, but already pertinent: interactive 3D
video-games using VR devices, such as the Nintendo Wii, is a good example of this
success. On another hand, many works have been already done in the field of dis-
tant/e-learning, illustrated for example by the notion of virtual classrooms/courses
proposed by some universities[4] with pertinent results[5]. Those distant learning are
now very useful in a world where people are traveling a lot, and where jobs opportu-
nities conducted them to move several times. This context makes (distant and mobile)
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 153–161, 2008.
© Springer-Verlag Berlin Heidelberg 2008
154 R. Chellali et al.
training as one of the most active domain in the field of Virtual Reality. Now, tech-
niques could go from expensive training environments in dedicated rooms to personal
computers or even laptops, while large-band networks allow easy access to distant and
heavy information.
Concerning Tele-operated systems, virtual reality techniques are very interesting
and used since the begins. Indeed it provide a good abstraction for real devices op-
erators have to operate with. VR allows simulations, but also interesting augmented
representations of real devices and this is very useful in executing tele-operated tasks.
Obviously, one can take advantage of these rich environments for pedagogical pur-
poses. We can cite the work of Bicchi[6], who applied e-learning techniques in the field
of automatic control, in order to learn the usage of a robotic arm for example[7].
Dongsik[8] was focused on simulating electronic circuits in distant virtual laboratory,
with the ability to apply the models on real equipment with webcams, in order to
validate the theoretical simulations.
SyTroN project is born in this context. Fundamental research and innovative solu-
tions have been developed in order to success in mixing VR techniques, mobile and
distant learning, and tele-operation applications. With those techniques, the funda-
mental goal of our project is to increase the learning efficiency in a two steps process:
first, students learn and simulate, and then they manipulate real devices. To do so,
SyTroN provide virtual classrooms supervised by a teacher, where distant students,
with standard supports such as distant reference books or media contents, have to train
by acquiring theoretical skills and learn more about processes and devices models.
Once an acceptable level reached, students can move to teleoperation of the real de-
vices to face and solve real problems. SyTroN system is a complete and functional
solution allowing for the moment simulation and training on 3 different tele-operated
devices. The system also takes advantage of VR for pedagogical add-ons not available
in the real world, in particular with dynamical and contextual information in the 3D
environment to help the training process.
The first part of this paper deals with the presentation of our models and its imple-
mentation. The next section gives an overview of the system usage. Finally, the last
section presents the field validation of SyTroN, based on a comparison of training
sessions with traditional techniques versus training sessions using SyTroN.
2 Contribution: SyTroN
We present in this section the models and the implementation of SyTroN.
2.2 Functions
In the following points, we are now introducing the set of functions attached to each
logical entity, according to the needs of usage.
Users functions. Except for some specific functions, students and teachers are con-
sidered as end-users and both are clients. The corresponding functions enable to access
the SyTroN services, namely, the device servers and the knowledge database. Two
main sets of functions are available: Communications, and Remote controls.
The communications functions enables the users to access to the contents and to
people within the SyTroN system. Namely, the students can:
– Access to lectures, tests and evaluations ;
– Contact teachers and other students ;
– Get the devices status and gateways ;
– Simulate devices;
Some specific functions are reserved to teachers : adding devices, adding lectures,
and setting tests and evaluations.
The remote control functions link users to physical devices. Users are enabled to
achieve remote physical interactions: they may send controls and receive measure-
ments in return, i.e. the device status as well as some information about the remote
environment.
The previous functions are implemented for the most used OSs: Windows, Unix and
Linux. Users terminals are designed for WEB-based interface as well as proprietary
interface. The communication blocks of each function are TCP-IP based.
Device functions are separated between a server and a controller.
Device server is the gateway between users and the remote world. It is composed of
two parts: the device gateway and the device controller. The device gateway connects
the users and the knowledge database to the device. The device controller pilots the
actuators and gets the sensors information. The device gateway is just a translator.
Depending on the considered device, the gateway formats the users commands and
passes the parameters to the controller. The concerned actuators execute the command
and respond to the device controller by giving the new position of the mobile. The
functions of the gateway are mainly dedicated to synchronous communication. TCP
protocol is used to guaranty the exchanges synchronicity and integrity.
Device controller is the interface between the real and the logical worlds. Its main
function is to close the local control loop. This component is device dependent: for each
device, one needs to write a specific driver. For SyTroN implementation, three devices
with three different dynamics have been tested. Due to the nature of the considered
devices, two control platform are used : the Matlab (from MathWorks) platform and a
proprietary platform (C++ based API).
Knowledge database functions. The knowledge database contains all the information
needed to manage the global system. Three sub-databases describes the three logical
units (the teacher, the student and the device) involved in a distant learning session, the
cross-relationships between the units, and a history of platform uses.
SyTroN: Virtual Desk for Collaborative, Tele-operated and Tele-learning System 157
3.1 Devices
The figure 2 presents two of the three devices we have already included in our system.
The heating board is dedicated to the study of heating processes, simulating com-
plex systems with numerous inputs and outputs in the state space (an algebraic ap-
proach of automatic control). The heating board is a basic MIMO system (Multiple
Inputs Multiple Outputs). The students have here to keep constant the temperature of
the board, regardless to the external temperature. The goal of the training session is to
learn MIMO control techniques, especially control laws.
The mobile Pekee robot is a classical robot produced by Wany Robotics SA. This
robot has two actuators, namely two DC motors enabling the robot to move with 3
degrees of freedom. On the other hand, Pekee has onboard a set of sensors (telemetry,
light, gyros, collision detector, etc.) enabling to handle the robot environment. Both
actuators and sensors are handled by the onboard PC. This last communicate with any
other PC using the wireless Ethernet based channel. The communication can concern
the state of the robot (sensors information) or robot controls (DC motors or a request for
a specific measurement).
3.2 Usage
Starting the session, the student download the documents explaining the lecture, the
manipulation protocol and the attended results. Following this, the simulation phase
can start on the virtual device (equivalent of the real one). This enables to acquire the
theoretical basis. In this virtual classroom(figure 3), students can easily manipulate
virtual devices, helped by VR pedagogical metaphors such as avatars, the use of
transparency, etc.
The heating board. The student can verify the results of the simulations by comparing
resulting curves to theoretical ones. From the derived model, the student can switch to
the real device to setup the control parameters, namely the PID constants. These pa-
rameters are sent via the asynchronous TCP channel to the device server and the real
test can start. The control closed loop is then operating to maintain the temperature to a
desired one. The client calculate the inputs of the heating board and send it via the TCP
synchronous channel. The device extract the current temperature and return it to the
client and so on. This question-response process is done at 1Hz frequency. Indeed, as
the response time of the system is greater than 15secondes, 1Hz frequency for the
control loop is enough to verify the Shanon theorem and to ensure the stability of the
system. To add more impact on the learner, a video feedback was added to this device.
The mobile Pekee robot. For the mobile robot, time constraints are stronger than for the
previous device. Indeed, the purpose here is to enable to a distant student to pilot a
mobile with a visual and a force feedback. It is obvious that the shorter the system
response time is the better is the manipulation. A UDP based protocol is used for this
service. As for the heating board, the first stages of the mobile robot manipulations are
concerned with the discovery of the device.
This is done using documents and simulations. For the tele-operated mode, the user
may use a force feedback joystick, a mouse or the computer keyboard to pilot the robot.
This generate motion commands that are sent to the robot. The robot execute these
commands, capture about 31 sensors measurements and reply to the client. These
measurements are then used to refresh the client virtual environment. As the robot
speed is about 0.5m/s, the user may react in less than 1/2 seconds to avoid obstacles in
front of the robot. This response time is obtained by tuning the range sensor dynamics
to 1m depth. Following that and considering transmissions time delays in the network,
any object about 0.25m has to be considered as very close obstacle. A hidden mobile
robot was added to the interface to take into account the effects of time delays and the
resulting offset between the user command and the real state of the remote robot.
The classroom metaphor : sharing the working environment. One of the main issues
of Sytron is to offer a collaborative platform. Collaborative services use the network to
share users activities to facilitate the realization of tasks and communication, by dis-
playing graphical feedback to show connected people and what they are doing. For
Sytron, the collaborative framework and the share of the 3D virtual environment are
built around a communication server named reflector. The Reflector supports a set of
Peer to Peer or broadcast-like services (chat or voice conferencing for instance). UDP
and TCP based protocols are used to handle these services with no specific constraints
except the users comfort.
The virtual white board is designed to let user share results and formula. With this
tool, users can draw whatever they want thanks to the pen tablet (or mouse). The chat
board display messages written by users who can type a message at any time. Users can
also see the avatars corresponding to the other users connected. When a user has a
service as focus, his correspondent avatar flies next to the object corresponding to this
service, in order to indicate to the other participants what he’s doing.
4 Evaluation
Before developing and writing codes for Sytron, we started by designing a paper proto-
type based on interviews we done with all actors. Teachers were asked to detail their
presence-based lectures regarding structural aspects, timing, exchanges (teacher-students,
students-students) during lectures, the most impacting parts, etc. This study leads us to
write a script with handmade screens and concepts to include into the virtual desk and the
virtual classroom. Our approach was not only objective: some of our considerations are
purely subjective to take into account personalities of teachers and how they transmit
knowledge. Students also evaluate the quality of knowledge transmission mostly
perceptually. They are more sensitive to the way the teacher introduces and explains
concepts than the validity of concepts. Following that, we opened the system to two
categories of students: normal program for young engineer (12 people) and adult program
for which students work mainly from home and nightly (10 people). Individual usages
were checked and we verified our pre-conception hypothesis: the interface is fluid and
users discovered services in a quasi natural way. The second verification was concerned
with the shared space and attached tools. Adults and young engineer reactions were
different: for the first category, chat and live talk were more used before and during ex-
periments. For young people exchanges were mainly concerned with simulations/real
160 R. Chellali et al.
experiments results. May be cultural gap exists between the two categories and ones are
more used with known channels (MSN or Skype for instance) and our offer was dis-
turbing their communication environment.
The last evaluation to do is concerned with the classroom metaphor. All participants
may use the system at the same time with the goal to replace the existing lectures. This
step is not of evidence and we have to reconsider the previous evaluation approach:
marks, credits and public evaluations are to be produced.
5 Conclusion
We detailed the Sytron system, its architecture, its components, its use and some pre-
liminary evaluations. First, we strongly concentrate our work on automatic control
aspects (closing high/low speed loops over Internet) which are not described here. The
second step was to pre-design the virtual desk and its components. Voice, video and
written messages channels were integrated within the system. A central system man-
aging technical and administrative matters finishes our core developments. We have an
operational system which can constitute a good basis for further improvements like
adding an ITS intelligent Tutoring System for automatic assistance for the teacher.
Indeed, teachers currently act in an old fashion by following learners. An ITS can help
to personalize more the relationship between each partner. The other contribution we
made and which must be also improved in the future is the use of real devices to let
learners confront to real problems. In addition VR techniques enable us to augment the
reality by adding guides and milestones. This let users use the maximum available
information at experimenting time.
Special acknowledgments to Amelie Imafouo. SyTroN can also be tested at http://
www.emn.fr/x-auto/reposit .
References
[1] Crison, F., Lecuyer, A., Mellet dHuart, D., Michel, G., Burkhardt, J.M., Dautin, J.L.: Virtual
technical trainer: Training to use milling machine with multi-sensory feedback in virtual
reality. In: IEEE VR 2005 (2005)
[2] Gerbaud, S., Mollet, N., Ganier, F., Arnaldi, B., Tisseau, J.: Gvt: a platform to create virtual
environments for procedural training. In: IEEE VR 2008 (2008)
[3] Gerbaud, S., Mollet, N., Arnaldi, B.: Virtual Environments for Training: From Individual
Learning to Collaboration with Humanoids. In: Hui, K.-c., Pan, Z., Chung, R.C.-k., Wang,
C.C.L., Jin, X., Göbel, S., Li, E.C.-L. (eds.) EDUTAINMENT 2007. LNCS, vol. 4469, pp.
116–127. Springer, Heidelberg (2007)
[4] Lepori, B., Cantoni, L., Succi, C.: The introduction of e-learning in european universities:
models and strategies. In: Kerres, M., Voss (eds.) Digitaler Campus, pp. 74–83 (2003)
[5] Kukk, V.: Analysis of experience: fully web-based introductory course in electrical engi-
neering. In: Proc. Int. Workshop e-Learning and Virtual and Remote Laboratories, pp.
111–118 (2004)
SyTroN: Virtual Desk for Collaborative, Tele-operated and Tele-learning System 161
[6] Balestrino, A., Bicchi, A., Caiti, A., Calabr, V., Cecchini, T., Coppelli, A., Pallottino, L.,
Tonietti, G.: From tele-laboratory to e-learning in automation curricula at the university of
pisa. In: IFAC 2005 (2005)
[7] Bicchi, A., Caiti, A., Pallottino, L., Tonietti, G.: On-line robotic experiments for
tele-education at the university of pisa. International Journal of Robotic Systems (2004)
[8] Kim, D., Choi, K., Lee, S., Jeon, C., Yoo, J.: Implementation of a web-based hybrid edu-
cational system for enhancing learning efficiency of engi-neering experiments. In: Tech-
nologies for E-Learning and Digital Entertainment -Edutainment 2007. LNCS, pp. 357–368
(2007)
An Examination of Students’ Perception of Blended
E-Learning in Chinese Higher Education
Jianhua Zhao
1 Introduction
Blended learning is hardly to say that it is a brand-new term, rather “an old friend gets a
new name” [1], which can be considered as a fairly new term in education lingo, but the
concept has been around for decades. Blended e-learning offers a new learning ap-
proach for combining different delivery modes, normally is online and face-to-face
teaching [2][3]. It changes the traditional face-to-face learning through bringing a
non-geographical and non-time limitation online learning when it is applied in
on-campus teaching/learning, because teachers and students, students and student can
meet online when they are off-campus. Blended learning is the fastest growing trend in
e-learning [4], and the successful e-learning program is or will become a blended
learning program [5]. Therefore, blended learning could be the next most popular term
which can replace e-learning. This is the reason for using blended e-learning rather than
blended learning in this study.
Numerous researchers have presented their efforts on exploring the meaning of
blended learning. For example, Smith defines blended learning as a method of educa-
tion at a distance that uses technology (high-tech, such as television and the Internet or
low-tech, such as voice mail or conference calls) combined with traditional (or,
stand-up) education or training [1]. Bielawski and Metcalf point out that blended
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 162–170, 2008.
© Springer-Verlag Berlin Heidelberg 2008
An Examination of Students’ Perception of Blended E-Learning 163
2 Method
Questionnaire survey research was used as method in this study. In order to compare
students’ perceptions of blended e-learning before and after the course, the pre- and
post-surveys were designed for collecting the relevant data. Except the closed questions,
the opened questions are involved in the questionnaire too in order to in-depth examine
students’ thinking and understanding.
20 undergraduate students at the School of Information Technology in Education
(SITE), South China Normal University (SCNU) registered in a bilingual course –
Computer in Education for eight-week studying. 20 students participated in the first
questionnaire survey at the beginning of the course. The responded rate and the valid
rate are all 100%. 16 students participated in the second questionnaire survey at the end
164 J. Zhao
of the course, and 13 of them gave their response back. The responded rate and the valid
rate are 81.25%.
The learning environment in this study includes two settings, which one is
ICT-based classroom with internet connection, and another is a VLE – WebCL which
is developed by Beijing Normal University. This is a good environment where it offers
lots of functions, such as group and class discussion forums, course management
module, group management module, and personal blogs etc.
3 Data Analysis
3.1 Students’ Experiences of Blended E-Learning
With respect to the first survey of this study, the question “do you have any experiences
of blended e-learning” was proposed to find out whether they had experiences or not
(see Table 1).
The data demonstrate only 20% students chose “used blended learning before”, and
80% students chose “never used it”. The data also illustrate 25% students heard the
term “blended learning” before, and 75% students “never heard this term”. The data
reveal that most continuing students were lack of the experiences of blended learning at
the beginning of this course.
The further analysis of students’ choice in questionnaire survey, when students were
asked to give “where you heard this term” if they chose “yes, I have used it before”,
they all replied “from this course” (100%). It is an interesting answer because it reveals
that students have no experience of blended e-learning before this course.
The qualitative data collected from the second questionnaire survey were analyzed
in order to examine how their experiences of blended e-learning developed through the
course.
Miriam: Blended e-learning combines the conventional educational approach
(classroom-based education) with the new educational approach (i.e.
web-based education) together. These two approaches can be complementary
each other and the new approach represents their advantages from that two
approaches. It can improve the quality and effectiveness of education.
Miss Mariam’s thinking was similar to the introduction from the relevant literature
[12][13], which demonstrates that she already formed appropriate understanding of
blended e-learning. Her attitude to blended learning was positive.
An Examination of Students’ Perception of Blended E-Learning 165
Onjcn: Blended learning combines online learning and FTF learning. Online
learning can improve the learning efficiency and to facilitate students learning.
In-depth understanding can be constructed through the communication among
students though the web. Meanwhile, students can consider the understanding
from the classroom as the foundation for further learning. Online learning can
be used to in-depth explore the relevant knowledge. FTF in the classroom and
online discussion are all very useful for improving their understanding.
Mr. Onjcn introduced his understanding of blended e-learning from online learning
and FTF learning. He explained how these two learning approaches were useful for
their understanding. He considered communication as the essential element. It could be
acquainted from classroom-based learning first and then was used as the foundation in
the online learning environment. The discussion was a useful means for building stu-
dents’ understanding.
Tina: In blended learning, tutor can give her/is guide, lecturing, and explana-
tion to students. It can improve the efficiency of the class. Web-based learning
resources can supply the shortage of the resources in the classroom setting.
Therefore, students have more chances to learn the relevant information.
Miss Tina explained the reason why students had more chances to acquaint more
information, because the online learning resources could remedy the shortage of the
resources in the classroom-based setting. Normally, it was difficult to get sufficient
learning resources in the classroom-based setting. However, when students were in the
online-based setting, they could be supplied plentiful learning resources. Blended
e-learning provided chance to compliment the shortage of classroom-based learning.
Gigi: In a blended learning environment, classroom-based and online-based
instruction can complement each other. Online course can assist class-
room-based instruction. There would be more time to collaborative learning
between students and tutor, students and students.
Miss Gigi described the relationship between tutor and students, students and
students in blended e-learning. They would have much time for collaborative learn-
ing, because online courses could assist their learning. Online learning could help
students to know more background of the relevant knowledge in their courses.
Meanwhile, they could also learn from online course themselves. It was useful to
improve the efficiency and effectiveness of collaborative learning when students and
tutor were getting together.
The data in table 2 reveal that “group learning” was the most important element for
the effective classroom-based learning, because fourteen students (73.7%) introduced it.
The second important element was “engagement learning” and ten students (52.6%)
mentioned it. The third important element included “learning environment” and
“teaching techniques”, and each one got 9 students’ (47.4%) proposition. 8 students
(42.1%) considered “heuristics teaching” method as the effective method that should be
used in classroom-based teaching. 6 students (31.6%) suggested that “interaction”
could facilitate teaching and learning.
Students’ experiences of online learning were examined at the beginning of the course
first. Among these 20 respondents, 38.3% of them had no any experience of online
learning, but 61.7% students had.
Eight students mentioned their experiences of online learning were formed in terms
of the different courses, i.e. the Theory of Television, Analog Electronic Technology,
Programming, Photoshop, Flash, Director, ASP, and Television Facilities. Some of
their descriptions are quoted and analyzed as follows:
An Examination of Students’ Perception of Blended E-Learning 167
other. She concerned students’ autonomy and the attraction of the issues on the web.
Her perspectives on online learning already went far beyond the superficial issues that
were formed at the beginning of the course, such as tutor’s guide, FTF communication,
and learning resources.
(2) The Effective Tools of the Online Learning Environments
In the second survey of the study, students introduced their thinking about the effective
tools of the online learning environment. The data about the effective tools of the online
learning environments is presented in Table 3.
“Discussion forum” got the significant response, which 84.6% students considered it
as an effective tool for the online learning environment. However, at the beginning of
the course, only 15% students chose this item, which illustrated that students liked to
use discussion forum to express or share their understandings. Students also introduced
their considerations to them.
Kinston: Discussion forum is an asynchronous communication tool. It can be
used effectively for those which need long time to ponder.
Mr. Kinston stated that discussion forum was an asynchronous communication tool.
It could facilitate students’ pondering if they took time for it.
Onjcn: Discussion forum can reflect students’ reality in time and more detailed.
The specialised discussion forum (sub-forum) results in the issues of discussion
more concentrated and professional.
Mr. Onjcn thought discussion forum could reflect students’ reality in time and more
detailed. He also mentioned the issues in discussion forum were more concentrated and
professional.
The second significant item was “email” (53.9%). At the beginning of the course,
only 25% students chose this item. This also demonstrated that students considered
email was an effective tool when they were engaging in the online learning activities.
34.5% students chose item “chatting” at the end of the course. However, 40% students
chose it at the beginning of the course. Some students changed their attitudes to the
“chatting” when they acquired the new experiences of online learning. It revealed that
“chatting” had some limitations, such as synchronous and the chatting contents could
not be easily saved. 15.6% students thought that “QQ (Quick Quest)” was the effective
An Examination of Students’ Perception of Blended E-Learning 169
tool for online learning. Only 7.8% students considered “learning resources” and
“blogs” were the effective tools.
4 Conclusions
This is a small-scale research on how Chinese students’ perception of blended
e-learning. It is only a short-term since blended e-learning has been using in the field of
Chinese higher education, which is important to know how Chinese students’ thinking
and understanding of blended e-learning, especially for knowing how they experience
this brand-new approach. The results presented in this study demonstrate that the stu-
dents have diverse understanding and experiences of blended e-learning and their at-
titudes are positive. The practical experiences would be crucial for forming and im-
proving students’ understanding of blended e-learning.
Acknowledgements
I thank for the students who enrolled in this course. Without their cooperation, this
study will not be finished. I also thank my PhD supervisor, Prof. McConnell for his
direction and encouragement.
References
1. Smith, T.: Asynchronous discussions: Importance, design, facilitation, and evaluation
(2001), http://www.ion.illinois.edu/pointers/2002_11/page1.html
2. Mason, R.: Guest editorial: blended learning. Education, Communication & Informa-
tion 5(3), 217–220 (2005)
3. Stubbs, M., Martin, I., Endlar, L.: The structuration of blended learning: putting holistic
design principles into practice. British Journal of Educational Technology 37(2), 163–175
(2006)
4. Ward, J., LaBranche, G.A.: Blended learning: The convergence of e-learning and meetings.
Franchising World 35(4), 22 (2003)
5. Bersin & Associates. Blended Learning: What Works? E-LearningGuru (2003),
http://www.e-learningguru.com/wpapers/blended_bersin.doc
6. Bielawski, L., Metcalf, D.: Blended eLearning: Integrating Knowledge, Performance,
Support, and Online Learning. HRD Press, Inc., Amherst (2003)
7. He, K.: Examining the new development of educational technology based on blended
learning (part I). Journal of Audio-Visual Education Research 131(3), 1–6 (2004)
8. Li, K., Zhao, J.: Blended learning: theory and application approach. Journal of Audio-Visual
Education Research 135(7), 1–6 (2004)
9. Ginns, P., Ellis, R.: Quality in blended learning: Exploring the relationship between online
and face-to-face teaching and learning. Internet and Higher Education 10(1), 53–64 (2007)
10. Delialioglu, O., Yildirim, Z.: Students’ perceptions on effective dimensions of inter-active
learning in a blended learning environment. Educational Technology & Society 10(2),
133–146 (2007)
170 J. Zhao
11. Reichlmayr, T.: Enhancing the student project team experience with blended learning tech-
niques. In: The Proceedings of 35th ASEE/IEEE Frontier in Education Conference, T4F-6,
Indianapolis, IN, October 19-22 (2005)
12. Tian, R.: Exploring on organising method of English learning group. Journal of Sheyang
College of Education 5(1), 53–56 (2003)
13. Gao, R., Kong, W.: Problem-based learning in the networked learning environment. Journal
of Chinese Educational Technology 211(8), 28–33 (2004)
14. He, K.: Constructivism: instructional mode, method, and design. Journal of Beijing Normal
University (Social Science) 143(5), 74–82 (1997)
Research and Application of Learning Activity
Management System in College and University E-Learning
1 LAMS Introduction
Learning activity management system (LAMS) is a system for creating and managing
sequences of Learning Activities. It was developed by Macquarie university, Sydney
Australia. LAMS has four main areas: authoring, monitoring, learner and admini-
stration [1].
LAMS can support many activities like chat, forum, notice board, notebook, Q &
A, share resources, submit files, survey, voting etc. Some activity can be finished by
learners themselves and some need to be finished by the learners’ collaboration.
Teacher can use drag-and-pull method to assemble activities into sequence easily. It is
a good tool for designing, managing and delivering online collaborative learning
activities. It provides teachers with a highly intuitive visual authoring environment for
creating sequences of learning activities [2]. LAMS can be used as a stand alone
system or in combination with other learning management system such as Moodle,
WebCT or BlackBoard.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 171–179, 2008.
© Springer-Verlag Berlin Heidelberg 2008
172 L. Yan et al.
Fig. 3. Learning sequence in “New Horizon College English, Book II” course, unit 5
4 Data Analysis
4.1 Data Analysis in “Modern Educational Technology” Course
4.1.1 Motivation
Question: Why did you sign in the activity about LAMS experience? (multiple choice)
Result: see Table 2.
Table 2. Result about motivation
From the result, we can see the students’ learning attitude and motivation are active
and correct. The new learning method can inspire students’ interesting.
From the result, we can see, the network infrastructure is the biggest problem when
learners use LAMS. It could be solved by the support of some departments of
university like computer center. For the factor 2, it needs the platform developers to
enhance the evaluation function of LAMS. For the factor 3, 5, 6, they are common
problems towards network learning. We need to enhance students’ ability and
literature about online learning. And we can see, for most college students, using
English version LAMS is not the problem.
4.1.4 Effect
The data about application effect with LAMS sees Table 4.
176 L. Yan et al.
Statistic
Condition set:
value
(1) Compared with 2 stands for “much more active”; 1
learning in classroom, your stands for “active”; 0 stands for “almost
0.62
thinking is: same”; -1 stands for “passive”; -2 stands
for “much more passive”
(2) Is this online learning
with LAMS helpful for you 2 stands for “very helpful”; 1 stands
1.29
to understand Educational for “helpful”; 0 stands for “not helpful”
Technology?
The result illustrates that for most students, learning with LAMS can make their
thinking more active and it is helpful for them mastering the knowledge and
understanding the concepts better in their professional field. In one word, with LAMS
this new e-learning platform, students can get better learning effect than traditional
learning method.
4.1.5 Attitude
Towards learner’s attitude, we set two questions let students voting. The first question
is: Which learning method do you prefer? The second is: Will you volunteer to use
LAMS later? The Fig.4 and Fig.5 can give us intuitionistic expression about the
result.
30 25
25 in Classroom
25 20 Certainly I
20
will
20 Probably I
with LAMS 15
15 will
10 Probably I
10 Network- 8 won’t
4 based,but not Certainly I
5 3 LAMS 5 3 3 won’t
2
Self-study
0 0
1 1
learning method your choice
Fig. 4. Learning methods the students prefer Fig. 5. The students’ choice towards LAMS
(“Educational Technology” course) (“Educational Technology” course)
The results sound conflict. Although most learners think using LAMS for e-
learning is better than traditional learning, and they probably use this method again,
they would like to go to classroom and learn knowledge from teacher face to face. We
analyze the reasons come from these factors: one, the learning sequence is too long.
Students should spend much time on it. This would cause that students feel tired.
Two, the students are not very familiar with LAMS platform, and without the
teacher’s direct guide and help, students will feel helpless sometime. Three, every
Research and Application of Learning Activity Management System 177
new thing needs time to be adapted by people. For this new learning method, students
need some time to adapt it and like it.
Through “survey” and “voting” tools, we collected students’ feedback and got the
statistic data about LAMS application result in this course. See Table 5, Fig.6. and
Fig.7.
total
Nomination
votes
Less than 3 learning hours 7
(1) Time spend 3 to 5 learning hours 12
More than 5 hours 3
With LAMS, the effect is better than traditional
(2) With LAMS, 12
learning.
how do you
With LAMS, the effect is less than traditional learning. 2
think about the
No obvious difference. 6
learning effect?
Open response 2
12 12 11
10
10 In classroom 10
8 7 With LAMS 8 Yes. I will.
6 No. I won’t.
6 6
No obvious Both are OK.
difference. 4
4 3 4 Open response
2 Open response
2 2 1
0 0
1 1
learning method further choice towards LAMS
Fig. 6. Learning methods the students prefer Fig. 7. The students’ choice towards LAMS
(“Advanced multimedia technology” course) (“Advanced multimedia technology” course)
From the feedback of students, we can see: using LAMS as e-learning platform,
students should spend much time than using traditional learning method. But, students
could receive more resources and obtain more knowledge towards same knowledge
concept. We don’t deny that with LAMS platform, there are some problems like the
interaction between teacher and students is affected by the limitation of time; when
students met problems, they could not get rapid response from teacher sometimes;
there are no audio and video instant interaction with LAMS; it is not convenient for
teacher demonstrates some technical operations with LAMS to students. However,
most students think that LAMS is a good e-learning platform and they can get better
learning effect with LAMS. The best advantages of LAMS from students’ opinion are
178 L. Yan et al.
represented with flexible learning time, self-learning method, colorful and rich
resources, good interaction etc.
4.3 Data Analysis in “New Horizon College English, Book II” Course
In this course, teacher got students’ feedback through communion with them.
Advantages in the eyes of students include: (1) New teaching and studying mode
arouses students’ interests; (2) The application of LAMS is close to the life of modern
students; (3) New teaching concept enriches the varieties of classroom activities;
(4)New classroom activities can explore students’ potentials, get students involved in
more activities leading to the use of the language; (5) LAMS can strengthen the
interaction between the teacher and students, thus a better learning environment is
established.
Disadvantages of e-learning with LAMS include: (1) The system is web-based,
however, the speed of the internet and other hardware can not meet the requirement;
(2) The system is not stable and students can’t make use of the system at the same
time; (3)The account numbers of students are hard to manage; (4) The teacher is not
familiar with all the skills in designing the sequences; (5)This course focuses more on
spoken English, while the system only can offer the communication based on written
English.
Acknowledgement
This research is supported by the Program of Introducing Talents of Discipline to
,
Universities Ministry of Education and State Administration of Foreign Experts
Research and Application of Learning Activity Management System 179
( )
Affairs of China NOB07042 , and National Great Project of Scientific and
Technical Supporting Programs ( NO. 2006BAH02A24 ).
References
1. Dalziel, J.: Implementing Learning Design: The Learning Activity Management System
(LAMS). LAMSTM Teacher’s Guide, November 2006, V2.0. Copyright © 2004-06 James
Dalziel, pp. 75–80 (2006)
2. About LAMS, http://www.lamsinternational.com
3. Dalziel, J.R.: From re-usable e-learning content to re-usable learning designs: Lessons
from LAMS [EB/OL] (2004)
4. http://www.lamsinternational.com/CD/html/resources/whitepapers/Dalziel.LAMS.doc
5. Britain, S.: A Review of Learning Design: Concept, Specifications and Tools[EB/OL]
(2004), http://www.jisc.ac.uk/uploaded_documents/ACF1ABB.doc
6. Russell, T., Varga-Atkins, T., Roberts, D.: Learning Activity Management System
(LAMS) Specialist Schools Trust Pilot Review. CRIPSAT, Centre for Lifelong Learning,
University of Liverpool. BECTA ICT Research (2005)
7. Marshall, S.: Leading and managing the development of e-learning environments: An
issue of comfort or discomfort? In: Atkinson, R., McBeath, C., Jonas-Dwyer, C., Phillips,
R. (eds.) Beyond the comfort zone: ASCILITE 2004. Proceedings of the 21st Annual
Conference of the Australasian Society for Computers in Learning in Tertiary Education,
Perth (2004)
8. Kenny: A research-based model for managing strategic educational change and innovation
projects. In: Annual Conference Proceedings of HERDSA (the Higher Education Research
and Development Society of Australasia), University of Canterbury, Christchurch, NZ, 6-9
July (2003)
9. Huang, R., Zhou, Y.: The Characteristic Analysis of distance learning. China Audiovisual
Education, pp. 75–79 (March 2003), pp. 69–71 (April 2003)
10. Wang, Y., Peng, H., Huang, R.: Scale of Distance Learners’ Learning Motivation
Development and Initial Application. Open Education Research, pp. 74–78 (May 2006)
Motivate the Learners to Practice English through
Playing with Chatbot CSIEC
1 Introduction
1.1 Motivation
English, as an international language, is treated as a key tool for the development and
cultivation of the cross-cultural communication ability. English language is now listed
as one of the three core courses in China’s elementary and secondary education, and
as a compulsory course in higher education. Statistical data shows that there were
more than 176 million people learning English in China in 2005 [1].
However, some problems exist in the English education in China. First of all, one
of the best ways to learn a foreign language is through spoken dialogue with native
speakers. But it isn’t practical in the classroom due to the one-to-one student/teacher
ratio it implies, especially in China and other countries with English as a foreign
language. A number of factors ranging from the lack of time to shyness or limited
opportunity for quality feedback hamper using the target language [2]. The language
environment and few qualified English teachers in China can’t supply enough chance
of authentic talking. So school teachers often complain of working burdens, and don’t
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 180–191, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Motivate the Learners to Practice English through Playing with Chatbot CSIEC 181
have enough time to converse with students in English. Secondly, although learning
English through communication and application has been emphasized recently,
passing examinations is the main motivation of many students to learn English.
Thirdly the grammar instruction is crucial to China’s English education, because
Chinese differs greatly from English in grammar [3]. Without basic grammar
knowledge, the students can’t make great progress, as they mostly only practice
English in school time, and can’t learn it spontaneously from the social environment.
A potential solution to these problems is to apply computer spoken dialogue
systems to role play a conversational partner. If we could design an interactive web-
based system which could chat with the English Learners anytime anywhere, their
great demand for learning partners could be fulfilled. Such a system should aim at
helping the learners improve their skills of using English through frequent chatting
with them in English, as well as encouraging them through playing and scoring
mechanism. Motivated by the great demand for English instruction, we in 2002 began
to design such a system. Our design principle is application and evaluation oriented.
So long as the system is applicable, we put it into free use in the Internet and get the
user feedback. We also cooperate with the English teachers and integrate the system
into English instruction. Through the systematic application and evaluation we get
more suggestions and critiques, which can direct our research more effectively.
Brennan defined a chatbot as "an artificial construct that is designed to converse with
human beings using natural language as input and output" [4]. ELIZA [5], the first
chatbot, used key words to analyse input sentence and created responses based on
reassembly rules associated with a decomposition of the input. The syntactic way of
NLP (Natural Language Processing) exemplified by ELIZA has been developed
significantly from 1960s up to now, leading to the development of various chatbots.
Since 1990s with the improvement of natural language processing, chatbots have
become more practical, and have also been applied in education.
Graesser et al. [6] used “AutoTutor”, an intelligent tutoring system with mixed-
initiative dialogue simulating a human tutor via conversation with the learner in
natural language, to enhance the learner's engagement and the learning depth.
Seneff [7] described several multilingual dialogue systems designed to address the
need for language learning and teaching. A student’s conversational interaction was
assisted by a software agent functioning as a tutor with translation assistance anytime.
Kerfoot et al. [8] described an experimental use of chatbots as a teaching adjuvant
in training medical students. Their web-based teaching using chatbots increased test
scores in four topics significantly and learning efficiency three-folds.
Abu Shawar and Atwell [9] developed algorithms for adapting a chatbot to chat in
the language and topic of the training corpus. The evaluation feedback from language
learners and teachers indicated that these adaptive chatbots offered a useful
autonomous alternative to traditional classroom-based conversation practice.
Kerly et al. [10] described an experiment to investigate the feasibility of using a
chatbot to support negotiation. Its result showed that most students liked the chatbot
as the chatbot helped them understand their learner model.
182 J. Jia and W. Chen
The related works above show the usage of chatbot systems in education is
drawing more attentions from related researchers. This trend confirms our
determination to further the development of the CSIEC system and its application in
English education.
J. Microsoft agent script formatter transforms the output text into VB scripts,
considering the selected agent character and speaking speed.
K. Browser/Server interface processes the http request from client machine and
responds with the robot output, either in text or with VB script.
Thirdly the chat topic between the user and the robot can be either free (unlimited)
or specific (limited). The unlimited dialogue simulation doesn’t specify the dialogue
topic and content. It suits the need of the users whose English is fluent or who are at
least good at written English, as well as users who are extroversive or conversational.
However, users whose English is poor, or who are introversive, have little to chat with
the virtual chatting partner. For those users, an instructive dialogue in a specific
scenario guided by the agent is more helpful.
In normal human talking these two chatting patterns are not absolutely separated,
but often interleave each other. Our system considers this interaction too. In the next 2
subsections we introduce the two patterns in details, as well as their relationship.
In the free chatting the users with different characters and personalities may choose
different types of chatting pattern. For example, some users may prefer to chat with
someone who listens to them quietly most of the time; while some others may hope
the chatting partner can tell stories, jokes or news. For the sake of user dialogue
personalization we designed five Microsoft agent characters which represent different
kinds of chatting patterns [15]. Christine always tells the user stories, jokes and world
news. Stephan prefers to listen quietly when the users share with him their own
experiences. Emina is a curious girl, and is fond of asking users all kinds of questions
related with the users’ input. Christopher provides comments, suggestions and advices
on the user’s input. Ingrid behaves as a comprehensive virtual chatting partner, who
gives users responses corresponding to both the input text and the discourse context.
Upon user registration to the chatting system the user's profile is obtained and
recorded, such as the gender, birthday, educational level, and province. So the
corresponding chatting topic and content can be generated based on the personal
information. Certainly if the user wishes to change the chatting topic during the
process of the robot’s narrating comments or asking questions, the robot should
terminate this process and transfer to another topic. If the user specifies a topic, for
example, “I want to talk about sport”, the robot changes the topic to it. If the user just
expresses the wish to change the topic, but doesn’t determine a topic, such as “I want
to talk about another topic”, the robot selects one from the waiting topics list which
has not been talked about with the given user.
The user's interests are also expressed in the input texts, e.g. the mentioned nouns
and verbs in the sentences. So the chatting topic can be triggered by nouns and verbs,
and their combination. More frequently one noun is or several related nouns are
talked about, the related topic is more emphasized. The chatting between the user and
robot can be regarded as guided chatting or chatting in some context.
Then we deal with the chatting on a given topic in two ways. The first one is
predefining some comments or asking some questions about this topic. By talking
about this topic only one statement or question will be randomly selected and given
out. The second way is to search the topic or related topic in the guided chatting
within a given scenario, and then transfer the chatting to the guided chatting in a
given scenario, which will be introduced more in next subsection. In Figure 1 the
arrow from the scenario dialogue handler to the communicational response indicates
this relationship.
Motivate the Learners to Practice English through Playing with Chatbot CSIEC 185
Summarily the goal of free chatting is to motivate the user’s talking desire. For this
purpose the robot tries to adapt itself to the user’s interest, and launch new topics.
We then explain our approach to guide the chatting with the robot in a given scenario.
The dialogue should be developed step by step around a red line or a topic for this
scenario. Due to the extreme complexity of natural language this dialogue
development is exceedingly nonlinearly complicated. It can be described by a
complex tree structure with many branches. These branches can be pragmatically and
semantically countable, but syntactically uncountable. So this dialogue scenario tree
is much more complicated than the classical chess decision making programs, which
has finite state change selections in every step.
We use scripts to describe the decision tree in the dialogue on a given topic. The
script is made up of lines of dialogue steps (states), every of which is a branch in the
decision tree. Suppose the robot speaks at first. In every line there must be the text
output from the robot and its order number in the dialogue. This output may be
triggered by specific user input, which we call the prerequisite of this output text. The
robot may also expect the user inputs certain texts, or some texts with specific
semantic or syntactical characters, which we call the expectation of this output text.
We write the line in the script with the following format:
Nr. <prerequisite> (text) <expectation>
The “Nr.” and (text) are the necessary two components in every line. The “Nr.” is
an integer which indicates the line order in the whole script, whereas the “text” can be
any text from the robot, either statement, or question, and so on, and it is written
within closed brackets.
In a script line the prerequisite and expectation are optional. If they appear they
must be written within closed sharp brackets. If the prerequisite exists and is satisfied
the output text can be given out by the robot. The expectation means the robot hopes
the user responses to this text with some specific syntactic and/or semantic features,
and can be applied to instructional goal. For example if the user’s input does not
satisfy the robot’s expectation he/she will face the previous robot output again, until
the expectation is fulfilled. This dialogue pattern can be used for drill. Another
alternative is that the user is given a high mark if his input satisfies the robot output,
otherwise a low mark, although the robot continues the next dialogue. This pattern
can be used in test or examination.
The format of the prerequisite is:
<Nr, variable 1: value 1, value 2...; variable 2: value 1, value 2...>
The format of the expectation is:
<variable 1: value 1, value 2...; variable 2: value 1, value 2...>
Both are almost the same form. Only the prerequisite needs an order number
indicating the expectation of which line this condition fulfills. There may be more
than one value for a given variable. This means if the variable equals any one of the
listed values, the condition is fulfilled, i.e. the values for a given variable have the
relation of logical disjunction. There may be also more than one variable and its
corresponding values. The relation among these variables is logical conjunction.
186 J. Jia and W. Chen
This discourse script is difficult to be written by normal authors, for example the
English teachers who want to use this program to train the students. Even an error
with a bracket will cause the misunderstanding of the computer program. Thus we
have designed a Java GUI, i.e. DSE (Discourse Script Editor) for editing the scripts
step by step more easily. With it a normal user such as an English teacher needs not to
pay attention to the writing format, but to the discourse content and process.
However, he/she has to spend much time on planning this discourse script between
the robot and the human user, just like a film director. This work is not just the
language teaching, but also the teaching of response strategies through natural
language.
We use the Microsoft agent technology to synthesize the output text, because the
agent’s voice is lifelike, the agent’s figures, movements as well as actions can be
designed very vividly, and it can also synchronously display the spoken text, which
facilitates the aural understanding and activates the user’s interests. We have also
designed seven facial expressions (neutral, happy, sad, feared, disgusted, angry and
surprised) for every agent character and hope the textual emotional expressions can be
accompanied by the agents’ facial changes. The robot’s reading speed can be adjusted
by the users at any time. We have also designed a text-reading webpage where the
agent can read any texts inputted or pasted by the user.
Different from the traditional audio technologies such as audio players, the user
confronts with unexpected robot text and voices, just like talking with a real human
being. So it is hoped that this function can benefit the user’s listening comprehension
and prompt response.
This function is designed to aid the user’s chatting on a given topic. With it the users
can watch the talk show of two robots before the human-computer interaction. The
talking texts are predefined by the teacher for the specific context or topic. However,
the actual texts for a given meaning can be expressed randomly. So this kind of talk
show is different from the monotone one presented in the traditional video or audio
cassette. It will enforce the learner’s spontaneous listening and understanding. The
talk show script texts can be readily written by the teachers with any text editor.
exercises, as well as the interface for learners to do these exercises and then to get the
automatic assessment results.
An example of gap-filling exercises is: “I ( ) a student.” The correct answer to the
gap can be: “am”, “want to be”, “will be”, “have been”, “need”, “help”, etc.
In order to motivate the users to learn English we trace users’ usage of different
functions and give them certain scores. The scoring principle is encouraging the usage
of chatting with agents, and with spelling and grammar checking. By the chatting on a
given context, the user is given a high mark if the input satisfies the robot output,
otherwise a low mark. This mark also contributes to the total score.
The user can review his performance and scores after entering the system. This
function is very important and helpful for self learning. A special user who is labeled
as the teacher can access the performance and scores of all the users who are
classified as his/her students. This automatic monitoring function is very necessary
for the teacher to assess the students’ learning behavior and progress.
The internet users get to the CSIEC website (www.csiec.com) mainly through search
engines, because our website has become one of the top 5s in the searching results of
famous search engines such as google.com, yahoo.com and baidu.com by related
keywords such as “chatbot”, “English chatbot”, “Online English learning” in Chinese
or in English, although we haven’t made any large-scale advertisement. The
effectiveness and attractiveness of the system’s adaption to English learning in China
has been somewhat demonstrated by this practical achievement.
With the human-computer dialogues recorded in the database, we make a
summarization of the system’s chatting function from Jan. 20th 2007 to June 20th
2007. The different users who accessed the CSIEC during this period count 1783. The
analysis of the demographic distribution of the users shows that more than half of the
users are undergraduate students. The second large user population is middle school
students. Except 45 students required to use the system in the evaluation, there are
still 377 free users. Totally more than 80% of the users are different kinds of students.
Table 1. The relation between the duration of dialogues and number of users
The average rounds number is 27.4 (48840/1783). The number of the rounds from
each user varies from 1 to 580. From table 1 we draw the conclusion that c.a. 49% of
the users chat with the robot briefly (<=10 rounds); c.a. 46% (38.42%+7.63%) of
them chat with it long or longer; and only few, c.a. 5%, chat with it very long (>100
rounds). Compared with our previous finding in [16] which is listed in the last column
of Table 1, the percentage of the brief chatting with the robot has decreased by
21.78%. Proportionally the percentage of the long and longer chatting has increased.
The dialogue is fluent. I hope the master to enrich the robot’s language.
The kind of communication can improve our English.
The negative comments point out either technical problems or content
shortcoming. Some complain that they can’t use the agent version or the agent voice
sounds curious. Other problems include: the access speed is too low, the robot
response is too slow, the dialogue for a given context is too short, in free chatting the
robot always repeat a same sentence, etc. These problems should be tackled in the
further improvement.
After discussing with the English teachers about the class integration and evaluation
of the CSIEC system we came to a decision that the instructional instruments are on
one side the talk show by two chatting robots and on the other side students’ talking
in English with one robot on a given topic corresponding to the textbook content. The
main application goal is to facilitate role-playing activities in the English classes.
45 high school students in Grade 2 attended the study, and the teacher required the
students to use the system together in the computer room. For the 10 units course
content we designed 40 scenario scripts for the role-play talk show and human-robot
chatting. During the whole term we formally evaluated it through questionnaires,
observations in the classes, surveys with teachers and students focus groups.
The survey contained 6 items about the students’ attitude toward the CSIEC’s
application in English instruction: enhancing fluency of English, enhancing
confidence in communication, enhancing learning interest, mastering practical
expression, improving listening skills, and reviewing key points in course units. All
they were measured with a five-points Likert agreement scale, i.e. the value 5
indicates the maximum best agreement, and 1 is no agreement. The mean is 2.5, 2.8,
3.3, 3.2, 2.9, and 3.3, respectively. It shows that high school students feel the CSIEC-
based English learning can help with course unit review, make them more confident,
improve their listening ability, and enhance the interest in language learning.
Another item in the questionnaire shows 60.5% of the students “liked” or “liked
very much” such a form of English learning, whereas only 2.3% disliked it. 60.5% of
them would continue using the system after class, even without the teacher’s request.
Through the integration and assessment of the system in English class instruction,
some new functions have been added to the system according to the students and
teachers’ suggestions and comments. They include the adjustment of speaking speed
of the agent character, two robots talk show, unlimited gap-filling exercises, etc. Thus
the application and evaluation guide the development of the CSIEC system in the
direction of users’ practical learning needs.
Acknowledgments
We are grateful for the support to our projects from Konrad-Adenauer-Foundation
Germany, Ministry of Education China, Peking University, Education Committee of
Capital Beijing, and Korea Foundation for Advanced Studies.
Motivate the Learners to Practice English through Playing with Chatbot CSIEC 191
References
1. Ministry of Education China. Annual Educational Statistics 2006. People’s Education
Press, Beijing (2006)
2. Fryer, L., Carpenter, R.: Emerging Technologies Bots as Language Learning Tools.
Language Learning &Technology 10(3), 8–14 (2006)
3. Liang, X.: Trial of Multimedia English Instruction Network. Journal of Hebei Medical
University 26(6), 573–574 (2005)
4. Brennan, K.: The Managed Teacher: Emotional Labour, Education, and Technology.
Educational Insights 10(2), 55–65 (2006)
5. Weizenbaum, J.: ELIZA–a Computer Program for the Study of Natural Language
Communications between Man & Machine. Communications of ACM 9(1), 36–45 (1966)
6. Graesser, A.C., Chipman, P., Haynes, B.C., Olney, A.: AutoTutor: An Intelligent Tutoring
System with Mixed-initiative Dialogue. IEEE Trans. on Education 48(4), 612–618 (2005)
7. Seneff, S.: Interactive Computer Aids for Acquiring Proficiency in Mandarin. In: Huo, Q.,
Ma, B., Chng, E.-S., Li, H. (eds.) ISCSLP 2006. LNCS (LNAI), vol. 4274, pp. 1–12.
Springer, Heidelberg (2006)
8. Kerfoot, P., et al.: A Multi-institutional Randomized Controlled Trial of Web-based
Teaching to Medical Students. Academic Medicine 81(3), 224–230 (2006)
9. Abu Shawar, B., Atwell, E.: Fostering language learner autonomy via adaptive
conversation. In: Proceedings of Corpus Linguistics (2007),
http://corpus.bham.ac.uk/corplingproceedings07/paper/
51_Paper.pdf
10. Kerly, A., Hall, P., Bull, S.: Bringing Chatbots into Education: Towards Natural Language
Negotiation of Open Learner Models. Knowledge-Based Systems 20(2), 177–185 (2007)
11. Frege, G.: Begriffsschrift, Eine der Arithmetischen Nachgebildete Formalsprache des
Reinen Denkens. Wissenschaftliche Buchgesellschaft, Darmstadt (1879)
12. Wittgenstein, L.: Tractatus logico-philosophicus. Suhrkamp, Frankfurt am Main (1918/21)
13. Jia, J., Ye, Y., Mainzer, K.: NLOMJ–Natural Language Object Model in Java. In: Shi, Z.,
et al. (eds.) Intelligent Information Processing II, pp. 201–209. Springer, NY (2004)
14. Fellbaum, C.: WordNet: An Electronic Lexical database. MIT Press, Cambridge (1998)
15. Jia, J., Chen, W., Hou, S.: Improving the CSIEC Project through Agent Technology &
Adapting It to English Instruction in China. In: Proceedings of ICCE 2006, pp. 47–54
(2006)
16. Jia, J.: The Study of the Application of a Web-Based Chatbot System on the Teaching of
Foreign Languages. In: Proceedings of SITE 2004, pp. 1201–1207. AACE, VA (2004)
17. The Design-Based Research Collective. Design-Based Research: An Emerging Paradigm
for Educational Inquiry. Educational Researcher, 32(1), 5–8 (2003)
18. Dagan, I., et al. (eds.): MLCW 2005. LNCS (LNAI), vol. 3944. Springer, Heidelberg
(2006)
A Strategy for Selecting Super-Peer in P2P and
Grid Based Hybrid System
1 Introduction
The rapid development of Internet bring us much convenience, following the ex-
plosive datum. Resolving data storage and access is significant.Many techniques
try to solve the problem, like as data warehouse, SAN, P2P, Grid, etc. Both
P2P and Grid can realize resources sharing and have different advantage and
shortage.Therefore, we may combine their merits and establish a convergence
system of P2P and Grid.
Napster,Gnutella and KaZaA are all well known P2P system, and Globus
Toolkit (GT) [16] has been accepted as the mature software toolkit on deploy-
ing Grid, which make it probability combining P2P and Grid. One of GT 4.0
component is Monitoring and Discovery System (MDS). MDS is a suite of web
services to monitor and discover resources and services on Grids. A service for
collecting above information is Index Service included in MDS. Indexes collect
information and publish those information as resource properties. It provides
information to clients by a Web Service interface. Clients may query and sub-
scribe resource properties from an Index. Index Service not only saves the local
useful datum, but also caches the remote datum, and maintains data updating
by lifetime management mechanism. In the large scale Grid, indexes can register
to each other in a hierarchical fashion in order to aggregate data at several levels.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 192–199, 2008.
c Springer-Verlag Berlin Heidelberg 2008
A Strategy for Selecting Super-Peer in P2P and Grid Based Hybrid System 193
2 Related Work
Making use of P2P scalability and dynamic, designing resources location and
discovery protocol for querying Grid resources to improve query success rate
and fault tolerance of system have been already involved into network systems
merging P2P and Grid[1][2][3][4][5][6]. These hybrid systems can efficiently and
reasonably implement resources utilization, which provide a new platform for
sharing digital resources. In these hybrid systems, super-peers may manage gen-
eral nodes concentratedly within certain range and integrate limit resources, in
which super-peers collaborate each other to form a decentralized P2P network
at a higher layer. B.Yang et al. studied pure P2P network based on super-peer
architectures[7]. KaZaA and Gnutella are two typical existed super-peer based
P2P system. Furthermore, Montresor et al. proposed opinions on super-peer se-
lection [8][9][10][11] in their systems, but did not do further research on how
to select. C.Mastroianni et al. devised Grid information service on the basis of
super-peer model[12][13]. C.Pasquale et al. presented a model for job assignment
across the Grid exploiting an underlying super-peer topology[14]. A hybrid and
unstructured network model based on P2P and Grid was introduced in [15].
But these models only gave or mentioned the concept of super-peer. Although
they have agreed on the opinions of how constructed super-peer architecture
according to P2P and Grid, they had not discussed super-peer selection and its
form in detail. Apparently, super-peer plays an important role in the network,
so its performance will influence the whole network performance directly. How
to reasonably select super-peer has been recognized as a new research hot.
This paper designs a super-peer selection algorithm based on capacity. The
rest of paper is organized as follows. Our super-peer selection algorithm based on
capacity is described in detail in section 3. Section 4 discusses the experimental
environments and evaluates the results. Section 5 concludes the paper.
Super-peers are selected from Grid nodes registered to AN, and they connect
to each other to form an overlay network at a higher level using P2P mechanism.
In a VO, the number of super-peer is limited. A super-peer and its managing
subnodes are called as a cluster. Through limiting the size of cluster, it can
guarantee that the amount of super-peers is not far too much.
Algorithm SSABC:
Step 1 Initialization: SP ← φ, CA ← φ
Step 2 if first node S1 joins the network then
SP ← SP ∪ {S1 };
CA ← CA ∪ {cas1 };
Step 3 if the ith node Si joins the network then
CA ← CA ∪ {casi };
Step 4 /* if exist super-peer is non-saturated in SP , then Si connects to the
non-saturated super- peer which is the nearest to Si . Using hop counts com-
putes the distance .*/
if ∃N Pj conn < M ∧ Sj ∈ SP then
Si connects to Sj ;
go to Step 3;
endif
Step 5 /*if super- peers in SP are all saturated, then select a new super-peer.*/
/* if Si ’s capacity is the maximal then select it as super-peer.*/
for each casj in CA
if casj < casi then
SP ← SP ∪ {Si };
CA ← CA − {casi };
else
/* Si ’s capacity is not the maximal, then select a node from CA
whose capacity is the maximal, assume it as Sk .*/
Disconnect Sk from its super-peer by AN announcement;
SP ← SP ∪ {Sk };
CA ← CA − {cask };
endif
endfor
Step 6 /* let every existed super-peers released the number of R nodes ,bal-
ancing the load*/
if Si or Sk as super peer then
R = M − M × |SA|/(|SA| + 1);
R × |SA| subnodes connect to Si or Sk ;
endif
While a client applies for resources, its node will submit application to its
super-peer. If super-peer couldn’t satisfy client’s resources application, it would
search its index table for finding a neighbor super-peer to satisfy client require-
ment. If it still can’t find the target super-peer, the origin super-peer will ran-
domly select a neighbor super-peer on behalf of its client and continue the same
query. In the process of query, we set TTL value to limit the query be carried out
in VO. When TTL decrease to zero, query would fail.Our experiments simulate
the behavior of client querying file resources and other resources such as free
CPU, available bandwidth, memory, etc.
bandwidth and current connection nodes are generated in topology files. The
number of file resources follows geometric distribution. The number of CPU,
memory and online time follow uniform distribution. These values are produced
by Matlab random functions. The literature [18] pointed that cluster size in
KaZaA system was between 60 and 100, while ultrapeers in Guntella system
reached 30 ∼ 40. So we set the proportion (Psp ) having super-peers as 1%∼ 5%
in one VO. The weight for properties of CPU,bandwidth,memory, online time
and the number of connection is 0.25,0.25,0.1,0.2 and 0.2,respectively, .
Query success rate is looked as success frequency divided by total number
of times of success. Successful query which means finding requesting resources
will produce many results, and each result has a latency. Average query delay is
total query latency divided by total query number of times. In order to explain
the advantage of capacity selection (CS) super-peer, we compare it with random
selection (RS) super-peer.
Fig. 1. File Duplicate Copies versus Fig. 2. File Duplicate Copies versus
Query Success Rate Query Delay
In Figure 1, while TTL is fixed, success rate appears ascending trend as the
duplicate copies of file rising from 80 to 120. File duplicate copies versus success
rate nearly presents linear relation. Figure 2 demonstrates that average query
delay doesn’t change obviously for a fixed value of TTL, and the curves are
relatively steady. When TTL increases, average query delay rise from 17ms to
45ms. Figures 1-2 show that our super-peer selection strategy is steady and
reasonable.
Success rate of between CS and RS with different TTL are shown in Figure 3
when Psp is changed from 1% to 5%, where query resources is expressed with ca-
pacity for the purpose of conveniences and simplicity. Capacity values is generated
by random function in program. This method doesn’t effect performance analy-
sis, because comparison of query success rate is based on same capacity,TTL and
Psp . By all appearance, for a fixed number of super-peers, the success rate in CS is
higher than RS. However, the number of super-peers hardly makes any difference
to success rate under random selection.
Fig. 3. Query Success Rate between CS Fig. 4. Query Success Rate between CS
and RS with Different TTL and RS with Different Number of Nodes
198 S.-H. Zhao et al.
Figure 4 shows the comparison of success rate between CS and RS with differ-
ent number of nodes. From figure 4, it is observed that the success rate of both
CS and RS are all not clearly change at the same Psp . On the condition of same
nodes, success rate of CS is still increasing with cluster size, while success rate
of RS doesn’t make better. It can be seen from Figure 3 and Figure 4, increasing
nodes doesn’t influence query success rate, which explains the capacity selection
super-peers having better stability,and shows our capacity selection super-peer
is priority to random selection super-peer and advances the query efficiency.
5 Conclusions
Super-peer model based network system has widely known, and how to se-
lect high performance super-peers is a key problem in these systems. However,
research on selecting super-peer mostly has been done in pure P2P systems.
Utilizing MDS4.0 index service to discover dynamic useful resources and using
resources properties to compute nodes’capacity, this paper presents a strategy for
selecting super-peer according to nodes’capacity in P2P and Grid based hybrid
system. Experimental results demonstrate that success rate of capacity selection
is obviously higher than that of random selection. Therefore, selecting higher ca-
pacity nodes as super-peers in the convergence system of P2P and Grid improves
query success rate and decreases query time, which provides a better platform
for sharing digital resources.
Acknowledgement
The work presented in this paper is supported by the Natural Foundation of
Anhui Provincial Education Department (No.2006KJ041B and No.KJ2007B073)
References
1. Puppin, D., Moncelli, S., Baraglia, R., Tonellotto, N., Silvestri, F.: A Peer-to-
peer Information Service for the Grid. In: Proc. of International Conference on
Broadband Networks, San Jose, CA, USA (2004)
2. Andrade, N., Costa, L., Germ’oglio, G., Cirne, W.: Peer-to-peer grid computing
with the OurGrid Community. In: 23rd Brazilian Symposium on Computer Net-
works - IV Special Tools Session (May 2005)
3. Amoretti, M., Reggiani, M., Zanichelli, F., Conte, G.: SP2A: Enabling Service-
Oriented Grids using a Peer-to-Peer Approach. In: 14th IEEE International Work-
shops on Enabling Technologies: Infrastructure for Collaborative Enterprise (WET-
ICE 2005), pp. 301–304 (2005)
4. Uppuluri, P., Jabisetti, N., Joshi, U., Lee, Y.: P2P Grid: Service Oriented Frame-
work for Distributed Resource Management. In: IEEE International Conference on
Services Computing (SCC 2005), pp. 347–350 (2005)
5. Iamnitchi, A., Foster, I.: A peer-to-peer Approach To Resource Location In Grid
Environments. In: Grid Resource Management. Kluwer Publishing, Dordrecht
(2003)
A Strategy for Selecting Super-Peer in P2P and Grid Based Hybrid System 199
6. Talia, D., Trunfio, P.: A P2P Grid Services-Based Protocol: Design and Evalua-
tion. In: Danelutto, M., Vanneschi, M., Laforenza, D. (eds.) Euro-Par 2004. LNCS,
vol. 3149, pp. 1022–1031. Springer, Heidelberg (2004)
7. Beverly Yang, B., Garcia-Molina, H.: Designing a Super-Peer Network. In: 19th
Int’l Conf. on Data Engineering. IEEE Computer Society Press, Los Alamitos
(2003)
8. Montresor, A.: A Robust Protocol for Building Superpeer Overlay Topologies. In:
Proc. Of the International Conference on Peer-to-Peer Computing, Zurich, Switzer-
land (2004)
9. Li, J., Vuong, S.: An Efficient Clustered Architecture for P2P Networks, In: 18th
International Conference on Advanced Information Networking and Applications
(AINA 2004), vol. 1, p. 278 (2004)
10. Lo, V., Zhou, D., Liu, Y., GauthierDickey, C., Li, J.: Scalable Supernode Selec-
tion in Peer-to-Peer Overlay Networks. In: Second International Workshop on Hot
Topics in Peer-to-Peer Systems, pp. 18–27 (2005)
11. Min, S.-H., Holliday, J., Cho, D.-S.: Optimal Super-peer Selection for Large-scale
P2P System. In: 2006 International Conference on Hybrid Information Technology
- Vol2 (ICHIT 2006), pp. 588–593 (2006)
12. Mastroianni, C., Talia, D., Verta, O.: A Super-Peer Model for Building Resource
Discovery Services in Grids: Design and Simulation Analysis. In: Sloot, P.M.A.,
Hoekstra, A.G., Priol, T., Reinefeld, A., Bubak, M. (eds.) EGC 2005. LNCS,
vol. 3470, pp. 132–143. Springer, Heidelberg (2005)
13. Puppin, D., Moncelli, S., Baraglia, R., Tonelotto, N., Silvestri, F.: A Grid Informa-
tion Service Based on Peer-to-Peer. In: Cunha, J.C., Medeiros, P.D. (eds.) Euro-Par
2005. LNCS, vol. 3648, pp. 454–464. Springer, Heidelberg (2005)
14. Cozza, P., Talia, D., Mastroianni, C., Taylor, I.: A Super-Peer Model for Multiple
Job Submission on a Grid. Technical report, TR-0067, Institute on Knowledge and
Data Management, Institute on Grid Systems, Tools and Environments, CoreGRID
- Network of Excellence (January 2007)
15. Guilin, C., Shenghui, Z., Zhengfeng, H.: A model of integrating P2P technology and
grid technology. Journal of Hefei University of Technology(Natural Science) 30(6)
(2007)
16. GT4.0, http://www.globus.org
17. BRITE, http://www.cs.bu.edu/brite/
18. Qiong, L., Peng, X., Hai-Tao, Y., Yun, P.: Research on Measurement of Peer-to-
Peer File Sharing System. Journal of Software 17(10) (October 2006)
Personal Knowledge Management in E-Learning Era
Abstract. In this paper, the author first introduces some relative theories of
E-Learning and personal knowledge management(PKM), including the concept,
content, objectives and their common ground, then puts forwards the require-
ments to personal knowledge management and how to use personal knowledge
management tools such as iSpace Desktop, iNota, Mybase in E-Learning envi-
ronment to improve personal knowledge literacy.
The new management master of the United States, Peter M. Senge has said "If a person
can’t update what he has learned at the race of 7 percent annually, then he will not be
able to adapt to social change." [1]. The knowledge economic era, which is characterized
by faster and faster speed of knowledge conversion and technical updates, is a new era of
learning which makes learning social and lifelong. E-Learning is derived from
E-commence, and different organizations have different definitions, among which the
definition made by the U.S. Department of Education is: E-Learning is the learning and
teaching activities mainly conducted through the Internet, which make full use of the
learning environment with new communication mechanism and rich resources provided
by modern information technology, so as to achieve a brand-new mode of learning.
The report Knowledge-oriented society Based on E-Learning, published by Ad-
vanced Learning Infrastructure Consortium of Japan in March 2003, pointed out that
E-Learning is a new-type learning environment, an active learning activity conducted by
learner through applying information technology and communication networks.
E-Learning has many features, such as customized courses, active and interactive
learning, learning outcomes and learning process easy-controlled, learning anytime &
anywhere and for anyone, transmission to the scattered learners, transmission quickly
and timely, learning content easy-archived and easy-reuse, etc. [2]. The term E-Learning
refers to the employment of new technology for learning purposes, and two key factors
in any E-Learning system are learning content creation and delivery [3]. The term
E-Learning refers to the digital learning process in digital learning environment, under
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 200–205, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Personal Knowledge Management in E-Learning Era 201
which the learner makes use of digital learning resources, digital learning environment,
digital learning resources and digital learning method are the three basic elements of
E-Learning [4].
For the concept of personal knowledge management, there is still no uniform definition.
Viewing from abroad, the broad definition by Paul, the United States is: PKM should
be seen as a set of problem-solving skills and methods in level of both logical concept
and practical operation [6]. Frand and Hixon believe that PKM refers to a strategy and
process to expand personal knowledge, during which individuals organize and con-
centrate their important information as a part of their own knowledge, and transfer
scattered fragments of information into systemic application information. In addition,
they believe that PKM also includes the expansion of personal knowledge and the
conversion from personal tacit knowledge to explicit knowledge [7]. In domestic re-
search, Dechao KONG believes that the PKM includes three meanings: First, managing
personal knowledge gained; Second, acquiring new knowledge through various
channels, learning from and drawing on the experience and strong points of others to
make up for their own deficiencies in thinking and knowledge, so as to constantly
construct their own knowledge characteristics; Third, achieving dominant change of
tacit knowledge and stimulating innovation of new knowledge by applying their mas-
tered knowledge and the long-standing views and ideas, and combining with other
people's ideological essence and disposing draft [8].
PKM is the management of knowledge resources to achieve personal goals. Based
on PKM, it emphasizes on recording and digging personal tacit knowledge, accessing
knowledge resources necessary for work and learning, promoting knowledge resources
orderly and in self-organization, and promoting reproduction and reuse of personal
knowledge. PKM mainly includes the management of basic personal knowledge
database and the management of personal thinking database. The former includes
personal communications management, personal time management, personal work
202 W. Li and Y. Liu
Professor Dorsey regards that PKM can include and define the following seven core
skills: retrieving information skills, evaluating information skills, organizing informa-
tion skills, analyzing information skills, presenting information skills, securing infor-
mation skills and collaborating around information skills.
If a person wants to obtain the survival and development capacity and to keep pace
with the rhythm, he must become a personal knowledge manager to enhance the ca-
pacity of creating and applying knowledge. In addition, during the process of PKM,
such as the discovery of knowledge, access of knowledge, storage of knowledge, share
of knowledge, pervasion of knowledge, application of knowledge and creation of
knowledge, and so on, involving methods, the exertion of strategy and means are in-
volved. So how to manage knowledge is an important lifelong learning content.
Therefore, it is very important for E-Learning to provide learners with corresponding
tools to help study. In many helpful tools of learning [12], during the learning process if
we can apply knowledge management technology to promote the conversion between
tacit and explicit knowledge, to promote the application of knowledge management
technology and tools and to promote knowledge innovation and sharing, it will be more
helpful to enhance the effectiveness of lifelong learning and to promote individual
lifelong learning more effective.
Knowledge has become the most important personal core resources in this era of life-
long learning. PKM is bound to produce an extensive and far-reaching impact on
education, and corresponding PKM tools will be fully applied in E-Learning. PKM
tools generally support several knowledge management processes, i.e. coding / de-
scribing, classifying / indexing, searching / filtering, sharing / disseminating and
knowledge innovation [13]. PKM tools are different according to their core functions
and can be classified as follows:
(1) Information capture / sharing tool
Such tools drag information (text, graphics, charts, links, etc.) from the Web page or
document to form new different type document. When information sources add new
content or information updates, the tool can inform users, and share the information
with others. The tool has a series of products of Entopia: Quantum Collect, Quantum
Collaborate and Quantum Capitalize, Web2one, Organizer, and so on.
(2) Encoding / describing tool
Such visualization tools can capture, organize and display (in the form of link and
concept) new concepts, and provide with functions of searching, enlarging, expanding
and navigating, etc.. Such tools are Mind map, Brain, etc..
(3) Search / indexing tool
Such tools index local and network drives, support keyword, full-text and natural
language, Boolean expressions, etc.. For example EnFish Find can look for and classify
information according to correlativity, names, documents, E-mail and URL, and then
provide online communities to support collaboration among individuals.
(4) Meta Search tool
Meta search engine will send questions to different database of several search en-
gines to search, and provide a single integrated and graded list. Such tools are
Scarch.com, Dogpile.com, Mamma.com, etc..
204 W. Li and Y. Liu
4 Conclusion
In the current knowledge economic era, knowledge has become an extremely important
factor for personal survival and development. The individual must skillfully master
PKM skills so as to gain the survival and development ability and become a constantly
Personal Knowledge Management in E-Learning Era 205
References
1. Lai, C.S., Lu, Z.H.: Initial Discussion on Personal Knowledge Management Method.
Jinzhou Medical College Journal 21, 76–78 (2000)
2. Advanced Learning Infrastructure Consortium: Knowledge-oriented Society Based on
E-Learning (2003)
3. Henry, P.: E-learning Technology, Content and Services. Education Training 43, 249–255
(2001)
4. Li, K.D.: Digital Learning (1)-Core of Information Technology and Curriculum Integration.
Education Research 100, 46–49 (2001)
5. Luo, Z.M., Chen, C.J.: Research on Definition and Characteristics of E-learning. Foreign
Language Teaching 26, 60–64 (2005)
6. Dorsey, P.A.: What is PKM?
http://www.millikin.edu/webmaster/seminar/PKMnet/whatispkm.htm
7. Personal Knowledge Management: Who, What, Why, When, Where, How?,
http://www.anderson.ucla.edu/faculty/jason.frand/researcher/speeches/educom98pkm/
sld001.htm
8. Kong, D.C.: Personal Knowledge Management. Library Development 3, 17–18 (2003)
9. Ke, P.: Introduction to Information Management, 2nd edn. Science Press, Beijing (2007)
10. Barth, S.: Personal Toolkit: A Framework for Personal Knowledge Management Tools,
http://www.kmworld.com/publications/magazine/index.cfm?action=readarticle&Article_I
D=1406&Publication_ID=83
11. Chen, L.X.: Personal Knowledge Management. Information Science 23, 1072–1075 (2005)
12. Zhong, Z.M.: The Research of Supporting Platform of Asynchronous Q & A for E-learning.
J. of Zhengzhou Univ (Nat. Sci. Ed.) 39, 55–59 (2007)
13. Gan, Y.C.: Personal Knowledge Management in E-Learning Environment. China Educa-
tional Technology 197, 20–24 (2003)
14. http://soft.ccw.com.cn/download (2007.12.06)
15. http://www.360doc.com/showweb/0/0/204026.aspx (2007.12.06)
16. http://www.skycn.com/soft/4039.html (2007.12.06)
Teaching Machine Learning to Design Students
Bram van der Vlist, Rick van de Westelaken, Christoph Bartneck, Jun Hu,
Rene Ahn, Emilia Barakova, Frank Delbressine, and Loe Feijs
1 Introduction
The Department of Industrial Design at the Eindhoven University of Technology
prepares students for a new type of engineering discipline: design and creation of
intelligent systems, products, and related services. These systems, products and ser-
vices require to be adapted to the user and thereby provide a new experience. In the
framework of our Masters program, we offer a course that familiarizes students with a
number of powerful conceptual and intellectual tools to understand and create adap-
tive behavior at a system level.
System level thinking has had and still has an enormous impact upon the develop-
ment of technology. When working at a system level one does not study individual
component behavior, such as Ohm's law for an electrical component; instead one
addresses bigger questions such as the stability of the feedback loops, information
throughput, or learning capacity. The learning objective includes classical control,
reinforcement learning and adaptive control and pattern recognition. The context of
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 206–217, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Teaching Machine Learning to Design Students 207
We will now provide a more in depth view on the content of the course, but we
would like to emphasize that the method may also be applied to teach different as-
pects of computer science. In our specific course, the goal is to teach the principles of
reinforcement learning and supervised learning to design students.
2.2 Participants
The participants of our course are all industrial design master students, who can be
classified into two types. The first group consists of students who have a certain affin-
ity with technology. These students like to explore technological principles that are
new to them. They have a good understanding about a wide range of technologies and
their applications. They also have considerable programming skills, with JAVA as
solid basis. This group of students is usually the smaller of the two groups and teach-
ing them machine learning is easier. They might even be satisfied with the traditional
non-embodied method, but using the Lego NXT platform considerably increases their
motivation.
The second group of students can be described as students who do not have an in-
herent affinity with technology. They have a limited understanding of technological
principles and master programming only up to a basic level. Teaching these students
machine learning is the true challenge. It still needs to be acknowledged that students
of either type are not mathematicians or computer scientists. These students are used
to the creative creation of artifacts and not to formulas and algorithms. The teaching
method needs to adapt to these characteristics.
Teaching Machine Learning to Design Students 209
3 Material
For an embodied intelligence course, software and equipment is necessary. While the
software is available for free, the hardware does require a certain budget. The basic
Lego Mindstorms Education NXT set is currently available for 285 Euro. Our practi-
cal experience shows that one set can be shared by a maximum of two students. We
will now discuss the required hardware and software in more detail.
3.1 Hardware
The NXT brick is part of the Lego Mindstorms set. The NXT is an embedded system
with a plastic casing compatible with the Lego brick system. This way it can easily be
integrated into a Lego construction that may also contain the sensors and actuators
[7]. Using Lego saves a lot of time in constructing mechanical components compared
to other methods. An educational version is available that includes the useful
rechargeable battery, a power supply and a storage box. The NXT specifications are:
• Atmel 32-bit ARM main processor (256 Kb flash, 64 Kb RAM, 48 MHz)
• Atmel 8-bit AVR Co-processor (4 Kb flash, 512 Byte RAM, 8 MHz)
• Bluetooth wireless communication (CSR BlueCoreTM 4 v2.0 +EDR System)
• USB 2.0 communication (Full speed port 12 Mbit/s)
• 4 input ports: 6-wire interface supporting both digital and analog interface
• 1 high speed port, IEC 61158 Type 4/EN 50170 compliant
• 3 output ports: 6-wire interface supporting input from encoders
• Display: 100 x 64 pixel LCD black & white graphical display
• Loudspeaker: Sound output channel with 8-bit resolution (Supporting a sam-
ple rate of 2-16 KHz)
• 4 button user-interface
• Power source: 6 AA batteries or rechargeable Lithium-Ion battery.
Lego has developed a number of sensors and actuators as part of the Lego Mind-
storms set. All these sensors are compatible with the Lego brick system. The basic
Lego NXT Education set contains the following sensors and actuators:
• Touch sensor – detects when it is being pressed by something and when it is
released again.
• Sound sensor – detects both decibels [dB] and adjusted decibel [dBA].
• Light sensor – reads the light intensity in a room and measure the light inten-
sity of colored surfaces.
• Ultrasonic sensor – measure distances from 0 to 255 centimeters with a preci-
sion of +/- 3 cm.
• Servo motor with build-in rotation sensor.
As result of the success of the Lego Mindstorms, other companies developed addi-
tional sensors and actuators. Some of these companies, such as HiTechnic Products
and Mindsensors.com, provide sensors for the NXT platform such as IR Link Sensor,
Gyro Sensor, IR Seeker Sensor, Compass Sensor, Color Sensor, Acceleration / Tilt
Sensor, Magnetic Compass, and Pneumatic Pressure Sensor. In addition to the Lego
210 B. van der Vlist et al.
NXT set, a standard computer is needed to write the programs. The programs are then
uploaded to the NXT using either USB or Bluetooth.
3.2 Software
Three software components are necessary for this course. All of them are available for
free and they replace the original Lego software. Lego’s own software development
tool is targeted at children and hence does not offer the flexibility and extendibility
required for a university course. An extensive tutorial on how to install the
components is available at: http://www.bartneck.de/work/education/masterClassLego/
javaInstallNXT/. We will now describe the components in detail.
Java is a platform independent, object-oriented programming language
(http://www.sun.com/java/). The language derives much of its syntax from C and C++
but has a simpler object model and fewer low-level facilities. Java applications are
typically compiled to bytecode, which can run on any Java virtual machine (JVM)
regardless of computer architecture. It is a popular language for embedded systems,
such as micro controllers and mobile phones and also the Lego NXT is capable of
execute Java programs.
It is advisable to use an integrated development environment (IDE) to write Java
programs. Eclipse is the powerful and widely used IDE that offers excellent support
for Java and the Lego NXT. Eclipse itself is written in Java and its installation is par-
ticularly easy.
To enable the Lego NXT to execute Java programs, its original firmware needs to
be replaced with the open source leJOS firmware [8]. The old firmware can be rein-
stalled at any point. Conveniently, leJOS includes a Java Virtual Machine so that no
further software installations on the NXT are necessary to execute Java programs. The
leJOS Java library is an extension to the standard Java and enables Java programs to
use the platform specific features of the NXT, such as sensors and actuators.
The Java Object Oriented Neural Engine (Joone) is an application that allows users
to build, test and train neural networks (http://www.joone.org). It features a conven-
ient graphical user interface. Neural networks trained with Joone can be exported and
called from any external Java program. It can therefore easily be integrated into more
general Java programs, such as Java programs for the Lego NXT.
movements, but after a few minutes it really finds a kind of rhythm allowing it to
move the arm and thereby move itself efficiently forward.
A second robot that was built by different students during this week was Johnny Q
(see Figure 2). It has wheels and left-right motor drives to move forward, backward,
rotate left and right. Johnny Q measures the brightness of the floor and “sees” the
distance from a wall or reference object. Inside is an NXT control brick, an embedded
processor programmed in Java to execute the reinforcement learning algorithm (Q-
learning). The reward is being tapped on the shoulder; a simple button serves to count
touches. Johnny Q learns by being trained. Depending on what the human user does
or does not reward, Johnny Q learns behaviors, such as turning away from a dark
spot, or running backwards near an obstacle. But it can also learn the opposite behav-
ior, bumping against the wall. It explores its possibilities and learns how to accumu-
late maximal rewards. The observer engages in a training session, teaching tricks and
little games, much like training a dog. Usually this algorithm is demonstrated through
screen demos but here the potential of embodied learning is visible in a truly embod-
ied model. From a semantic point of view, it is interesting to sculpt the behavior
which (of course) requires some patience. Johnny Q will gradually forget although
desired behavior can be maintained through continued training.
We will now discuss the Q-learning theory in more detail to enable the reader to form
a better judgment of the difficulty that the students were able to overcome during one
week by using our teaching method. Q-learning is a common and well known rein-
forcement learning algorithm [9]. Reinforcement learning is a method that allows a
machine to learn behavior through receiving rewards and punishments. When a ma-
chine performs an action in a certain state it can get a positive reward, negative re-
ward (punishment) or no reward. These rewards reflect the design and goal of
the machine. The Q-learning algorithm [10] works by constructing an action-value
212 B. van der Vlist et al.
Fig. 2. Johnny Q
function that gives an estimate of the expected value Q(s,a) (the total award that may
eventually be accumulated) when taking a given action “a” in a given state “s”.
Through experience the machine achieves better and better estimates of the action
values. The behavior of the machine is given in terms of a policy. The policy deter-
mines the probability that the machine will take a certain action in a certain state. An
important dilemma in determining the policy is whether the machine should exploit
its knowledge and choose the actions that lead to the biggest reward or that it should
explore new actions in certain states to discover better ways to retrieve even more
rewards later on.
The strength of Q-learning is that it will adapt to its environment without knowing
it and without being programmed. Q-learning, as well as other reinforcement learning
principles, works because it tries to optimize a given reward. Q-learning requires a
finite set of environment states, a fixed set of actions, and a reward function:
π (s) : Α (s) → [0,1]
∀ s ∑ a∈Α ( s) π (s, a) = 1
⎧∞ ⎫
Q π (s, a) = Επ ⎨∑ γ kτ t + k +1 | st = s, at = a ⎬
⎩ k =0 ⎭
π
Q (s, a) = max Q (s, a)
*
π
We write ∫ for the exploration factor, γ for discounting factor, α for learning factor,
π for policy ( ∫ -greedy), s for state, a for action, t for (discrete) time, Α (s) for action
set, Q(s,a) for expected return in state s after action a, under current policy, Q*(s,a) for
expected return in state s after action a, under optimal policy, and Ε for expectation.
training remotely on a PC, is that the network itself is unlikely to stand by itself. Most
likely, it would be integrated into other software. This software needs to be created on
the computer anyway and hence there was little reason to renounce the use of a com-
puter. Once the neural network is trained, it can be transferred back to the NXT. It can
then be used as a standalone application or as part in another program. The students
were able to take advantage of their previous design education to create a highly us-
able GUI for the software. Hopefully, this will encourage other Lego users to take
advantage of their software.
1
Oj = m
− ∑ xi ∗wij
1+ e i=1
where,
O j = output value of neuron j
( ) (
δ j = t j − O j ⋅ O j ⋅ 1 − O j , where, )
δ j = error signal for neuron j (in output layer)
t j = desired output
O j = actual output
(
δ j = Oj 1 − Oj )∑ δ k
⋅ wkj , where,
k
O j = actual output
wij (t + 1) = new weight for synapse between neuron i and j (in all layers)
η = learning rate
δ j = error signal on output
6 Conclusions
We described the embodied intelligence method to teach machine learning to design
students. By using a tangible embodiment as a platform for machine learning, the
environment of the machine-learning program does not need to be simulated. But
more importantly, the embodiment provides the student with a tangible tool to under-
stand and interact with a learning system. Lego Mindstorms NXT is a good platform
for this embodiment. The Lego system allows the students to quickly build a machine
and thereby enables students to focus on the machine learning. In addition Lego NXT
provides a Java Virtual machine on which students can execute Java programs. Java is
a widely used object-oriented programming language. The combination of the Lego
construction system and Java is a very low hurdle that even students who do not have
an affinity toward technology can overcome.
Many of the students played with Lego during their childhood. This positive mem-
ory might have lowered inner barriers that technophobic students might have built up.
It might have allowed them to approach the course with a more open attitude and
thereby increased the opportunity for learning. A second factor that might have had
positive influence on the students is the behavior of the robots. The Crawler robot
demonstrates that even simple learning behavior embodied in Lego has the power to
create affection and empathy with human observers. This might have further moti-
vated the students to experiment with the machine-learning program.
However, the embodied intelligence method does not only offer advantages for less
technophile students, but it also offers enough room for advanced development.
Within only three days certain students were able to build and use neural networks.
They then continued to build their own neural network program from scratch, utiliz-
ing on the theory they learned in the preceding week. In the end, they were able to
create neural network software that is user friendly enough for the general Lego en-
thusiast. As an example application, they built a voice command system, which en-
ables the Lego NXT to operate as a stand-alone voice controlled device. Again, we
have to emphasize that these were neither computer science students nor mathemati-
cians. These were design students that normally create artifacts.
Only by enabling design students to understand, use and develop machine-learning
systems, we can ensure that they will be able to create truly intelligent systems, prod-
ucts, and related services. The embodied intelligence teaching method can help
achieving this goal and our experiences suggest that this has the potential to signifi-
cantly help students who do not have an inhering affinity towards technology.
References
1. Voorhees, R.A.: Measuring what matters: competency-based learning models in higher
education. Jossey-Bass, San Francisco (2001)
2. Weert, v.T.J.: ICT-rich and Competency Based Learning in Higher Education. In: Kallen-
berg, A.J., Ven, M.J.J.M. (eds.) The New Educational Benefits of ICT in Higher Educa-
tion. Erasmus University Rotterdam, Rotterdam (2002)
Teaching Machine Learning to Design Students 217
3. Kyffin, S., Feijs, L.: The New Industrial Design Program and Faculty in Eindhoven. De-
signing Designers - International Convention of University Courses in Design, Milan
(2003)
4. Bartneck, C., Jun, H.: Rapid Prototyping for Interactive Robots. In: Groen, F., Amato, N.,
Bonarini, A., Yoshida, E., Krose, B. (eds.) 8th Conference on Intelligent Autonomous Sys-
tems (IAS-8), pp. 136–145. IOS Press, Amsterdam (2004)
5. Bishop, C.M.: Pattern recognition and machine learning. Springer, New York (2006)
6. Nehmzow, U.: Mobile robotics: a practical introduction. Springer, London (2003)
7. Gasperi, M., Hurbain, P., Hurbain, I.: Extreme NXT: Extending the LEGO Mindstorms
NXT to the Next Level. Apress, Berkeley (2007)
8. Solorzano, J.: LeJos (2002), http://lejos.sourceforge.net/
9. Sutton, R.S., Barto, A.G.: Reinforcement learning: an introduction. MIT Press, Cambridge
(1998)
10. Watkins, C.: Learning from Delayed Rewards. King’s College, vol. PhD. Cambridge Uni-
versity, Cambridge (1989)
11. Haykin, S.S.: Neural networks: a comprehensive foundation. Prentice Hall, Upper Saddle
River (1999)
A Survey on Use of “New Perspective English Learning
System” among University Students—Case Study on
Jiangxi Normal University
1
School of Communication, Jiangxi Normal University, Nanchang, P.R. China
2
WanNian Middle School, WanNian, P.R. China
zjing727@163.com
1 Introduction
New Perspective English Learning System (NPELS) was designed and developed by
Shanghai Foreign Language Education Press according to Instructional Requirements
of College English Course (for trial implementation) published by Ministry of
Education of PRC in 2003[1]. As an autonomous e-learning system based on the
modern concept of teaching foreign language and supported fully by network
technology, NPELS consists of three parts, i.e., interface of system administrator,
interface of teacher management and Learning Hall for students. Among these,
Learning Hall is the main part of the most significance because it is a public English
learning platform open to the non-English major university students and it focuses on
providing rich out-of-class resources of English learning and constructing an
individualized autonomous e-learning environment.
So far as we know, over 180 China’s universities and colleges have put this
e-learning system into practice. However, there is no scientific evidence or data
available which can indicate the effectiveness of NPELS use. Has it really improved
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 218–229, 2008.
© Springer-Verlag Berlin Heidelberg 2008
A Survey on Use of “New Perspective English Learning System” 219
2 Theory
Although NPELS has been widely used in a number of China’s universities, little
research has been done directed towards its effectiveness. Up to now only three papers
associated with NPELS can been searched in China National Knowledge Infrastructure
Databases and these papers are conducted in a highly empirical way without full
evidence from quantitative analysis. Moreover, only one master dissertation was found
in Master Dissertations Full-text Database of JNU with its focus on NPELS-based
autonomous learning strategy of university students instead of NPELS effectiveness
evaluation. Hence a full-scale research conducted in a highly scientific manner must be
undertaken. This research aims to provide theoretical and statistical analysis to examine
different aspects of use of Learning Hall.
The theoretical constructs pertinent to this research are constructivism and the
theory of web-based instruction evaluation.
Firstly, the constructivism believes that learners will construct their own knowledge
and make a meaningful learning if they are put into a well-organized socially
cultural-rich context [2].
Secondly, the theory of web-based instruction evaluation (He Kekang, 2002) is
chosen to be the basis for this research because of its solid theoretical foundation and
the fact that it has been proven successful in numerous empirical studies. From the
perspective of this theory, the factors of web-based instruction evaluation include
learners, teachers, web-based instructional Support Platform, instructional contents and
learning support service [3].
3 Methodology
3.1 Hypotheses
Based on the theory of web-based instructional evaluation as the above discussed, and
according to the practical situation of NPELS use, we assume five factors when
examining the effectiveness and correctness of Learning Hall, which are respectively
learners themselves, learning resources, support from teachers, e-learning Support
Platform and learning support service. And we have the following hypotheses to
explain what attributes a highly satisfactory e-learning system should have:
H1: Students are supposed to have authentic experience of entering Learning Hall
and e-learning before they make any comment towards it.
H2: The effectiveness of learning resources is positively associated with learning
content providing, structure and guiding of the interface and interaction function.
H3: The satisfaction to support from teachers is positively associated with their
attitude and how they offer materials to students.
H4: The effectiveness of e-learning Support Platform is positively associated with
its instructional and technical function.
H5: The satisfaction to learning support service is positively associated with “help”
function.
First of all, through collecting basic information of the students, the respondents cover
a wide range of different majors background. As Table 1 shows, the rates of the
A Survey on Use of “New Perspective English Learning System” 221
investigated students are in a fairly balance seen from their grade and sex. The rate of
the art and science students, however, shows a visible difference, which is mostly
because the enrollment of science major students surpasses considerably that of art
major students in JNU recent years. Hence the validity to this research results is
ensured.
The other single question asking when you usually login Learning Hall is used to
measure the students’ attitude towards learning in Learning Hall. We find out that
39.87% of the respondents login in normal times whereas 28.76% login typically when
an English exam is approaching, left the rest 31.37% login in uncertain time. It shows
that nearly one third of the students are motivated to use Learning Hall by exams rather
than their interest.
Learning content. The learning process in Learning Hall can be divided into several
stages as follows—login, different-level test, class choosing, modular learning, modular
test, promotion to upper module and promotion to upper grade.
Firstly, seen from the data about “the most satisfactory stage and the most
dissatisfactory stage in the whole learning process”, it is clear that the students are most
satisfied with “modular learning” and most dissatisfied with “modular testing”(see
Figure 1). Meanwhile it is found easily that some of the students don’t express an
explicit opinion by choosing “none”, causing the rate of most satisfactory reaches
29.14% and that of most dissatisfactory 24.50%, which produces a close result.
When being asked the reason of choosing “none”, some students explain that they
are not familiar with the whole learning process in Learning Hall. Some of them only
enter and keep opening the platform interface for a long time without further visit to the
learning content patiently and earnestly so that they can meet the demand from the
Academic Affair Office of JNU (i.e. ten hours’ log on Learning Hall accumulatively
each semester) and avoid failure in written English exam of class-teaching. Then why
the most dissatisfactory stage is modular test? In the interview, the students give a
common explanation that should be taken into account. Learning Hall requires that the
students can come into “modular test” stage only after having learned in the prior
“modular learning” stage for a fairly long time. Hence it keeps many students out of
reach of “modular test” stage, leading to a dissatisfied inclination.
When being asked the reason of choosing “none”, some students explain that they
are not familiar with the whole learning process in Learning Hall. Some of them only
222 J. Zhang and M. Li
Fig. 1. Most satisfactory stage and most dissatisfactory stage in the whole learning process
enter and keep opening the platform interface for a long time without further visit to the
learning content patiently and earnestly so that they can meet the demand from the
Academic Affair Office of JNU (i.e. ten hours’ log on Learning Hall accumulatively
each semester) and avoid failure in written English exam of class-teaching. Then why
the most dissatisfactory stage is modular test? In the interview, the students give a
common explanation that should be taken into account. Learning Hall requires that the
students can come into “modular test” stage only after having learned in the prior
“modular learning” stage for a fairly long time. Hence it keeps many students out of
reach of “modular test” stage, leading to a dissatisfied inclination.
“Modular learning” is the most important stage in the whole process and it contains
various units. The question of “the most satisfactory unit and the most dissatisfactory
unit in modular learning” is asked and data is analyzed (see Figure 2).
Figure 2 displays that totally the students are most satisfied with “supplementary
resources for learning” with a high rate of 43.70%. It has some relevance into the fact
that the students have easy access to all kinds of rich complementary resources for
learning in this stage. And the most dissatisfactory unit is “meeting teacher”. Ideally, by
“meeting teachers”, students can make an online appointment with a desired teacher
first and then get a face-to-face instruction from the teacher. But the premise of ful-
filling this is that students must pass unit test prior to meeting teachers. However, vir-
tually few students meet teachers due to the difficulty in passing unit test for various
reasons although its function is quiet desired by students.
Structure and guide of Learning Hall. Among the 151 respondents, 79 students
(accounting for 52.32%)think the structure and guide of Learning Hall is designed ra-
tionally which always prevent them from getting lost in the process of e-learning; 39
students (accounting for 25.83%) think it unreasonable and the rest 25 students (ac-
counting for 21.85%) are not clear about this point.
A Survey on Use of “New Perspective English Learning System” 223
Fig. 2. Most satisfactory unit and most dissatisfactory unit in modular learning
The data reflects that totally the structure and guide design of Learning Hall is well
accepted and effective. However, “getting lost in information and network” will also
occur when learning resources are presented unreasonably. Therefore structure and
guide design and learning resources design both play a key role in preventing getting
lost and producing effective learning result.
Interaction. Interaction is a significant sub-dimension of learning resources and it is
embodied in the following questions and data analysis. As Table 2 shows, nearly 60%
students approve that Learning Hall provides a fully autonomous learning setting but at
the same time almost half students feel somewhat lonely during learning in Learning
Hall. What’s more, over 70% students are positively interested in trying to use
“meeting teachers” if it is available.
Fig. 3. On which aspects should teachers improve their function in Learning Hall
As for the Multi-optional question of “on which aspects should teachers improve
their function in Learning Hall”, Figure 3 displays that an overwhelming majority
among students hopes their teachers to improve their function on various aspects. The
top three aspects they want teachers to improve are respectively to offer more extensive
English learning materials, to give timely online feedback and to organize online in-
teractive communication or arrange cooperative learning activities among students.
Offering learning materials. The second dimension presenting supports from teachers
is the situation of their offer learning materials to students, which are mainly embodied
in the quantity and the update frequency of the uploaded resources by teachers. Figure 4
below reflects that the investigated students are not very satisfied with the uploaded
materials from their teachers. It’s general that the uploaded materials are not in variety
and lacks frequent update.
A Survey on Use of “New Perspective English Learning System” 225
As the forth dimension of the questionnaire, E-learning support platform’s function can
be evaluated instructionally and technically.
Instructional function. Instructional function of e-learning support platform is typi-
cally reflected in its support to students’ learning by providing various resources and
tools. To the question of “whether the resources and tools of Learning Hall can meet
your needs of autonomous learning”, 52.32% students choose “no” whereas only
22.51% students choose “yes”, left the rest 25.17% have an unclear opinion on it. In the
further interview, some students complain that many tools perform practically no
function like forum, meeting teachers, not to mention tools for exploratory or col-
laborative learning.
Technological system. When being asked “ how often tech breakdown happens when
you are using Learning Hall”, the number of respondents who choose “often” or “oc-
casionally” accounts for 97% in comparison to the 3% students who says they seldom
meet tech breakdown. And the top three kinds of tech breakdown told in the interview
are failure to log on, low speed connection to Internet and getting disconnected with
line suddenly.
necessary helps (e.g. English learning strategic helps and e-learning helper) is not in-
cluded, left only 3.31% respondents’ affirmative opinion on the help service.
Through the above data collection and analysis, undoubtedly NPELS Learning Hall is
faced with many various problems in both its initial functional design and practical
implementation in school. There is an urgent need to solve these problems and improve
the function of the whole Learning Hall platform in many ways. Hereby some discus-
sion and corresponding suggestions are proposed.
Both the questionnaire and interview reflect that some students show a very ambiguous
opinion on many basic attributes of Learning Hall such as learning process, internal
structure and function by choosing the answers like “not clear”, “None” or “I don’t
know”. It is mostly because it’s too easy for the students to meet the demand from the
school—if a student has stayed online in Learning Hall for over ten hours in one se-
mester accumulatively and then he or she will be qualified whatever his or her attitude
and performance in the whole learning process. We believe only a time demand is not
enough. An effective learning process monitoring and assessment is neglected in both
the design and the implementation of Learning Hall and it is urgent to be developed at
the earliest.
To some extent the teachers in NPELS should be accountable to the inactive per-
formance of the students. The teachers should more care for students and analyze what
the students need by participating in their learning activities and sharing their learning
experiences. Rich materials uploading and frequent updating should be made. More-
over, to arrange cooperative learning activities among students, and to give timely
online feedback should also be indispensable. In this way, we believe students will
view Learning Hall as an effective approach to improve English learning.
Some problems on NPELS itself are discovered in the survey and interview which the
platform developers should take into account. On the other hand, e-learning platform
urgently need to optimize its function in providing interactive and cooperative learning
tools such as forum, Netmeeting and chatting room, etc. They should be included and
work practically instead of being an empty shell in NPELS. Particularly “meeting
teachers” function should be more accessible and open to students without the con-
straint of passing a test. On the other hand, learning support service should be more
individualized and works in humanity manner. Besides the system using help,
e-learning strategy helper, English learning method assistant should be embraced in the
whole learning support service. In this way, students will be motivated to make an
active learning.
6 Conclusion
This research, focusing on the case study of JNU, reflects the common problems
sharing among the universities and colleges of the same kind in China. This paper
suggest a fully improvement be obtained on the aspects of system management,
teachers’ support, platform function and learning support service. Hopefully, the ef-
fectiveness and correctness of NPELS Learning Hall will be promoted after a series of
improvement and optimization in overall function of the system.
Like most research, this study is not without limitations. The results in this study
may have limited to the universities of different types in China. Future research could
choose various kinds of universities or colleges as samples so that a more comprehen-
sive result may be achieved.
228 J. Zhang and M. Li
References
1 Introduction
Game Industry, as a complex of medium which includes story telling, artworks, sound
and techniques such as Math, Physics, rendering, has been developed fast enough to be
the mainstream of cultural industry. When the Graphics and Sound techniques have
reached at certain level, the game users want the games more realistic and interesting.
For these reasons, the Artificial Intelligence (AI) technique has become more and more
important in game industry [1]. As computer games become more complex and con-
sumers demand more sophisticated computer-controlled NPCs, the game developers
are required to place a great emphasis on the AI. In order to make the game more re-
alistic and smarter, high level AI has been strongly demanded by game users since
1990s. For that, game programmers can educate/train game NPCs using various AI
techniques. When designing the specific AI, it is important to choose adaptive AI
method/algorithm. If we use low-level algorithm, people always can predict the action
of the NPCs. It will make the game boring. Recently there has been much interest in
combining evolutionary algorithms and artificial neural networks [2], [3], [4], [5], [6],
[7]. Evolutionary Artificial Neural Networks (EANNs) has been highly effective in
training NPCs in video games, because EANNs can evolve the whole structure of the
Neural Networks which can be represented as behavior of NPCs [16].
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 230–239, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Evolving Game NPCs Based on Concurrent Evolutionary Neural Networks 231
As we can know from the real football games, every player not only has to consider
his own position, his team members’ positions, enemy team members’ positions, and
ball position and so on, but also has to decide his action to win the game. When we
simulate the whole things in digital football games, it is too hard to check every cir-
cumstance. In other words, we hardly use rule-based systems or Finite State Machine
(FSM) to design AI. However, it is possible to use EANNs. We can encode the virtual
player in a form of Neural Network controller, all the positions we described above set
to be neural inputs and the actions to be the outputs.
As we described above, the player must pay attention to so many positions that the
evolutionary algorithm can be considered as the optimization problem of the high
dimensional search space. We focus on (1) choosing a principled method of high di-
mensional data space, (2) designing adaptive fitness functions which can make the right
evolution. We describe the Concurrent Evolutionary Neural Networks (CENNs) based
on EANNs for competitive team game playing behaviors by teams of virtual football
game players. We choose Darwin Platform as our test bed to show its efficiency.
Darwin is ergonomics A.I. game platform developed in the project of Ergonomics
Game Intelligence [8].
2 Related Works
In the area of using EANNs to design game AI, one of the main problem is finding
suitable structures for evolving adaptive game NPCs [16].
Enforced Sub-Populations (ESP) [9], is a kind of EANNs system based on Symbi-
otic, Adaptive Neuro-Evolution (SANE) [10]. In ESP, populations of neurons are
evolved to form a neural network. It differs from other EANNs systems in that it
evolves a population of neurons instead of complete networks. A neuron is selected
from each population to form the hidden layer of the network, which is then evaluated
on the problem and its fitness passed back to the participating neurons. These neurons
are combined to form neural networks that are then evaluated on a given problem. ESP
differs from SANE in that neurons for the different positions in the network are evolved
in separate subpopulations. ESP has been used in training the agents in some video
games [11] and was shown to be significantly faster than other EANNs methods.
Another EANNs method used in developing game AI is NeuroEvolution of Augmenting
Topologies (NEAT) [15]. NEAT starts from the neural networks with small structures and
becomes increasingly sophisticated over generations. This technique is appropriate for the
real-time game where the agents are repeatedly created and eliminated.
In sum, the approach in using EANNs to design game AI is a useful integration. In
our work, we try to design the AI of football game NPCs where team-cooperative be-
havior is expected and propose a method which can evolve the weights while decreasing
the number of connection of the neural networks, as we will discuss in Section 3.
2.2 Darwin
Darwin, a game platform, based on agent that developers embody AI easily and capable
of proposing AI test with module that makes them find strategic action and
232 X.H. Jin, D.H. Jang, and T.Y. Kim
then evaluate achievement results through making agent used strategic module that
Darwin offers.
Darwin consists of AI-Level and Game-Level (Fig. 1 shows the framework of
Darwin Platform). Game designer can make DLL using the template that Darwin
provides. When AI designer develops a DLL, Darwin-Manager in Game Level pro-
vides the main program functions of Neural Networks, Genetic Algorithms, FSM, basic
EANNs and so on. DLL developed by game AI designer can be loaded by
Darwin-Loader in Darwin-Manager and then the NPCs’ action can be seen from Dar-
win-Viewer.
8. Check criterion. Repeat 3-7 until the fitness value reaches at the criterion.
9. Save and load. Save the best architecture in the whole process and load in the
game.
4 Experiment
In the virtual soccer game, played by Blue team and Red team, the two field players in
Red team will be evolved during the game play. The Blue team is pre-programmed by
Rule-Based System to help the neural network players evolve successful. We name the
target field players as NEPlayer evolving with CENNs algorithm.
NEPlayer has two separate neural network systems depending on the status that the
player has the ball or has not the ball.
1) When NEPlayer has not the ball: The network takes the inputs from the coordi-
nates of friendly players, enemy players and the ball. Darwin test bed provides the
sensing ability of each positions of the players or the ball. The sensored inputs are
transferred to the outputs through the weighted nodes of the hidden and output neurons
resulting from sigmoid functions. The outputs control the moving direction or target
position of NEPlayer without the ball. The combination of output and the corre-
sponding directions is shown in Table 1.
Output Behavior
000 Move to the midpoint between the ball and the enemy goal
001 Move to the midpoint between the nearest friendly player and ball
010 Move to the midpoint between the second nearest friendly player and ball
011 Move to the midpoint between the third nearest friendly player and ball
100 Move to the ball position
101 Move by 0 degree direction
110 Move by 120 degree direction
111 Move by 240 degree direction
2) When NEPlayer has the ball: The network takes the same inputs as in the previous
case except the ball position. The outputs control not only the moving direction or
target position of NEPlayer but also the power to kick the ball. We have prepared
different expected behavior from the Table 1. For example, if the first to third outputs
are 000 and the fourth one and fifth one is 00 (the fourth, fifth ones are one of 00~11
controls the kicking power of the ball ), then it will kick the ball at specified direction in
Table 2 with 1/4 multiplied by the maximum kick power. The combination of output
and the corresponding directions is shown in Table 2.
Output Behavior
000 Kick to the nearest friendly player
001 Kick to the second nearest friendly player
010 Kick to the third nearest friendly player
011 Kick to the midpoint between the third nearest enemy player and the farthest
enemy player
100 Kick to the midpoint between the second nearest enemy player and the farthest
enemy player
101 Kick by 0 degree direction
110 Kick by 120 degree direction
111 Kick by 240 degree direction
act as a human player. We must consider as many situations as possible that can be
taken place during the game play. Like the Neural Network architecture, the fitness
function has two separate cases depending on the status that the player has the ball or
has not the ball.
The game is initialized when either team get a goal. At the initialization step, the
fitness functions are evaluated considering the three situations as follows.
1) When NEPlayer has not the ball: In this case, we can calculate the fitness based on
which team has the ball and how well the members are distributed. The all distances
between each player are summed and normalized to calculate the distribution bonus in
Darwin Platform. Let f be the fitness score, NF be the number of the friendly team has
the ball until any player gets a goal, NE be the number of the enemy team player has the
ball until any player gets a goal, NBonus be the number of the players distributed well, T
be the frames from starting to the moment of getting a goal, TF be the time limit. In our
experiment, considering that it takes about 1800 frame to get a goal in the Darwin
Platform, we set the time limit by 3000 frame. The formulas are shown below:
i. When friendly team gets a goal:
f = ( N F − N E + N Bonus ) / T (3)
ii. When enemy team gets a goal:
f = −( N F − N E + N Bonus ) / T (4)
iii. When neither team gets a goal until time limit:
2) When NEPlayer has the ball: In this case, the fitness function evaluates its fitness
according to the following rules. If the teammate gets the ball, it adds 2 points to the
fitness value but if the enemy player gets the ball, it loses 2 points. When nobody gets
the ball, it calculate the nearest player to the ball then, it will add or subtract 1 point
depending on the nearest player's team. When the game ends, these values are summed
up and set to the fitness value of each game. Let f be the fitness score, NF be the number
236 X.H. Jin, D.H. Jang, and T.Y. Kim
of the NEPlayer correctly kick to the friendly team member until any player gets a goal,
NE be the number of the NEPlayer kick to the enemy team player until any player gets a
goal, NNF be the number that the nearest player to the ball is friendly team player when
nobody takes the ball, NNE be the number that the nearest player to the ball is enemy
team player when nobody takes the ball. The formulas are as follows.
i. When friendly team gets a goal:
f = ( S F * N F + S E * N E + S NF * N NF + S NE * N NE ) / T (6)
ii. When enemy team gets a goal:
f = −( S F * N F + S E * N E + S NF * N NF + S NE * N NE ) / T (7)
iii. There is no team get a goal during time limit:
f = 0.5 * ( S F * N F + S E * N E + S NF * N NF + S NE * N NE ) / T
(8)
All the parameters used in our experiment is as shown in Table 3.
Table 4. Approximate averaging number of the evaluation of CENNs, Binary Genetic Algo-
rithms (Binary GAs) and ESP
In order to evaluate the evolution process and the result, we evolve the same structure
of Neural Networks using Binary Genetic Algorithms [13], ESP and CENNs. The
approximate number of the averaging result is as shown in Table 4.
Table 4 shows that CENNs is an appropriate system in evolving football game NPCs.
CENNs takes much fewer evaluations than Binary GAs and ESP, showing that in the
high dimensional optimization problems sometimes we need concurrent events to make
the evolution faster.
To prove its usability and efficiency, we have matched the untrained NEPlayers and
the trained NEPlayer against the pre-programmed players with FSM for 100 games. The
match result of untrained NEPlayers against the FSM Players is as shown in figure3.
Evolving Game NPCs Based on Concurrent Evolutionary Neural Networks 237
Fig. 3. The match result of untrained NEPlayers against the FSM Players
Fig. 4. The match result of trained NEPlayers against the FSM Players
Darwin Platform provides save and load NE Players which get the best fitness score
in the game (Fig. 5 shows the screenshot). The match result of trained NEPlayers with
the best fitness value against the FSM Players is as shown in Fig. 4.
As we can know from all the results above, CENNs not only can speed up the evo-
lutionary computation time, but also is an efficient algorithm with the adaptive fitness
function in video football game. Although enemy team becomes stronger, the NE-
Players trained by CENNs act cooperatively in team plays and win the games.
Acknowledgments
This research was supported by the ITRC (Information Technology Research Center,
MIC) program and Seoul R&BD program, Korea.
References
1. Rollings, A., Morris, D.: Game Architecture and Design (2000)
2. Kitano, H.: Designing neural networks using genetic algorithms with graph generation
system. Complex Systems 4, 461–476 (1990)
3. Koza, J.R., Rice, J.P.: Genetic generalization of both the weights and architecture for a
neural network. In: International Joint Conference on Neural Networks, vol. 2, pp. 397–404.
IEEE, New York (2000)
4. Liu, Y., Yao, X.: A population-based learning algorithm which learns both architectures and
weights of neural networks. Chinese Journal of Advanced Software Research 3(1) (1996)
5. Hrstka O, K.: A Search for optimization method on multidimensional real domains. In:
Contributions to Mechanics of Materials and Structures. CTU Reports, vol. 4, pp. 87–104.
Czech Technical University, Prague (2000)
6. Whitley, D., Starkweather, T., Bogart, C.: Genetic algorithms and neural networks: Opti-
mizing connections and connectivity. Parallel Computing 14, 347–361 (1990)
7. Yao, X., Liu, Y.: Evolving artificial neural networks. Proceedings of the IEEE 87(9) (Sep-
tember 1999)
8. Im, C.-S., Kim, T.Y., Um, S.-W., Baek, S.H.: Flexible Platform Archi-tecture for Devel-
oping Game NPC Intelligence with Load Sharing. In: The Workshop of 2005 International
Conference on Computational Intelligence and Security, Xi’an, China (2005)
9. Gomez, F., Miikkulainen, R.: Incremental evolution of complex general behavior. Adaptive
Behavior 5, 317–342 (1997)
10. Moriarty, D.E.: Symbiotic Evolution of Neural Networks in Sequential Decision Tasks.
Technical Report AI, Austin, pp.97–257 (1997)
Evolving Game NPCs Based on Concurrent Evolutionary Neural Networks 239
11. Yong, C.H., Miikkulainen, R.: Cooperative Coevolution of Multi-Agent Systems. Technical
Report AI, pp. 01–287. University of Texas at Austin (2001)
12. SADE, http://klobouk.fsv.cvut.cz/~ondra/sade/sade.html
13. Holland, J.H.: Adaptation in natural and artificial systems. Internal Report. University of
Michigan, Ann Arbor (1975)
14. García-Pedrajas, N., Ortiz-Boyer, D., Hervás-Martínez, C.: An alternative approach for
neural network evolution with a genetic algorithm: Crossover by combinatorial optimiza-
tion. Neural Networks 19(4), 514–528 (2006)
15. Stanley, K.O., Bryant, B.D., Miikkulainen, R.: Real-time neuroevolution in the NERO
video game. IEEE Transactions on Evolutionary Computation 9(6), 653–668 (2005)
16. Buckland, M.: AI Techniques for game programming (2004)
Knowledge Discovery by Network Visualization
{wuyc,zhouhong,pazuchan,huamin}@cs.ust.hk,
{xzm,lxm}@pku.edu.cn
Abstract. Hyperlinks among webpages are very important information and are
widely used for webpage clustering and webpage ranking. With the explosive
growth in the number of webpages available online, the exploration of hyperlinks
among webpages becomes a very challenging problem. Information visualiza-
tion provides an effective way of visualizing hyperlinks and can help users gain
insights into the relationships of webpages.
In this paper, we present some novel computer graphics techniques to visualize
the hyperlinks among webpages. We propose a visual encoding scheme for five
dimensional hyperlinks data and two constrained 3D layout techniques for incom-
ing and outgoing links of a single webpage. To reveal the hierarchical structure
of webpages as well as the hyperlinks information, we extend the treemap rep-
resentation. Our representations are visually appealing and can effectively reveal
linkage patterns among webpages. Experimental results and a user study demon-
strate the effectiveness of our system. Our system can facilitate E-learning and
help students understand the complex structures and hidden patterns in network
datasets.
1 Introduction
Hyperlinks among webpages are important and useful features of the Internet. They can
reveal interesting information about a webpage. For example, hyperlinks can be used
for clustering webpages because relevant webpages are usually connected via hyper-
links. In addition, hyperlinks can be used to rank webpages. Many hyperlinks pointing
to a webpage usually mean that this webpage is important. Recently, search engines
have been relying more and more on linkage data to determine webpage rankings. Hy-
perlinks are also widely used in data mining and other applications. Sometimes, users
want to manually explore hyperlinks data to reveal hidden linkage patterns associated
with a webpage or a network. Simply reading the results from database queries may be
tedious and inefficient. A picture is worth a thousand words. Computer graphics and
imaging techniques have thus been introduced to help users explore linkage data. With
the rapid growth in the number of webpages, visually exploring links among webpages
has become a critically important technique. The linkage patterns also pose special chal-
lenges for students. It is difficult to understand the various attributes of webpages and
their complicated relationships.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 240–251, 2008.
c Springer-Verlag Berlin Heidelberg 2008
Knowledge Discovery by Network Visualization 241
2 Related Work
Website Maps. There have been some works on website maps which visualize the
linking relations among webpages in that website. This can help users navigate and
search complex websites with the big picture of the website structure in mind and avoid
getting lost during browsing. WebTracer1 , which is being developed by Tom Betts, is a
tool for mapping the structure of websites. This freeware tool utilizes a 3D molecular
model visualization to show hyperlinks of a given site. Spheres are used to represent
webpages - the larger the sphere, the heavier the linkage to and from this page. The
red and blue edges represent links between pages. Tree Studio, developed by Inxight
Software (a spin off from Xerox PARC), provides a neat fisheye style interactive website
map. There are various notable works on visualizing the structures and the evolution of
websites and the Internet [1, 2, 3]. In this paper, we only focus on one important aspect
of the Internet, i.e., the hyperlinks among webpages.
1 http://www.nullpointer.co.uk/-/webtracer.htm
242 H. Zhou et al.
Links among Webpages and Network Visualization. The links among webpages are
usually represented as either a matrix or a node-link diagram. The node-link diagram
representation is more popular. Munzner [4] gave an excellent survey for network vi-
sualization. One of the earliest network visualization systems is SeeNet, which has had
profound influences on subsequent works. The simple node-and-link map is used, in
which the color and thickness of the lines represent the strength of the relationship while
the glyphs encode statistics data associated with the nodes. Cox et al. further extended
SeeNet to SeeNet3D [5] by exploiting more available space in 3D to decrease line in-
tersections and reduce visual clutter. Such simple graphical primitives were amenable
for slow computers with limited graphics capacity at that time. However, they have lim-
ited ability to encode multivariate information associated with network data and have
poor scalability for massive data. Visual metaphors and interaction are two main fac-
tors affecting visual scalability [6] and many studies have been conducted in these two
directions. Rafier and Curial [7] proposed an effective visualization method for large
networks through sampling.
Treemaps. Treemaps are very effective graphical representations for data with hierar-
chical structures. In treemaps, the size and color of the individual rectangles are sig-
nificant and can be used to encode data attributes. For example, if the tree represents
a file system hierarchy, the size may be proportional to the size of the respective file
and the color indicates the file type. The layout algorithms for the treemap and its var-
ious extensions have been thoroughly studied. An excellent survey can be found [8].
Some notable, variations of treemaps include Cushing treemaps, 3D treemaps [9, 10],
and Voronoi treemaps. An interesting technique has been introduced by Fekete [11] to
visualize a graph as a treemap with overlaid links.
Domains Domains
links
Cluster
Institutes Institutes
links
Cluster
Webhosts Webhosts
links
Cluster
Webpages Webpages
Hyperlinks
The webpages used in this experiment were collected and processed in the network
laboratory of our department. The webpages were first retrieved by crawlers and then
hyperlinks information was extracted. After analyzing the webpages and with the help
Knowledge Discovery by Network Visualization 243
Low High
Size Number of Webpages
.com
Center
.org Angle Domain
.net
.edu
To visualize linkage information for one single node, we can draw a simple graph or
node-link diagram where a node represents a webpage, a webhost, an institution, or a
domain and a link between two nodes indicates that they are connected via links. To
distinguish the node (i.e., a sphere or other graphics primitive) used in the node-link di-
agram and the node (e.g., a web host) in the linkage data, we use the graph node and the
web node, respectively, if ambiguity arises. In this section, we first introduce a graphical
representation which can encode four dimensional linkage data for a single node. After
244 H. Zhou et al.
that, we present two techniques that can dramatically improve the effectiveness of our
encoding scheme.
(a) (b)
Encoding all these information into one image can help users find the possible corre-
lations between different attributes of nodes and links. We design the following visual
encoding scheme for five dimensional linkage data:
– The size of a web node is encoded using the size of the corresponding graph node.
– The number of outgoing links from the central web node to another web node is
encoded using the distance between them. The larger the number of links, the closer
the two nodes. The distance can naturally represent the strength of the relationship
between two nodes.
– The number of all incoming links of a web node is encoded using the color of the
corresponding graph node.
– The incoming links to the central node from an outside node are represented by
the line between the two node. If there is no outgoing link from a web node to the
central web node, no line will be drawn between them. The color of the lines/links
encodes the number of outgoing links from the web node to the central node.
– Nodes are clustered according to their domains. For example, if two webpages
belong to the same domain, they will cluster together. We use a pie-style graph
(See Figure 3b) to show the percentage of each domain among all domains.
Figure 2 shows our visual encoding scheme. Figure 3 shows all the nodes connected
to one single node via links and the associated statistics information using our 4D visual
encoding scheme.
Knowledge Discovery by Network Visualization 245
(a) (b)
Fig. 4. Distance histogram equalization: (a) before equalization; (b) after equalization
Fig. 5. Constrained 3D layouts: (a) Original layout; (b) Cylindrical layout; (c) Semi-spherical or
dome layout; (d) Transparent dome layout
be re-positioned on the surface of the semi-sphere or cylinder. Figure 5 shows the two
3D constrained layouts. Users can switch between cylindrical view and semi-spherical
view. The transparency of the cylindrical surfaces and semi-spherical surfaces can also
be adjusted. These layouts have the following advantages:
1. There is more space to layout the nodes. The nodes can be positioned onto a 2D
surface instead of a 1D circle.
2. It is natural for users to understand that the nodes on the semi-spherical or cylindri-
cal surface have a similar distance to the central node.
3. The constrained 3D layout can overcome some of the disadvantages of other gen-
eral 3D layouts. One major disadvantage of using 3D layouts is the visibility and
occlusion problem which may be confusing for some users. However, our con-
strained 3D layouts have the advantage that more space can be used but no serious
occlusion problems will be caused.
As mentioned earlier, webpages can be naturally organized into a hierarchy. This hi-
erarchical structure may be useful for many applications. To display this hierarchical
structure along with linkage information, we exploit the treemap representation. The
treemap is a classic visualization technique to show data with hierarchical structures.
Usually, treemaps can use the following features to encode three dimensional data into
one image: the hierarchical structure, the size, and the color of the boxes.
In our example, we have more than three dimensional data. To encode the hyperlinks
information with the associated multi-variate statistics data, we introduce a 3D treemap
representation. Our method first shows a 2D traditional treemap. The hierarchical struc-
ture of boxes encodes the hierarchical structure of webpages. Users can then choose to
use the size of the boxes and the color of the boxes to encode any two of the following
four node attributes: size of this node, number of self links, number of outgoing links,
or number of incoming links. Users can zoom in/zoom out the treemap and click on any
box then the upper level boxes will be highlighted and their associated node information
will be shown on a separate message window. In addition, after a user clicks on a box
which represents a node, all other nodes connected with this node will be drawn using
Knowledge Discovery by Network Visualization 247
3D boxes, where the height of the boxes can encode another statistic data attribute (see
Figure 6b).
Compared with traditional treemap representations, our 3D treemap can encode two
more node attributes, i.e., the relationship of nodes represented as the dimension of
boxes (3D vs. 2D) and another attribute represented as the height of the box. The 3D
treemap has been used before to visualize file system hierarchies by Bladh et al. [9].
They used the height of the box to encode the depth in the file tree. To the best of our
knowledge, it is the first time that the 3D treemap is used to encode linkage data. Figure
6 compares the traditional 2D treemap and our 3D treemap.
(a) (b)
Fig. 6. Treemap for linkage data: (a) 2D treemap encoding 3D data; (b) 3D treemap encoding 4D
data
6 User Study
A real system has been developed based on our methods to visualize the linkage data
we collected. The system is installed on a Pentium(R) 3.2GHz PC with 1GB RAM and
a 256MB Nvidia Geforce 6800 Ultra graphics card. MySQL is used to handle queries
from users. To test the effectiveness of our encoding schemes and the usability of our
system, we conducted a user study.
6.1 Procedures
The data we used consists of all the webpages in the “edu.cn” domain collected at
Peking University. We obtained about 820 Chinese educational institutions and 4,333
webhosts. Ten students with normal color vision, which consist of five undergraduate
students and five post-graduate students in the computer science department of Hong
Kong University of Science and Technology, participated in our user study. All are ex-
perienced computer users but have no prior experience with our system or any similar
information visualization systems. Before starting the user study, each user received a
fifteen-minute training about the visualization and interaction techniques of our system.
They were also provided a diagram similar to Figure 2 , which shows our visual encod-
ing scheme. After the training, each participant was asked to perform four tasks. The
first task was designed to test the usability of our system. The participants were asked to
do conjunctive queries using our system. The other three tasks were designed to test the
248 H. Zhou et al.
effectiveness of our visual encoding schemes. Ten subjects are divided into two groups.
These two groups then explored the data using different encoding schemes (i.e., layouts
with and without distance histogram equalization; 2D layout v.s. 3D constrained layout;
straight-line layout v.s. curve layout). The four tasks procedures are listed as follows.
Task 1. We chose one specific webhost (www.pku.edu.cn) and then asked each sub-
ject to find out one or more nodes which are connected to this webhost and have : 1) a
large number of webpages; 2) a large number of incoming links; and 3) relatively weak
connection to www.pku.edu.cn. It is used to test the encoding scheme in Section 4.1.
After finishing this task, each subject was asked about their feelings.
Task 2 & 3. We randomly divided the 10 subjects into two groups of five, to test our
distance histogram equalization and constrained 3D layouts techniques.
In Task 2, the systems with and without distance histogram equalization were used
by group A and group B respectively. Both groups were asked to perform a task sim-
ilar to the first task, but the chosen webhost www.pku.edu.cn was changed to webhost
www.edu.cn, and the three conditions were also changed slightly.
In Task 3, group A was asked to use only the 2D layouts, while group B was asked to
use the 3D layouts. Distance histogram equalization was used by both groups. The ques-
tion was still the same, but the chosen webhost was changed to www.tsinghua.edu.cn
and the three conditions were changed slightly again.
After finishing task 2 and 3, each participant was asked about their experiences of
answering these two questions.
After finishing all the three tasks, each subject was asked to play with our system for
at least ten minutes. We observed their usage of our system and interviewed them about
their experience and opinions after their exploration.
F 1
F: www.phil.pku.edu.cn
E 2
B: www.tsinghua.edu.cn
Webhost
A: www.edu.cn
D 3
D: www.cau.edu.cn
C 1
C: e.pku.edu.cn
B 2
E: bbs.pku.edu.cn
A 9
0 2 4 6 8 10
Number of persons
www.pku.edu.cn
(a) (b)
Fig. 7. Results for Task 1: (a) The linkage pattern for www.pku.edu.cn and the six webhosts chosen
by the users; (b) The distribution of the user answers
Basically our system is good at those subjective tasks because users can easily get the
overview of all the attribute values simultaneously. To formulate these three conditions
using conjunctive queries in SQL is difficult because the large, hot and strong connec-
tions are relative and highly depend on the context. It is hard to decide the thresholds
for the SQL queries. The results are also subjective and need human judgment. But this
information is important for knowledge discovery and may reveal some patterns about
the web data. A good solution is to show the overall distributions of all the attributes
together and then let users to choose the answer using their own judgment. Our multi-
variate data encoding scheme can facilitate this process and help users for knowledge
discovery.
G 1 G 2
F 1
K 1
Webhost
Webhost
E 1
D 1 J 1
C 2
I 2
B 1
A 2 H 1
0 0.5 1 1.5 2 2.5 0 0.5 1 1.5 2 2.5
Number of persons Number of persons
(a) (b)
Fig. 8. Results for Task 2: (a) Result from group A. The average response time is 39.6 seconds;
(b) Result from group B. The average response time is 21 seconds.
Task 2 & 3. For Task 2, the results are showed in Figure 8. The average response time
of group A is about two times longer than that of group B. Only webhost G (See Figure
8) was pointed out by both groups, and more webhosts are pointed out by group A than
by group B. From Figure 4, we can see that with the distance histogram equalization
the usage of space is dramatically improved. Thus, the users can easily shrink the range
of candidates and find out the final results faster after the equalization. In addition, the
answers may be more reasonable and accurate.
For Task 3, the results are showed in Figure 9. The average response time of group B
is about three times longer than that of group A, because the subjects in group B spent
more time on 3D rotation and exploration. After exploring the constrained 3D layouts,
each subject of group B chose only one webhost while two group A members picked
out multiple choices.
For the question about their experiences of answering these two questions, two group
A members disliked the layout (i.e., without distance histogram equalization) in task 2.
They complained that too many nodes were clustered together. One group B member
said that the cylindrical layout was very cool and really helpful. However, another per-
son thought that using the 2D layouts was easier than using the 3D layouts to find the
answer because of the longer time spent on the rotation.
Most participants mainly used the visualization scheme presented in Section 4.1 to
explore the data. They are fascinated by our system. When playing with our system,
most users can find some interesting information and some emitted comments like “Oh,
my favorite college is so blue!” and “Why don’t these two universities have many links
between them?” They were usually attracted to some noticeable but unknown web-
hosts. Some of them even opened the web browser to take a look at these webhosts.
250 H. Zhou et al.
F 1 F 1
E 1
E 1
Webhost
Webhost
D 1
I 1
C 2
H 1
B 1
A 1 G 1
0 0.5 1 1.5 2 2.5 0 0.5 1 1.5
Number of persons Number of persons
(a) (b)
Fig. 9. Results for Task 3: (a) Result from group A. The average response time is 36.2 seconds;
(b) Result from group B. The average response time is 86 seconds.
One example is a Chinese educational portal webhost (i.e., www.edu.cn), which was
noticed by all users, because this webhost is very big, red, and nearly all the Chinese
educational institutions are linked to it. Another webhost discovered by most subjects
is www.gaokao.edu.cn, which is a very large and blue webhost. Most subjects are not
aware of the existence of these webhosts before the user study. Some unexpected link-
age patterns were also found out by our users. For example, someone commented that
the big and red nodes are usually far away from the central web node. For example,
in Figure 5a, 7a, and 3a, many big and red nodes are not very close to the central web
node. We found that this is also a common fact for many other webhosts. This is an
interesting finding, which means, among the Chinese education institutions, many web-
hosts are connected to some big and hot webhosts but their connection strengths are not
as high as we expected. We were surprised by their findings because as developers we
did not notice these patterns before.
The 3D treemap display was also played by some participants. They liked to change
the encoding scheme of treemap frequently and click those attractive boxes to see what
they are. One user pointed out that nearly all the institutions contain a BBS webhost
(e.g., The University of Science and Technology of China has bbs.ustc.edu.cn) which
has a large number of webpages and many incoming links.
7 Conclusion
In this paper, we introduced some novel computer graphics techniques for visualiz-
ing hyperlinks among webpages. We proposed a 4D encoding scheme and constrained
3D layouts to visualize incoming/outgoing links for a single webhost. We further ex-
tended the treemap representation so the links information can be shown along with the
hierarchical structure of webpages. A user study was conducted to demonstrate the ef-
fectiveness of our encoding schemes and the usability of our system. Some unexpected
patterns have been detected by students using our system.
Our current work only focuses on effective visualization techniques for medium size
data. With the improvement of data collection techniques, more data such as the click
stream for hyperlinks and history of webpages will be available for users to explore.
How to visualize large data and encode more than five dimensional data is worth further
study. We also plan to extend our methods for general network visualization.
Knowledge Discovery by Network Visualization 251
Acknowledgments
This work was supported by Hong Kong RGC grant CERG 618706 and NSFC grant
60573166.
References
1. Cheswick, B., Burch, H., Branigan, S.: Mapping and visualizing the internet. USENIX, pp.
1–12 (2000)
2. Chi, E.H., Pitkow, J., Mackinlay, J., Pirolli, P., Gossweiler, R., Card, S.K.: Visualizing the
evolution of web ecologies. In: Proceedings of the SIGCHI conference on Human factors in
computing systems, pp. 400–407 (1998)
3. Chi, E.H., Card, S.K.: Sensemaking of evolving web sites using visualization spreadsheets.
In: Proceedings of the 1999 IEEE Symposium on Information Visualization, pp. 18–25
(1999)
4. Munzner, T.: Interactive visualization of large graphs and networks. Ph.D. dissertation, Stan-
ford University (2000)
5. Cox, K., Eick, S., He, T.: 3d geographics network displays. SIGMOD Record 25(4), 50–54
(1996)
6. Eick, S., Karr, A.: Visual scalability. Journal of Computational and Graphical Statistics 11(1),
22–43 (2002)
7. Rafiei, D., Curial, S.: Effectively visualizing large networks through sampling. IEEE Visual-
ization, 48–55 (2005)
8. Shneiderman, B.: Treemaps for space-constrained visualization of hierarchies,
http://www.cs.umd.edu/hcil/treemap-history/
9. Bladh, T., Carr, D., Scholl, J.: Extending tree-maps to three dimensions: A comparative study.
In: Proceedings of 6th Asia Pacific Conference on Computer Human Interaction, pp. 50–59
(2004)
10. Tanaka, Y., Okada, Y., Niijima, K.: Treecube: Visualization tool for browsing 3d multimedia
data. In: Proceedings of Information Visualization 2003, pp. 427–432 (2003)
11. Fekete, J.D.: The infovis toolkit. In: IEEE Symposium on Information Visualization, pp.
167–174 (2004)
Research on Emotional Vocabulary-Driven
Personalized Music Retrieval
1 Introduction
With the rapid development of multimedia and in-depth popularity of Internet tech-
nologies, the music information is growing rapidly. How effective retrieval for music
is becoming an important research area in modern information retrieval. A typical
traditional practice is to use text formats such as names of songs, singers or lyrics
keywords, as users must remember that relevant information. However music is
wildly different with text data, the user's needs are often based on the music semantics
as melodies, form, style and emotion, and so on. In the means of semantic-based re-
trieval, the user only needs to present the musical content of a semantic description,
for example: some music samples, such as a user playing or singing the melody repeat
[1], a school or style ( classical, rock, etc.), or some kind of emotional acoustic sens-
ing features (sorrow, joy, etc.).
Different retrieval conditions of semantic information are corresponding to differ-
ent music retrieval technology. After the pioneering papers [1] in ACM multimedia
seminar, a considerable amount of research on “Query by Humming” retrieval system
is reported, which adopts a more natural way of interaction, so that users only need to
remember songs melody.
In the study of music retrieval based on the style or emotion, the premise work
is organizing the music in semantic information. Island of Music System [2] intro-
duces a method based on the similarity of perception and reduced-dimension to ex-
tract rhythm patterns, use self-organizing map (SOM) data analysis methods for data
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 252–261, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Research on Emotional Vocabulary-Driven Personalized Music Retrieval 253
classification, and then use smoothing data histogram (SDH) for visualization and
browsing. This system is particularly suited to the music retrieval requests such as
"sounds like how” and "what emotional color with the works". However it is a pity
that the system itself does not provide the retrieval application.
In article [3] a pioneering work in emotion-driven music retrieval is presented,
which implements the retrieval mode in 200 pieces of MIDI music database based on
the binary encoding interactive genetic algorithm, whose genetic operating is evolved
in the music characteristic space, the user's choice of emotional needs is integrated
into the interactive process. However, in the retrieval process, users’ fatigues are
occurred easily, and the emotional expression model is also not match with the psy-
chology research.
Juslin’s researches [4] on music and its emotion connotation is much representa-
tive, but there is little discussion about music retrieval based on emotional informa-
tion. This paper is studying on the emotional retrieval method of music, which adopts
emotional vocabulary as input conditions.
8 1
Vigorous Dignified
7 2
Exciting Sad
6 3
Joyous Dreamy
5 4
Graceful Soothing
For instance, after some reasoning process, the emotional connotation of a certain
music M can be expressed as follows: (0.2, 0.6, 0.9, 0.4, 0.3, 0.1, 0.0, 0.1), and so its
dominant emotion is "longing", with emotional semantic similarity value is 0.9, while
the music also includes a "sad" emotional connotation, with a little lower value as 0.6,
which can be termed as the secondary dominant emotion. Anyway, the emotional
connotation of music M can be described as "very longing, and some sadness".
The preliminary work of music retrieval is automatic identification and emotional
machinery tagging for music clips database. Figure 2 shows the emotional music of
tagging and database building process [6]. First, we build the music clips database
with independent emotional semantic through music feature recognition, and using
the theme melody to represent the music as abstract. And then we tag music clips with
emotion vector based on the IGEP algorithm, which achieves mapping from a 10-
dimensional feature space to the 8-dimensional emotional space.
Key
Key
Melody
Melody
Recognition
Emotion
vector
help of outside advice. In all, the needs of user are uncertain, as we call this mode of
retrieval as the personalized retrieval. As the communication among peoples, only
interactive process helps to mutual understanding and tacit agreement. As shown in
figure 3, interactive genetic algorithm (IGA) require user to evaluate the satisfactory
degree of the newly- evolutional individual through the interactive process, which is
used to measure its distance with psychological space, and then the system performs
subsequent operation and produce a new generation of groups by the corresponding
fitness function, and constantly repeat to get an optimize individual meeting the user's
requirements. The purpose of interactive process is to integrate the human wisdom
and computer technology for improving the performance of genetic algorithms. Since
the preference, intuitive, emotional and other psychological factors of users can be
identified as a fitness value of the characteristics that can be binding to the target
system, the IGA algorithm is also applied in the music retrieval [3], [7].
Start
Genetic operation
Y
New-generation
End
The same as the genetic algorithm, IGA also requires fewer individuals of each
group and more minimized evolutionary generation as possible, in order to speed up
the retrieval speed because of music’s timing characteristic, with which a person's
memory is limited therefore the presented individual must be restricted. According to
the same reasoning, evolution generation must also be limited.
The most serious problem of IGA is how to reduce fatigue in the interactive pro-
gress. Since people should communicate with computer, giving assessed value of each
individual in each generation, larger the number of individuals, longer evolutionary
time, more easily result in fatigue, especially in the case of facing these individuals
with little difference and difficulty to distinguish, which will induce greater psycho-
logical pressure and thus more easily produce fatigue.
256 B. Zhu and T. Liu
In IGA for emotional music retrieval, not only the specific characteristics of emo-
tional linguistic value of music but also some specific genetic operators should be
taken into account. Since music clips database has been established through the emo-
tional tagging process, the supervenient music retrieval can be performed directly in
the music emotional space.
By now, the commonly used method of coding is the binary-coded and Real-coded.
According to the definition of emotion vector, this thesis is based on the real number
of eight dimensional vector, ranging between [0, 1], so the higher accuracy is required
(5 location after decimal point). If the binary code is adapted, its length of solution is
136
8 * 17 = 136, the facing search space is 2 , and that is very big. In addition, the use
of Real-coded approach can also help development of the specialized genetic operator
for music retrieval. Therefore, this paper uses Real-coded method [8].
Similar with the definition of emotion vector, style of gene series
Xi (i = 1, 2,...N )
is expressed as
(ai1, ai 2,..., ain )
, in which we have
ak ∈ D(ak ) ⊂ R (k = 1, 2,..., n) ,
n = 8 for the dimension of space, and N for the population size (this paper evaluated
as 8).
For the average retrieval process, the selection of first generation is very important.
The retrieval process will be speed up when the music has certain characteristics con-
sistent with the user’s needs. For the greatest possible diversity of the sample, the
system chose one from eight pieces of music whose value of similarity with dominant
emotion is biggest as the first generation sample groups. Those eight pieces of music
would certainly include the characteristics of user’s demand, so it is bound to increase
the efficiency of retrieval, but also diversity of samples can be insured.
Genetic operations play vital role in IGA, which achieve the survival of the fittest of
the evolutionary process by genetic operators (such as choice operator, crossover
operator and mutation operator, etc.) and the degree of adaptation to the environment.
1. Choice operator for fine breed
To ensure individual with greater fitness will be able to be retained in the next gen-
eration, while accelerating the process of the convergence of GA, the evaluation of
music achieved by users will be cognized as the criterion for choice to choose 8 sub-
chromosomes. The calculation process is as follows:
(1). Consider numerical evaluation of the users as the coefficients corresponding to
emotional vector of each individual, and calculate the reference emotional vector
according to multi-source information fusion method based on music emotional lin-
guistic value model [5];
(2). Calculate the similarity between the emotion vector of each music individual
sample and reference emotional vector, and evaluate fitness value fi of this individual
i, and arrange those individuals by descending of fitness values;
Research on Emotional Vocabulary-Driven Personalized Music Retrieval 257
∑f
i =1
i
ing the general assumptions, supposed the last k component is the smaller, and then
two child generations will be defined as:
sz = (v1(1) ,..., vk (1) , α k +1vk (1) + (1 − α k +1 )vk (2) ,..., α n vn (1) + (1 − α n )vn (2) ) (4)
sw = (v1(2) ,..., vk (2) , α k +1vk (2) + (1 − α k +1 )vk (1) ,..., α n vn (2) + (1 − α n )vn (1) ) (5)
The parameter
t = 0,...,3,...
notes that the recent evolution number.
In addition, similarity between every two reference vector can be calculated
as Sik , i, k = 1,..., 4 and i ≠ k , then the general similarity metric of all reference
4
Si = ∑
k =1, k ≠ i
Sik
vector is . By using these four similarities as counterpart contribution
on the curve interpolation corresponding to every vertex, we get the matrix expression
of the curve:
⎡ 1 0 0 0 ⎤
⎢ 11 3 1 ⎥⎡S X ⎤
⎢− 1 − ⎥ 1 1
⎢ 6 2 3 ⎥ ⎢S X ⎥
1⎥ ⎢
2 2⎥
P = P (t ) = [1 t t 2 t3 ] ⎢ 1 (6)
⎢ 1 2 − ⎥ ⎢ S3 X 3 ⎥
⎢ 2 2⎥ ⎢ ⎥
1 ⎥⎣ 4 4⎦
S X
⎢ 1 1 1
⎢− − − ⎥
⎣ 6 2 2 6 ⎦
And then the interpolation curve can be calculated by using the four reference vec-
tor to predict the next generation (t = 4) emotional reference vector. After the fitness
value of new generation are calculated by system instead of user’s interaction accord-
ing to the formula above, the autonomous GA process will be restart.
After a series of genetic operation, the chromosome of group has been changed for
some new type of chromosome produced. Since limitations of storage capacity of
music clips database results that there may exit no be new corresponding items of
chromosome in the database. This paper produces a new generation by recently
neighbors Law.
Since evolutionary termination conditions have sth to with the quality of the solu-
tion, if we adopt a pre-determined the number of generation as a termination condi-
tions, the complexity and convergence of the algorithm are difficult to control because
different users have different cognitive ability, so it will be very difficult set a reason-
able number of generation. In this paper, when the majority of music clips which
users feel satisfied (generally preferable for 5) exit, the system identify that users have
found that the needs of music and terminate operations.
Research on Emotional Vocabulary-Driven Personalized Music Retrieval 259
3.4 Experiment
This method does not require users to have a prior clear impression of goal music and
the user only has to select and score the music represented. In the interactive process,
users can change the target music demands at any time. The whole process is de-
scribed as follows:
A) Produce the initial individual groups automatically;
B) Evaluate the individual sample according to the relevance of the goal subject:
users choose some musicians (above two) to positive or negative assessment and the
fitness values of each individual will be calculated;
C) Check if the termination conditions of evolution are meeting: N to D), Y to the
end;
D) Genetic operate on the group according to the fitness values calculate by step
B): implement choice, crossover and mutation operation, and produce a new genera-
tion of individuals. The crossover probability defaults to 0.3, which users can change
at any time;
E) Check the user fatigue and the number of generation, and if users feel fatigue
and the number of generation has more than four, system goes to step F), the self-
genetic processes, or to B);
E) Calculate the fitness value of reference vector according to the records of out-
standing individual evaluated by users;
F) Calculate the fitness values of each individual of the generation and go to
step C).
In the algorithm, with the deepening of the process of evolution, the target music
chosen by the user should be increasingly clear. And so we can have better judge
rules, and get more satisfactory results for users.
In our experiment, every generation provided eight music clips to users, and users
give the evaluation of satisfaction about every song based on his or her audition.
We invite six boys and six girls to our experiment and divided them into six groups
to test the algorithm contained absolute of entire operators designed above in a data-
base including 1,190 musicians, and the goal music emotional linguistic words are
ILWWQHVVYDOXH
(IIHFWRQHIILFLHQF\RIUHWULHYDO
&KRLFH2SHUDWRU
&URVVRYHU2SHUDWRU
6HOIJHQHWLF2SHUDWRU
&RPSUHKHQVLYH
([SHULPHQW
1RRSHUDWRUVXVHG
QXPEHURIJHQHUDWLRQ
"romantic", "fantasy" and "sanctity", "dark", "anger" and "surprised". If the number of
generation reaches 15, the retrieval is considered failure.
Figure 4 gives the average number of generation which users find the satisfactory
music. These experiments show that the genetic operators designed above can effec-
tively accelerate the convergence rate, so that we can better reduce the user’s fatigue.
From the experimental results, the ratio of music clips in each generation meeting
with the needs of user can be gradually increased through the interactive process. In
all operators, the contribution of choice operator for the improvement of genetic con-
vergence is the most notable. And under the operators synthetically, algorithm can
effectively accelerate convergence, and find the music matching the emotional needs
of the users.
4 Conclusions
In this paper, several major works are as follows:
(1) To identify and tag for emotional connotation of music clips based on linguistic
value model of music emotion automatically;
(2) To design an improved interactive genetic algorithm against personalized re-
trieval demand, and emotional language-driven music retrieval;
(3) To design a series of strategies in order to effectively accelerate convergence
and eliminate user’s fatigue produced in the interactive progress, such as: choice op-
erator for fine breed, crossover operator holding the dominant element of emotion
vector, and self-genetic operation strategy.
References
1. Ghias, A., et al.: Query by humming: Musical information retrieval in an audio database. In:
Proc. of ACM Multimedia 1995, November 1995, pp. 231–236 (1995)
2. Pampalk, E.: Islands of music analysis, organization, and visualization of music archives.
OGAI Journal (Oesterreichische Gesellschaft fuer Artificial Intelligence) 22(4), 20–23
(2003)
3. Cho, S.-B.: Emotional image and musical information retrieval with interactive genetic al-
gorithm. Proceedings of the IEEE 92(4), 702–711 (2004)
4. Juslin, P.N., Friberg, A., Bresin, R.: Toward a computational model of expression in music
performance: The GERM model. Musicae Scientiae, Special Issue 2001-2002, pp. 63–122
(2002)
5. Sun, S., Wang, X., Liu, T., Tang, Y.: Study on linguistic computing for music emotion (in
Chinese). Journal of Beijing University of Posts and Telecommunications 29 (z2), 35–40
(2006)
Research on Emotional Vocabulary-Driven Personalized Music Retrieval 261
6. Yang, C., Sun, S., Zhang, K., Liu, T.: Study on music emotion cognition model based on
applying the improved gene expression programming. Journal of computational information
systems (to be accepted)
7. Wang, S., Chen, E., Wang, S., Wang, X.: A Kansei Image Retrieval Based on Emotion
Model (in Chinese). Journal of Circuits and Systems (6)(2003)
8. Herrera, F., Lozano, M., Verdegay, J.L.: Tackling Real-Coded Genetic Algorithms: Opera-
tors and tools for the Behaviour Analysis. Artificial Intelligence Review (12), 265–319
(1998)
Research on Update Service in Learning Resources
Management System
1 Introduction
The construction of learning resources is a systems engineering which includes design,
development, usage, management and evaluation. It is not a one-step but a circular and
dynamic process of lack-supply-balance-lack-supply. Resources update service is a
necessary and significant service in learning recourses management system [1]. Since
new resources are accumulating continuously and users are expanding rapidly, an
obvious problem appears that users urgently ask for new resources while new resources
are not timely available to the user. There are two resources update patterns at the
present. One is the new resources disks delivered by people or mailed to users; the other
is a modern update pattern that the new resources are published on Web, and users can
download them. The former requires a long update cycle and wastes lots of manpower,
material resources and finances. As the construction of informatization foundation has
been consummating and Internet has been widely used, more researches focus on the
latter [2].
In order to realize resources share effectively and resources update in time, we
designed a Distributed Push System of Learning Resources, called DPSLR. In this
system, we use ACO to resolve the schedule between servers and users. It has been
*
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 262–269, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Research on Update Service in Learning Resources Management System 263
proved that DPSLR was an effective Internet-based update service and solved disad-
vantages of traditional update pattern.
2 System Design
Fig. 1 shows the framework of DPSLR. In this system, there are several equal and
distributed servers locating in different place. Every server can support update service
for users independently under the control of the schedule server. Users and distributed
servers are both managed by the schedule server. It is fit for global system control. In
addition, the system has expansibility and fault-tolerant capability. When a new server
is added, it is only need to register its information in the schedule server, and then the
new server can work. When a server breaks down, the schedule server will automati-
cally detect the error and collaborate with the users’ requests again.
3 Schedule Server
The Schedule Server is composed of task receiver, task scheduler, task dispatcher, task
inspector, original task queen, optimized task queen, task log and database. Its
framework is as Fig. 2.
The working flows of schedule server are as follows:
(1) The task receiver receives users’ requests and collects related information to
form the original task queen.
(2) The task scheduler re-arranged to form the optimized task queen using some
optimization algorithm according to the original task queen and database.
(3) The task dispatcher assigns the related distributed servers to respond to users’
requests according to the optimized task queen.
(4) The task inspector adjusts schedule strategy and re-optimizes task queen
according to task log.
From a macroscopic viewpoint, the schedule server balances the load of every dis-
tributed server to avoid overload. From a microscopic viewpoint, optimization algo-
rithm is used to design the most reasonable schedule scheme between distributed
servers and users. In this way, the performance of the system will be near to the best.
The pre-alarming mechanism is designed for exceptions. The task inspector can watch
the working state of the system. When exceptions happen, the urgent process program
will be automatically triggered to protect the system.
Research on Update Service in Learning Resources Management System 265
(2)
Where, allowed k represents a list of the users which can be accessed by ant k. Vis-
ited users will be eliminated from the list during each phase. τ ij (t ) represents the
pheromone density between u i and s j at the certain moment t. ηij represents the
heuristic degree of moving from s j to u i . α represents the weight of pheromone
density and β represents the weight of heuristic degree. At the beginning, the
pheromone densities of all paths are equal and τ ij (0) = C (C is a constant). ηij is
determined by some heuristic algorithm. In this paper, ηij is as follows:
(3)
The deposit of pheromone:
(4)
Where, ρ represents evaporation degree of pheromone. Q is a constant.
Steps of the algorithm:
(1) Initialize the value of ρ ,Q,C, α , β .
(2) ocÅoc represents outside counter).
Research on Update Service in Learning Resources Management System 267
4 Experiment
To evaluate the efficiency and validity of schedule algorithm, we have done a com-
petitive experiment. The data are as follows:
m=3, n=20, Max =Max =6, Max =8, all users’ tasks are the same and the size of
s1 s2 s3
tasks is 50M.
The maximum of all users connected to servers is 1. The download velocities be-
tween users and servers are limited from 100k/s to 900k/s and randomized. In addition,
constants in ACO are initialized as follows:
α =2, β =2, ρ =0.1, Q=1, C=0.1. Through 200 generations, we got the values of
optimization function E as Fig. 3.
th
From Fig. 3, it shows the better result is obtained at the 106 generation. The value
of optimization function E is 1243.5s. In order to prove the optimal performance of
ACO, we compared with traditional load-balance algorithms in dealing with the
schedule between distributed servers and users, such as cycle, random, the least re-
sponse time and cycle with weight [6]. The comparison result of ACO and traditional
algorithms is showed in Fig. 4.
From Fig. 4, it shows schedule based on ACO uses fewer time than FIFO obviously, not
only in the download time of all tasks but also in solitude download time of many tasks.
Acknowledgment
Our research is supported the National Torch Key Program Foundation of China.
Research on Update Service in Learning Resources Management System 269
References
1. Seels, B.B., Bichy, R.: Instructional Technology: The Definition and Domains of the Field.
In: Association for Educational Communications and Technology, Washington, DC (1994)
2. Pei-jun, D., Dong-dai, Z., Xiao-chun, C.: The Research on Individual Service Model of
Education Resources Based on Multi-Agent Technology. Journal of Northeast Normal
University(Natural Science Edition) 38, 31–35 (2006)
3. Yi-xing, Y., Yang, G.: Research on Resource Scheduling in the Grid. Application Research
of Computers 5, 23–26 (2005)
4. Abido, M.A.: Multiobjective Evolutionary Algorithms for Electric Power Dispatch Problem.
IEEE Transactions on Evolutionary Computation 10, 315–329 (2006)
5. Dorigo, M., Stutzle, T.: Ant Colony Optimization. MIT Press, Cambridge (2004)
6. Dewei, P., Yanxiang, H.: Study of Application on Framework of Load Balancing Based on
Agent. Computer Engineering and Applications 5, 153–156 (2005)
On Retrieval of Flash Animations Based on Visual
Features
1 Introduction
Flash, as a vector-based media format, is widely used for cartoon, music TV, game,
VR on desktop, commercial advertisement, e-postcard, etc, with the advantages of
small size, easy composition, bright dynamic effect, powerful interactivity, etc[1].
Since its advent in 1997, Flash animation has experienced an explosive growth and
become one of the most prevalent formats on the Web. The retrieval, management
and utilization of Flash resources become the interesting issues in IR community.
Nowadays, some multimedia search engines offer Flash retrieval on keywords. How-
ever, it is difficult to describe its contents and visual features using two or three
keywords due to the complicated content and the visual effect of Flash animation.
Content-based multimedia retrieval (CBMR) has been developed since 1990s [2].
Unfortunately, the research efforts on CBMR are mainly aiming at the retrieval of
image, video and audio, and there are quite limited research papers involved in con-
tent-based retrieval and management of Flash resources. Yang et al. [3] proposed a
generic framework named FLAME (Flash animation Access and Management Envi-
ronment) through the analysis of content structure, which embodies a three-tiered
architecture for the retrieval of Flash animations. The FLAME has implemented the
retrieval at different levels of details, including object level, event level and interac-
tion level. Ding et al. [4] suggested a semantic model with co-occurrence analysis
for improving the performance of retrieval of Flash. No other relevant research on
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 270–277, 2008.
© Springer-Verlag Berlin Heidelberg 2008
On Retrieval of Flash Animations Based on Visual Features 271
content-based index and retrieval of Flash animation has been reported (to the best of
our knowledge).
Flash animation is a kind of streaming media with heterogeneous components (in-
cluding texts, graphics, images, videos and sounds), dynamic effects and interactions.
On the view of movie playing, a Flash animation is composed of a series of scenes
like a video stream, therefore it may be presented and indexed similar to video based
on its shots and scenes. The approaches proposed by [3], [4] have described compre-
hensively the content structure of Flash animation overlooking the visual effects and
scene structure. On the view of visual effects of Flash animation, Flash animation is
segmented into a series of scenes based on the visual effects showing on screen, and
the key visual features of scenes are extracted to represent Flash animation based on
its scene structure. A Gif animation image is constructed with the typical frames (one
extracted from every scene) to be used as a summary displaying the main playing
screen of Flash animation. Finally, a system of Flash animation retrieval and quick
browsing is realized, based on scene structure and visual features.
The rest of the paper is organized as follows. In Section 2, the structure of scenes
of a Flash animation is analyzed and a 3-layer architecture to represent the content of
Flash animation is presented. The method of keyframe extraction and scene segmen-
tation of a Flash animation is elaborated in Section 3. The retrieval of Flash anima-
tions based on visual features of scenes is described in Section 4. The conclusion is
given and promising future directions are suggested in Section 5.
Fig. 1. The architecture of scenes in a Flash animation, ○ means the typical frames
272 X. Meng and L. Liu
former is the frame in which some objects, object properties, changes, actions or in-
teractions are set or defined by the producers. The later is generated automatically
through interpolation calculation with the help of the before and after keyframes by
Flash animation making tools. A sequence of frames which can be automatically
played in Flash animation is called a segment. All segments are arranged on the time-
line of production and they are linked through the jump scripts in keyframes on hy-
perlink structure of Flash animation, as illustrated in Fig.1 (upper part). The hyperlink
structure of Flash animation indicates the relationship between segments, but it cannot
reveal the visual effect of scenes in Flash animation. In a Flash animation, the visual
features may sometimes vary greatly in a segment, while sometimes they may vary
less among several neighboring segments. On the view of visual effects, the frame
sequence on the production time-line is segmented to a series of scenes, each of which
is composed of the contiguous frame sequences with similar visual features (looking
similar on visual effects). Thus a Flash animation can be viewed as a movie composed
of a series of scenes with different visual features, as illustrated in Fig. 1 (lower part).
On the basis of scene segmentation and scene analysis, combined with extraction
of meta-information of Flash animation, the content of Flash animation can be de-
scribed in a 3-layer structure, as Fig. 2.
The first layer is the global features of Flash animation. The thematic contents of
Flash animation are represented by some key phrases, which are extracted from the
internal embedded texts of Flash animation, such as nouns, verbs, adjectives, phrases,
etc. The type refers to the types of subject matter, including cartoon music TV,
game, VR in desktop, CAI courseware, etc. GIF is composed of the representative
frames of all the scenes, used to display the major content and general picture of Flash
animation scenes. A representative frame is the keyframe in the middle of scene,
basically representing the visual features of scene picture. The second layer refers to
the scene structure of Flash animation. The third layer is the visual features of scenes.
The color is the average color of the representative frame image. The complexity is
measured by the average edge density (the greater the density is, the more compli-
cated the scene is, and the greater the complexity is.), and the length refers to the
frame count of scenes.
On Retrieval of Flash Animations Based on Visual Features 273
A keyframe is the frame in which a key action or a change on some content is de-
fined. But in the SWF file, the keyframes and the generated frames use the same for-
mat of definition and the same method of storage. No special markings to distinguish
the keyframes from the generated frames in SWF file. In order to find out the key-
frames in all frames, the raw data of SWF file need to be parsed.
The SWF file is makeup by a series of tagged data blocks and the structure is simi-
lar to the one of a XML file. In essence, a SWF file can be regarded as an encoded
XML. There are three types of tags, namely Definition Tags, Control Tags and Dis-
play List Tags. Definition Tags are used to define the content of the SWF file, such as
the shapes, texts, bitmaps, sounds, and so on. Control Tags are used to create and
manipulate rendered instances of characters in the dictionary, and control the flow of
SWF file. Display List Tags are used to add characters and character attributes to a
display list.
A keyframe should meet one of the following five circumstances: adding charac-
ters to the display list, removing characters from the display list, modifying attributes
of characters, morphing or including ActionScript code. Adding characters, removing
characters may be simply judged respectively by PlaceObject tag, RemoveObject tag
and RemoveObject2 tag. Modifying the attributes of characters in the display list may
be judged by PlaceObject2 tag and PlaceObject3 tag. The morphing and ActionScript
code can be simply recognized by DefineMorph tag and DefineAction tag.
In the progress of keyframe extraction, it is very important to detect the motion
tweens. It is hard to distinguish keyframes from general frames if only using Display
List tags, because in the motion tweens, PlaceObject2 tags and PlaceObject3 tags are
used to modify the character’s properties in both keyframes and general frames. Ac-
cording to the SWF file structure, the motion tween are recognized by finding that
some same tags are used to modify a character’s properties continuously. In such
situation, the first frame and the last frame are the keyframes.
Due to the complexity of internal structure and storage structure, the screenshot
function provided by Flash animation Player ActiveX is used to save the keyframes
instead of analyzing of the contents of the keyframes after the keyframes are indexed.
Experimental results show that the screenshot method has a higher efficiency than the
content analysis of keyframes does.
Scene segmentation is mainly depend-
ent on the visual features of the scene.
Color is an important perceptual feature
of the image and it is not sensitive to
the change of position and direction.
Color features of an image are more ro-
bust than geometric features. Scene seg-
mentation is fulfilled by comparing the
color difference of two adjacent key-
Fig. 3. The sub-blocks and their weights of
frames rather than frame by frame con- a frame image
sidering the little difference among the
274 X. Meng and L. Liu
middle frames between two keyframes and improving the speed of calculation. The
color difference of two keyframes are calculated on 9 blocks divided with different
size and weight, as illustrated in Fig.3. The color difference between two keyframes is
the sum of the weighted difference of each block.
In order to calculate the color difference rationally, an improved HSI color model
is adopted. In the model, the hue (H) is kept invariant and the saturation (S) and
brightness (I) are adjusted. And the entire color space is compressed into a sphere
with the radius of 0.5. The transformation between our improved HSI model and RGB
model are as follows.
⎧ θ G≥B
H =⎨ (1)
⎩2π − θ G<B
⎧ I
⎪ S 0 0.5 − (Ym 0 − 0.5) 1 − ( − 1) 2 I < Ym 0
2 2
⎪ Ym 0
S =⎨ (2)
⎪ S 0.5 2 − (Y − 0.5) 2 1 − ( I − Ym 0 ) 2 I ≥ Ym 0
⎪ 0 m0
1 − Ym 0
⎩
Y2
I= (3)
Ym
⎛ ( R − G) + ( R − B) ⎞
θ = arccos ⎜ ⎟ (4)
⎜ 2 ( R − G ) 2 + ( R − B )(G − B ) ⎟
⎝ ⎠
3min ( R,G ,B )
S0 =1-
( R +G +B ) (6)
Where, Ym is the maximum possible value when hue and saturation are certain values.
Ym0 is the value of Ym when the saturation is 1.
In the improved HSI color space, the distance (d12) of any two colors (C1, C2) is be-
tween 0 and 1. (the distance between white and black or any other two complemen-
tary colors is 1 )
d12 = ( I1 - I 2 ) 2 + S12 + S 2 2 - 2 S1 S 2 cos( H 1 - H 2 ) (7)
When calculated the value of color difference between two keyframes, a threshold
value is set to determine the border of the scene. If the value of color difference is
greater than the threshold value, the two keyframes divided to different scene. Other-
wise, the two keyframes belong to the same scene. Then the middle keyframe as the
representative frame of the scene is selected and the mean color and image complex-
ity are calculated. Finally, all of representative frames arranged on the production
time-line are converted to a GIF animation to be used to browse the scenes shown in
Flash animation.
Experiments show that scene segmentation based on weighted region has higher
sensitivity and the accuracy rate is 85% compared with the artificial segmentation.
On Retrieval of Flash Animations Based on Visual Features 275
S = D1 × D2 × S t ×
( Sc + Sf ) (8)
2
AIB
St = (9)
A+B
{ {
S c = 1 − max min d ( ci , c j )
i j
}} (10)
S f = mean { R j } (11)
1
Rj =
⎛ δ −δ j ⎞
2
(12)
1+ ⎜ m ⎟
⎝ Δ ⎠
⎧ δj ;δ j ≤ δ m
Δ=⎨ (13)
⎩1 − δ j ;δ j > δ m
Where, D1, D2 is the logical value which expresses the match degree of the size and
the type respectively. St, Sc, Sf is the fuzzy similarity of theme, color and complexity
respectively. A , B , A I B is the number of query keywords, the number of keywords
indexed and the number of the synonymous words of them respectively. ci, cj is the
desired and the indexed color respectively. d(ci,cj) is the Euclidean distance of them in
our improved HSI color space. Rj is the satisfactory calculated with formula (12) at
(
the complexity δj of scene j under δm δm=0, 0.5, 1 meaning respectively to retrieve
)
complex, medium or simple scene . Sf is the average value of Rj.
To demonstrate the feasibility and effectiveness of our approach, an experimental
prototype as a Flash animation retrieval system based on Visual Features of Scenes
has been built. The prototype system has indexed roughly 100,000 Flash animations
downloaded from related Web pages in total. The interface of the system is displayed
in standard Web browser and may be accessed remotely over the Internet supporting
query by keywords (to theme), color and complexity (to scenes). 50 Flash animations
with top similarity calculated based on formulation (8) are retrieved and ranked in
276 X. Meng and L. Liu
descending order. The GIF animation of scenes, the address of related web page, title
(linking to the file address) and the file size of each animation retrieved are displayed
10 per page, as illustrated in Fig. 4. The average retrieval accuracy of 300 retrieving
trials with the combination of 20 themes, 5 colors and 3 complexities, is 83%, much
higher than 57% only with 20 themes.
5 Conclusions
This paper has investigated the problem of retrieval of Flash animation based on vis-
ual features, which is helpful to better utilization of the proliferating Flash animation
resource in web. As the major contribution of this paper, a promising approach of
Flash animation retrieval based on the visual features of scenes is proposed. The pri-
mary experiment on a database with roughly 100,000 Flash animations demonstrates
the flexibility and the effectiveness of our approach. However, there still remain many
issues for future research, for example, extraction of visual features of components,
presentation of logical architecture of the segments and their scenes and components,
retrieval based on integration of visual features, components, logical architecture and
semantic of Flash animations, etc. Some of the issues are our research direction in the
future.
References
1. Flash Player statistics, http://www.adobe.com/products/player_census/flashplayer/
2. Lee, T., Sheng, L., Bozkaya, T., Ozsoyoglu, G., Ozsoyoglu, M.: Querying Multimedia Pres-
entations based on Content. IEEE Trans. Knowledge and Data Engineering 11(3), 361–387
(1999)
On Retrieval of Flash Animations Based on Visual Features 277
3. Yang, J., Li, Q., Wenyin, L., Zhuang, Y.: FLAME: A Generic Framework for Content-
based Flash Retrieval. In: 4th International Workshop on Multimedia Information Retrieval,
in conjunction with ACM Multimedia 2002, Juan-les-Pins, France (2002)
4. Ding, D., Li, Q., Feng, B., Wenyin, L.: A Semantic Model for Flash Retrieval Using Co-
occurrence Analysis. In: Proc. ACM Multimedia 2003, Berkeley, CA (2003)
5. Chang, S.F., Chen, W., Meng, H.J., Sundaram, H., Zhong, D.: VideoQ: An Automated Con-
tent based Video Search System Using Visual Cues. In: Proc. ACM Int. Multimedia Conf.,
pp. 313–324 (1997)
6. Flash Player Developer SDKs, http://www.adobe.com/licensing/developer/
7. Meng, X., Wang, L., Li, H., Zhong, Y.: A Method of Image Retrieval with Color Names.
Chinese Graphic and Image 10(3), 349–353 (2005)
The Design of Web-Based Intelligent Item Bank
1 Introduction
Item bank is “a set of items for a special subject constructed by computer system
according to the education measurement theory”. Item bank following the education
measurement theory is an education measurement tool based on strict mathematical
model.
Item bank construction is a complex system engineering. First, establish the
mathematical model. Then set the attributes indexes of the items and the structure of
the test papers, input the scientific effectual items. Following that, evaluate the testing
effect and situation of the students’ leaning. After putting into operation, adjust adap-
tively according to the testing level of the students. To ensure the scientificalness and
effectiveness, the relative attributes of the items and test papers should not only be set
by experts but also be sample tested using large numbers of sampling data, which can
adjust the effectiveness of the parameters. Editing and testing task for a relatively
complete item bank based on the classical test theory is such a formidable task, which
can not be done well by average institutions. [1]
Presently with the development of information technology, item bank are estab-
lished by companies, schools or their cooperation. They have made certain achieve-
ments in practical application, but with some problems.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 278–289, 2008.
© Springer-Verlag Berlin Heidelberg 2008
The Design of Web-Based Intelligent Item Bank 279
(1) Independent of the other teaching links, used un-widely, so can not be improved
based on sufficient sample data.
(2) Providing reverence answers without education evaluation and intelligent in-
structing.
(3) The adaptability of the system is not so good, for example: the difficulty index
keeps invariable and papers can not be generated automatically.
(4) Having little characteristics of automatic marking and reasonable statistics analy-
sis functions.
In order to solve the problems above, we improved the existed internet item bank
and developed a general item bank system successfully. It consists of subsystems of
item management, automatic test paper generation, autonomous learning, testing, and
study evaluation, which can provide a series of services: automatic test paper genera-
tion, testing, analysis and evaluation.
2 System Design
In order to solve the problem of running independent of other teaching links, the sys-
tem is designed with the relative systems such as testing, evaluation and autonomous
learning. The overall design is illustrated by figure 1.
Students
database
Strategy
database
Score
database
Knowledge
database
Item
Testing Subsystem
database
Test paper
database
b ) Generating by students: The students can input the parameters of paper genera-
tion (different from the one used by teachers) according to their own condition,
then the system generates the paper for exercise of pertinence which can be
printed or saved as web pages.
c) Static generation strategies: There are some common generation strategies for
the common testing. The teachers can choose the ready generation strategies in-
stead of inputting the complicated parameters.
(3) Testing Subsystem
This subsystem can use the papers generated by the system or the ones generated
by the paper generation subsystem. Students take the test online or take the paper
home via storage device to test, and then submit their answers to the analysis mod-
ules. This subsystem includes three links.
a) On-line examination: The students can choose the special paper via browser,
take the test and submit their answers. The system stores the answers into the
student information database, and then the teachers can take them out and do the
marking.
b ) On-line marking: After login in the system, the teachers can choose the paper
they should mark, and then choose the students who took the examination to take
the students’ papers and corresponding answers. The objective questions can be
judged by the system automatically. The non-objective questions are marked by
teachers which are stored into the score database with the marking information
together after submitted by teachers.
c) Results requiring: After login in, the students can choose the papers they tested
and look over the testing contents, standard answer and marking condition, etc.
(4) Autonomous learning subsystem
Autonomous learning subsystem has a close relationship with the testing subsys-
tem. On one hand, it calls the paper generation subsystem to generate the papers, do
exercise autonomously and transfer the results to the analysis modules. On the other
hand, it can give the corresponding advices on knowledge review and learning strat-
egy according the test.
(5) Study Evaluation Subsystem
This subsystem is used to do statistic analysis of questions, papers and the teaching
process according to the history data of the students’ testing which is analyzed by S-P
table. It has two important parameters of difference coefficient and alarm coefficient,
and they can separately provide important information for analysis of teaching situa-
tion and teaching effect. So, this subsystem can analyze the reasonable degree for
individual situation, trends of all the students and improvement of the teaching effect.
It includes the following aspects.
a) Statistics analysis of the papers: The distribution of the items such as students’
points can be shown by line graph and histogram and all the abnormal questions
in the exam can be captured such as all the answers are right or wrong in whole.
Analysis on the following subjects: the reliability, validation and average diffi-
culty of the paper; the maximum and minimum points, the total number of each
grade section, average point and standard deviation ; the original points and the
transformed points for each student.
282 S. Zhong et al.
3 Evaluation Strategies
S-P table is an information processing method of analyzing the teaching based on the
students’ points table of the questions. The method can evaluate the learning situation
of individual student and the trends of all the students; also it can evaluate propriety
of all the questions. Difference coefficient and alarm coefficient are two important
parameters used for S-P analysis. Calculation of them can provide important informa-
tion for analysis of teaching situation and teaching effect.
M students answer n questions, mark 1 when right, mark 0 when wrong, and a matrix
of student – question is created .In the matrix, Uij represents the point student i gets
for answering question j. Because of adoption of 0-1 score method, the matrix is con-
sisted of only 0 and 1.The score matrix can be shown concretely as the condition in
figure 2.
The Design of Web-Based Intelligent Item Bank 283
Student-Problem Matrix
P1 P2 P3 P4 P5 P6 P7 P8 P9 P10 Score
S1 0 1 1 0 1 1 0 1 1 0 6
S2 0 1 1 1 1 0 0 0 0 0 4
S3 1 1 1 1 1 0 1 1 0 1 8
S4 0 1 1 1 1 1 1 1 0 0 7
S5 1 1 1 1 1 1 1 0 0 1 8
S6 0 1 1 0 0 0 0 1 0 0 3
S7 1 1 1 1 1 1 1 1 0 1 9
S8 1 0 0 1 1 1 1 1 0 1 7
S9 1 1 1 1 1 1 1 1 1 1 10
S10 1 1 1 0 1 1 0 0 0 0 5
S11 0 1 0 1 1 1 1 0 0 1 6
S12 0 1 1 0 1 1 1 1 1 0 7
S13 0 1 1 1 1 1 1 0 1 1 8
S14 0 1 1 0 1 0 1 0 0 0 4
S15 0 1 1 1 1 1 1 1 1 0 8
Correct 6 14 13 10 14 11 11 9 5 7
Times
The column on the right lists the scores of the students (the number of the right
ones); the bottom lists the number of right answers for each question. Before the ma-
trix is processed, it can only provide the point of each student and the correct rate of
each question. It can provide much significant information after processed by the
following rules.
a) Students will be arranged by descending order according to the scores. Namely,
exchange the students scoring line, making the high score above in the row, low
score firms bottom.
b) Problems will be arranged by descending order according to the number of cor-
rect answers from left to right. Namely, exchange the problems line, making the
problem line of giving more correct answers to it on the left, the problem of low
frequency of correct answers row on the right side.
c) For the same score line, first of all, find out the sum of correct answers which the
student obtained the errors of each problem in the same column. And then, ar-
range the line of the same score according to the sum; line get high score firms
the top.
284 S. Zhong et al.
d) For the same score column, first of all, find out the sum of the student obtained
the errors of each problem. And then, arrange the column of the same score ac-
cording to the sum; column get low score firms the top.
e) Making the S line. Draw the vertical line for each of the students. Making the
number of the problems on the left side of the vertical equal to the students’
score. And then, drawing the striping between lines, link each vertical line to
form a ladder of the curve, called S line. As Figure 3 shows the solid line.
f) Making the P line. Draw the ledgement for each of problems. Making the num-
ber of students above the ledgement equal to the number of correct answers. And
then, drawing the vertical line between columns, link each ledgement to form a
ladder of the curve, called P line. As Figure 3 shows the dashed line.
After processing, the student-question matrix is transformed into ordered list of S
(solid line) and P (dashed line), which is called S-P table.
S-P Table
P5 P2 P3 P7 P6 P4 P8 P10 P1 P9 Score
S9 1 1 1 1 1 1 1 1 1 1 10
S7 1 1 1 1 1 1 1 1 1 0 9
S3 1 1 1 1 0 1 1 1 1 0 8
S13 1 1 1 1 1 1 0 1 0 1 8
S15 1 1 1 1 1 1 1 0 0 1 8
S5 1 1 1 1 1 1 0 1 1 0 8
S8 1 0 0 1 1 1 1 1 1 0 7
S12 1 1 1 1 1 0 1 0 0 1 7
S4 1 1 1 1 1 1 1 0 0 0 7
S1 1 1 1 0 1 0 1 0 0 1 6
S11 1 1 0 1 1 1 0 1 0 0 6
S10 1 1 1 0 1 0 0 0 1 0 5
S2 1 1 1 0 0 1 0 0 0 1 4
S14 1 1 1 1 0 0 0 0 0 0 4
S6 0 1 1 0 0 0 1 0 0 0 3
Correct 14 14 13 11 11 10 9 7 6 5
Times
Based on the data relationship, it is not difficult to see that the S-P table has the fol-
lowing basic properties.
Because the score of students equals the total time of the right answers, the left-
hand area of S is equal to the area above the P; S is not only the score curve of the
students but also the number curve according to the cumulative scores. P is the accu-
mulating distribution model of the right answer number for each question. S and P are
always intersectant, and when superpositioned, the left end of P is always above S and
the right end of P is always under S; the area gap between S and P is called dispersion
of the two lines which can describe the relationship between the question difficulty
and the students’ response. If S is as same as P, the condition is that the learning state
of all the students is absolute stable.
According to S-P table, holistic analysis can only be done for whole and individual
analysis can only be done for individual.
(1) Holistic Analysis
It includes the difference analysis between the students and the questions and the
distribution analysis between the questions and students’ response. The dispersion of
the two lines reflects the evenness degree of distribution. The increase in dispersion
can be associated with the increase in un-even distribution. The evenness of distribu-
tion can be described as follows:
As for common tests, the value of D ranges from 0.25 to 0.35, and can not over 0.5
usually. If it is over 0.5, the relationship between the questions and students’ response
is abnormal .For example, if a difficult problem is solved by students getting low
points and not solved by students getting high points.
The difference between students and questions can be reflected by the faultage be-
tween S and P. Faultage stands for the beeline range between the domestic wirings.
Figure 4 illustrates the faultage.
If the line S keeps horizontal for quite long, it suggests the danger of polarization.
If the line P keeps vertical for quite long, it suggests that there is sufficient difficulty
difference which may influence the validity of the testing.
(2) Individual Analysis
It evaluates both the learning situation of the students and the pertinence of the
questions. When the dispersion between S and P is quite large, some responses of the
students or the questions are abnormal. As the S-P table in figure 3, the rate of right
answer of Problem P1 is 8/15, which is quite high. But the number who gives a right
answer is just a half both above and below line P, which shows whether the problem
286 S. Zhong et al.
Amplitude of S Line
(a)
P
en
iL
Pf
o
deu
itl
p
m
A
(b)
Fig. 4. Faultage
can be answered correctly is random distributed, and the students no matter with high
or low points answered the questions correctly random. So the differentiation index of
the problem is quite low and need to be checked. For example, there are two students
who both answered 5 problems right, but the problems for each student distributes
nearly halfly on both sides of line S ,which is also quite random and need to pay at-
tentions. The abnormality degree of the students and questions can be described by
alarm coefficient.
Calculation formula for the alarm coefficient:
COV j ( P)
W j (P) = 1 − (3)
COV j (C )
Wj(P) represents the alarm coefficient of problem j; COVj(P) represents the covari-
ance of the response pattern of problem j and the general response pattern; COVj(C)
represents the covariance of the general response pattern and the complete response
pattern. Complete response pattern means the response pattern that above the curve of
S are all 1s and the blow are all 0s.
Calculation formula for the covariance of the response pattern of problem j and the
general response pattern:
1 m 1
COV j ( p ) = ∑ ( X ij − m PX i )(SX i − v)
m i =1
(4)
In the formula, v means the average score of the students, and it can be calculated
as follows:
1 m
v= ∑ SX i
m i =1
(5)
The Design of Web-Based Intelligent Item Bank 287
COV j (P) ∑ X ij S X i − P X jV
W j (P) = 1 − = 1− i =1
PX (8)
C O V j (C ) j
∑ SX
i =1
i − P X jV
In the formula, COVi(S) presents the covariance of the response pattern and the
general response pattern; COVi(C) represents the covariance of the general response
pattern and the response pattern of the student. Response pattern here means that the
left-side of S are all 1s and the right-side are all 0s. As calculating the alarm coeffi-
cient for problems, calculate COVi(S) and COVi(C), and then Wi(s):
n
∑X
j =1
ij PX j − SX i u
Wi (S ) = 1 − SX i (10)
∑ PX
j =1
j − SX i u
In the formula, u means the average correct time for each question:
1 n
u= ∑ PX j
n j =1
(11)
When the alarm coefficient of the student is high, it suggests that this student
makes mistakes in the relative easy problems but solves the relative hard ones. The
phenomena may result from the problems in the learning conditions .For example,
students are not of high learning enthusiasm such as to answer careless, they don’t
exert the potential abilities or answer just by guessing, etc.
When the alarm coefficient of the student is over high, it suggests that this problem
is answered wrong by students with high points but answered correctly by students
with low points. So the differentiation index of the problem is low and its use value
decreases in terms of score levels.
Experiments indicates that: when the value of Wj(P)or Wi(S) is over 0.6,teacher
should pay sufficient attentions to the problems of the students and courage them to
make great effort. Also the problem should be deleted or modified. [3]
288 S. Zhong et al.
5 Conclusion
At present, the system has already established and been on the test-run stage with the
basic functions. It can solve the shortcomings of running independent of the other
teaching links, lacking of education evaluation, intelligent instructing and statistics
analysis well. Yet further research and development should be made.
(1) Further Application of Information Push Technology
Although the system push information such as evaluation, knowledge points to
review after the functions such as statistics analysis and intelligent instructing,
yet the technology is necessary to extend its application in other aspects. There
are some examples. Push the teaching strategies after mining personalized in-
formation such as students’ appetite on study media and learning regulation;
Push more authoritive tests; Push the adaptive tests to students according to their
knowledge regarding knowledge points. [5]
(2) Componentizing the System
The system provides a series of services of automatic test paper generation, test-
ing, analysis and evaluation, which can provide strong support to the distant
education. Each link of this system is also an important link in the current web-
based distant education. The system can be separated into functional subsystems
based on component, which can integrated into the distant education supporting
platform seamlessly and combine with the network courses closely.
(3) Applying the Item Response Theory to the Adaptive Testing
Compare the IRT with classical test theory, the theoretical system of the former is
built based on more complicate mathematical model with more accurate concepts
and theoretical derivation. It’s suitable for the adaptive testing. The testing system
asks the student one or more questions, then chooses the questions which can
evaluate the student most well and truly by analyzing the completed questions. In
this way, on one hand it can make the test corresponds with the student’s ability
level and then provides the more accurate information to evaluate the students’
ability. On the other hand, it can shorten the testing duration effectively.
The Design of Web-Based Intelligent Item Bank 289
References
1. Yu, S., He, K.: Design and Realization of the Network Test Bank System. J. China Distance
Education 164, 53–57 (2000)
2. Jing, Y.J., Zhong, S.C., Li, X., Li, J.-N., Cheng, X.C.: Using instruction strategy for a Web-
based Intelligent Tutoring System. In: Technologies For E-learning and digital entertain-
ment, vol. 3942, pp. 132–139 (2006)
3. Fu, D.: Education Information Processing. Peking (2001)
4. Zhong, Y.J., Liu, J., Zhong, S.C., Zhang, Y.M., Cheng, X.C.: Programming of informatized
instructional design platform for physics. In: Technologies for E-learning and Digital Enter-
tainment, vol. 3942, pp. 171–177 (2006)
5. Liu, L., Wang, W.: Design of Web-Based Exam-question with Self-study and Adaptive Ad-
justing. J. Computer System Application 4, 45–47 (2006)
Methods on Educational Resource Development and
Application
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 290–301, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Methods on Educational Resource Development and Application 291
environment, the instructional design basis and so on. Thus, the developed resource
cannot bring advantages of network into play maximally, cannot solve key difficulties
in learning and teaching process. The developed resource and software are too loose
in global area and too centralized in local area, which have no share functions.
Teachers cannot use them smoothly. On rules and approaches of subject-specific
resource development, animation developing tools, resource integration tools and
excellent resource structure, there are lack of normative studies. These lead to the
lower efficiency and quality of resource establishment. On function, structure and
technology of resource and their support software, less work has been done. The
developed resource and software cannot meet teachers’ needs. On categoricalness and
consistency of resource, there exist more problems such as incomplete, incompatible
and conflicting phenomenon [4].
On resource application and service system, there are no more modes and
approaches for the integration of IT and subject, namely, no operating regulation. The
existing application modes and approaches have no enough scientificaalness,
hierarchy, excellent evaluation system (for example, what is the good integration
class?), effective service support system and share system.
Instructional design under the environment of network means to provide students
materials by utilizing network resource. The communication between teachers and
students or among students mainly uses words. There are no communication ways for
the process of solving problems. Instructional activities under the environment of
network lack of support system on learning, directing, supervising, evaluating and
feedback (for example, the supporting system in the course of investigative learning).
More important thing is that the development and change of IT and instructional
modes are so fast that teachers cannot adapt to these dynamically. Clearly, how to
provide teachers effective application support system has become more and more
important [5].
According to the above, this paper presents methods of ERDA in the course of
curriculum implementation, the curriculum resource structure and integration model,
evaluation factors of resource and curriculum integration, and methods on integrating
intelligent resource and establishing distributed resource system by experimental
study.
Definition 2 curriculum resource means multimedia data and software saved in the
computer, which can be spread through network for students’ learning and teachers’
teaching. These multimedia data and software can bear knowledge, transfer
information, record data, make interaction, make control, make computation, make
simulation and so on. Multimedia types include word, picture, image, video, audio,
animation, software and their compound.
Definition 3 goals on curriculum resource development mean to develop resource
that can help teachers and students to solve difficulties existing before class, in class
and after class, which cannot be solved by the general instructional means. These
resources are divided into the following types: just for teachers, just for students, for
teachers and students, etc.
Methods on Educational Resource Development and Application 293
3 Methods of ERDA
ERDA should be considered from the sufficient integration of IT and curriculum.
Developer should grasp the nature of integration. Methods of ERDA are multifarious,
each school and teacher have their own methods according to their concrete
conditions. Each method has its advantages and shortcomings. At present, there is no
ideal method of ERDA yet. In the following, we contribute a set of ERDA methods
based on the course of curriculum implementation:
Among these steps, instructional mode and process design are the most important
step. If they are not reasonable and scientific, the diagnosis results are unacceptable.
In the course of instructional mode and process design, some factors should be
considered. The following are the key factors:
(1) what are instructional purposes? Where do students’ learning motivations come
from and how to inspire them? Students’ learning motivations come from their needs
of life instinct and curiosity. The learning purposes are to educate abilities, improve
diathesis and master knowledge and information. Concretely, they include mastering
not only basic knowledge but also corresponding knowledge system, logical thinking
approach and essential information based on the frame of abilities. Thus, from the
angle of psychology and education, learning contents that are suitable for the needs of
students’ life instinct and curiosity should be arranged for students.
(2) factors on affecting learning efficiency are multifarious, thereinto, age and
information form are the most important factors among them. Different ages need
different learning modes. People’s means for obtaining information are mainly from
five sense organs. Information forms affect directly the received efficiency. Thus,
from the angle of psychology and education, the effective learning environment,
reasonable instructional process and mode that is subject to the age character should
be made for students.
(3) factors on relationship between knowledge education and innovation education
should be considered. On knowledge education, its goal is to master knowledge and
then change them into skills. Normally, Students educated in this way just can do the
repeated work. However, society needs persons with innovation ability. Thus, the
instructional result should be to make students think about more questions instead of
no questions.
The role that IT plays in the course of curriculum implementation is subject to the
sufficient analysis on IT advantages and features. Analysis on IT advantages and
features should be from multi-angles such as in which parts of instruction IT should
be applied, what kinds of learning environment can be made by IT, etc.
The advantages and features of IT include dynamic interactivity, independent
usability, from stillness to movement, from abstract to concrete, from micro-state to
normal-state, virtuality, simulation, fast transmissibility, real-time, amplitude, etc.
These features are good for inspiring students’ learning interest, integrating excellent
teachers’ and experts’ wisdom together and inheriting effective instructional modes
and approaches.
In the course of IT and curriculum integration, we should do more research on IT
method. IT can set up multifarious instructional environment. Instructional
environment is divided into the following types: multimedia classroom, computer
classroom, campus network and internet. It is necessary to study their features and
application. For each instructional environment, we analyze it from the following
aspects: instructional modes, instructional steps and needed condition, etc.
Methods on Educational Resource Development and Application 295
4.4 Internet
5 Resource Architecture
Definition 6: curriculum resource system means all kinds of resource types, which
includes some based on curriculum criteria, some based on textbook version, some
based on test resource, some based on public resource among curriculums and so on.
Definition 7: textbook version resource means instructional resources organized by
some textbook versions. These resources mainly focus on resolving difficulties that
general instructional means cannot solve during the course of instructional modes
implementation. They are organized by curriculum unit, subject, chapter, passage,
item and so on.
Definition 8: material granularity means knowledge and skill, process and method as
well as the scale of sensibilities value implemented by these materials.
Definition 9: material share means materials that can be used in two or more than two
courseware. There are two means for material share, one is courseware share for the
same content, and another is courseware share for the different content.
Definition 10: curriculum criterion resources mean resources that are organized
through basic knowledge system, encyclopedia information, special topics, test
training questions, learning rules, teaching rules, etc. They don’t depend on
textbook’s version. Some textbook’s resources can be built based on them.
Definition 11: public resource means resource that reflect people’s living environment
and can be shared by subjects. It includes natural and humanistic resources. Typical
public resources include information on nation, region, city, organization and people
as well as some natural environment information on mountain, river, lake, ocean and
so on.
Definition 12 test training resource means resource that can be used to test if students
understand curriculum knowledge or not and to discover where problems exist. It
includes curriculum forward model, curriculum backward model as well as related
test questions and papers.
Figure 2 illustrates the structure of curriculum resources.
Curriculum resource can be divided into three hierarchies from figure 2, namely,
basic hierarchy, category hierarchy, application hierarchy. Resources in basic
hierarchy are the basic curriculum resource types; they have the maximal share
degree, normally, they are 100 percent share. Resources in category hierarchy are
integrated through specific questions; they have the higher share degree. Resources in
application hierarchy are integrated for the concrete applications; they have the lower
share degree.
Methods on Educational Resource Development and Application 297
Resource evaluation should be considered from multiple factors. The following are
key factors:
(1) the proportion of usable resource in each class;
(2) the proportion of share resource in each class;
(3) the granularity of share resource;
(4) the distributing uniformity of total resources;
(5) the reliability of resource design
(6) the validation degree of resource used in instruction;
(7) the freshness of resource update.
(8) the satisfaction degree between resource and curriculum instructional needs;
(9) the leading degree of resource.
The function of each factor is different, some of them are sufficient, some of them
are necessary and some of them are sufficient and necessary. Each factor should be
assigned to a weight in the course of evaluation.
B. Formula
After getting all information, compute the total score by using the following formula,
0≤S≤100 .
(1)
S=p1×p2×p3×(p4×r4+p5×r5+p6×r6+ p7×r7+ p8×r8)
C. Weight Assignment
8 Conclusions
In the course of ERDA, it should be to make clear the orientation; study curriculum
types and difficulties existing in instructional process; arrange learning contents that
are suitable for students; select learning process and environment that are helpful for
students; cultivate logical thinking ability while applying visual technology; think
about integration between subjects; establish teachers’ continuing education system;
improve application and management system; take integration as a system
engineering.
Try to form a leading modes and methods system: modes and methods system on
the integration of IT and different content type; on the establishment of resource and
software structure; on the application of resource and software. ERDA needs
educational experts, IT expert, officer and teachers work together to make great
progresses.
Methods on Educational Resource Development and Application 301
References
1. Nan, G.: Several theoretical and practical problems on education informatization. J. E-
Education Research 11, 3–6 (2002)
2. Sang, X., et al.: Strategy on district educational informatization. J. E-Education Research 3,
8–11 (2005)
3. Jin, Y., Zhong, S.: Analysis on modern long-distance instructional modes and support
platform. J. Educational Technology Research 4 (2003)
4. He, K.: Nature of e-Learning–integration of IT and curriculum. J.E-Education Research 1,
3–6 (2002)
5. Gu, X., Zhu, Z.: Implementation mode on teachers professional development. J. China
Educational Technology 3, 5–8 (2005)
6. Wu, J., Yu, S.: General problems and countermeasure for IT and curriculum integration. J.
E-Educational Research 10, 70–72 (2006)
7. Tang, Q.: Analysis on errors of IT and curriculum integration. J. E-Education Research 10,
70–72 (2006)
8. Zhong, S.: Research on resource and software of IT and curriculum integration. J. E-
Education Research 3, 53–56 (2005)
9. Zhong, S.: Model and method on IT and curriculum integration. J. China Educational
Technology Research 9 (2003)
Research on Management of Resource
Virtualization Based on Network
1 Introduction
In practice of university education, it is becoming an increasing important topic
to discuss how to deal with the problem about lack of resources. Especially, in
recent years, those things, such as number of students in university increased
remarkably and finance of education relatively decreased, bring a serous lack
of resources in universities of China. The shortage of resources exists in many
aspects, for example, equipments, locations and human resources, and so on.
Obviously, increasing resources is a direct way to change the situation, but it is
not the best selection for us since it can’t economically solve the problems indeed.
In fact, there are many resources, which are not used efficiently and sufficiently, in
the universities. The main reason for this is that the resources mentioned before
are exclusively occupied by different universities, which results the resources
can not be shared with other universities. So it is an effective and natural way
to solve the shortage of resources by reforming traditional pattern of resource
management based on information technology, which can share resources and
enhance efficiency of resource using.
The main task of university is to cultivate students and satisfy various de-
mands of learners. The goal of instructing student, the pattern of education and
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 302–309, 2008.
c Springer-Verlag Berlin Heidelberg 2008
Research on Management of Resource Virtualization Based on Network 303
the frame of organization are deeply effected changed by the coming informa-
tion age. In university, Teaching environment rapidly improves and the limits
between the different schools and subjects are being broken. The educational
pattern based on the teacher’s instruction is lagged with the times in industrial-
ized era. Today, we need a new educational pattern to satisfy the individualized
demands of students. At the same time, we also need to offer some environments
and resources of learning with relevant patterns and corresponding management
system to satisfy the demands of students.
Based on the situation mentioned above, we have to reform today’s resource
management pattern of universities to fully display all the teaching resources
and to satisfy the learners’ individualized demands. In this paper, we proposes
a strategy to realize the resource management virtualization based on the in-
formation technology. The rest of paper is organized as follows. Firstly, we dis-
cuss the value and model of resource management virtualization and the related
work in Section 2, then give a general framework of resource virtualization man-
agement in Section 3. Some suggestions about detailed implementation of the
model is proposed in Section 4, followed by our conclusion of the framework
and discussion about the significance of management virtualization pattern in
Section 5.
benefits. Considered from the user, virtualization is only considered uses, did not
consider possesses.
By analyzing the kinds of teaching resources, three kinds of attribution should
be found. The first, attachment. That is to say, which department the resources
belong to. Certainly, in view of overall aspects, the school has all ownerships. But
according to current circumstance, it usually still belongs to a certain section in
the school. The second, management . The resources needs to be managed and
maintenance, this attribute usually has relation to attachment, in anther words,
it is managed and maintenance by whom it belongs to. The third, usage. It is the
attribute that serve the teaching and play an important role. This is the most
important part, it is also a part that customer to compete with each other fiercely.
in current circumstance, three kinds of above-mentioned attributes of teaching
resources are generally and all managed by a department, it is also the traditional
management style that we usually say. the share of resources is limited by this
kind of mode, because proprietor of resources usually provide a service toward
other customers after satisfying own demands, but if the usage application of
other customers conflict with the department benefits, the proprietor will carry
on restriction to the resources share.In this kind of mode, all resources have to
unify together.It can’t provide the learner individuation study space, and it can’t
satisfy the learner’s study need.
Management of resources management is putting forward for resolving the
above mentioned problems, the purpose is to maximize the usage attribute, and
provide individuation study environment for the learner.
So we can give such definition: management of resources Virtualization of
the university and college base on the foundation of network, on the premise
of keeping management and attachment attribute unchanged, usage attribute
may be separated from other components ,constitute a logic resources layer(or
a virtual resources layer), and build up a specialized resources Virtualization
scheduling center to schedule resources, according to the certain management
strategy. Any user can apply for logic resources to handle his mission.
according the strategy of the school and the fact instances, can decide the re-
sources whether can be shared, or the range of sharing.
Providing the user interface, a portal. It includes some shared information re-
sources, user login interface, and so on. It is the only interface for user applying
for resources. Actually, It is users who apply for resource using from portal.
Realization of resource sharing has become the consensus of the majority higher
education workers, but in the specific practice, there are still many factors af-
fecting share. Aim to the management model of the virtual resources present in
this paper, we put forward the following proposals for its implementation.
The most important is still the concept and it is need to promote the concept
changed by various ways such as propagation and education. Judging from univer-
sities’ reality, we know that majority branches are still habit monopolizing various
resource in self, especially the teaching unit. Except the factor of management,
concept is a very important factor. We deal with several viewpoints here.
Firstly, school leads’ concept. Come to say resource’s the sort , lead to be to
be ready to come true tier. But in putting process into practice concretely, still
existence some affect the factor leading a wish’s. For instance, the management
boundary between different administers would be mixed up because of share. In
other words, it is hard to find junior organization. Obviously, it is a question of
habit. If we do canonical management, these questions will not appear on the
way of our work.
Secondly, functional branches’ concept. Functional branches always think that
the management is its patent. Secondly, functional branches’ concept. Functional
branches always think that the management is its patent., the virtualization
management will affect the management way of the functional branches without
a doubt, even will be able to form the feeling which one kind of management
Research on Management of Resource Virtualization Based on Network 307
jurisdiction will reduce. If this level cannot change their idea, it will seriously af-
fect the virtualization management implementation. Thirdly, branch of resources
ownership and user’s idea. In the university, user and ownership branch’s role
are mutually transform in some time. In the tradition, everyone is accustomed
to use monopolistically.
5 Conclusion
The pattern of the resources virtualization management which designed in this
paper based on the existing resources can enhance the efficiency of using re-
sources and sharing resource, through the information and standardized man-
agement, also can provide the individuality study space and the environment for
the learners. From the experiment which already developed see, which had the
following several mean.
Acknowledgement
The work presented in this paper is supported by the Natural Foundation of An-
hui Provincial Education Department (No.2006KJ041B and No.KJ2007B073),
the Teaching Research Key Project of Anhui Province of China and the Natu-
ral Science Foundation of the Jiangsu Higher Education Institutions of China
(No. 2007jyxm115 and No. 07KJD460108).
References
1. Trow, M.: Lifelong Learning through the new information Technologies [J]. Higher
Education Policy 12(2), 201–217 (1999)
2. Trow, M.: Some Consequences of the New Information and Communication Tech-
nologies for Higher Education[EB/OL],
http://cshe.berkeley.edu/publications/docs/
3. Egol, M.: The future of higher education[J]. Educause review 41(4), 72–73 (2006)
4. Maldonado, M.F.: Virtualization in a nutshell: A pattern point of view[EB/OL]
(June 2006), http://www.ibm.com/developerworks/library/gr-virt/
5. Garbacki, P., Naik, V.K.: Efficient Resource Virtualization and Sharing Strategies
for Heterogeneous Grid Environments. In: 10th IFIP/IEEE Symposium on Inte-
grated Management (IM 2007), Munich, Germany, May 2007, pp. 40–49 (2007)
The F-R Model of Teaching in Chinese Universities
Modern Education Technology Center, Central South University 410083, Changsha, China
huizh@csu.edu.cn
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 310–315, 2008.
© Springer-Verlag Berlin Heidelberg 2008
The F-R Model of Teaching in Chinese Universities 311
negligence of the students’ cognition process, the students are in passive situation and
can’t carry on deep thought. So their thinking mode and innovative abilities are
fettered.
The information technology instruction that is grafted on the traditional model of
teaching consolidates the traditional model of teaching and doesn’t make the real
modern education technology instruction come true. Thus in order to promote the
fusion between the information technology and the instruction in university and
improve the quality of instruction in university, the paper tries to make some explore on
the model of teaching in Chinese universities and puts forward the F-R model of
teaching for the first time. The new model synthesizes the rigid instruction based on
class in universities and the flexible instruction based on education of individual
information accomplishment. And the new model makes the realization of instruction
according to person that based on learning task come true with the help of information
technology and web-based teaching platforms. In this case, the students can study
independently and construct their knowledge meaningfully under the teacher’s
instruction. So it can help to develop the students’ innovated ability.
About more than 1000 years ago there was “Si Shu” that the teacher exerted
individuality education on the students in China. That the artisan guiding the prentice
was another pedagogical activity based on the study skill. Both the “Si Shu” teaching
and the apprenticeship teaching may belong to flexible model of teaching. The
characteristic of flexible teaching system is that under teacher’s instruction students can
make asynchronous study come true. Or in other words, the content and the progress of
the students may be different. The teacher manages the whole process. According to the
diversity and the desire of students, the teacher devises different plan of instruction for
different students and adjusts the teaching goal. The flexible model of teaching can
reach multidimensional cultivation so the students can devote themselves to the
development of society more quickly.
The obvious characteristic of the flexible model of teaching is to implement
individual instruction to the students. It is good for the teachers to discover students
individuality, instruct the students to study and develop their personal thinking. Under
this model of teaching, the teacher plays an important role in the improvement of the
teaching equality. The teacher manages each student's entire studying process;
therefore the amount of students instructed by each teacher is limited. So the flexible
model of teaching is not adapted to popular education.
312 H. Zhao, Y. Huang, and J. Zhang
After entering the 18th century, the assembly line started to be applied to the
mechanical industry to make large scale production. Correspondingly, the same
characteristics come up in the pedagogical activity. That is to divide the teaching
content into certain units: required curriculum, technical required curriculum,
professional required curriculum and professional curriculum. The process of
instruction was similar to the assembly line. After passing the “assembly line", the
students turned into professionals. This kind of model enhanced the teaching efficiency
enormously, so a large amount of professionals may be “produced” in a short time.
The rigid model of teaching emphasizes the collective design of instruction.
Teaching plan is made according to most of the students in class and the teaching goal
is explicit. So it is easier for teaching management. The rigid model of teaching exerts
level- teaching according to the structure of knowledge so it is easier to shape the
thinking process of the students. While the disadvantage is that it ignores individual
difference so that one part of the students can’t exert their potential adequately and
another part of the students deal passively and can’t acquire knowledge in true sense.
their study resources and is convenient for teachers to instruct students to study
independently, to explore new domain and to develop students’ innovative ability.
Presently the teaching system emphasizes the influence to students from the outside
and the one-way transmission of knowledge from teachers to students in the full-time
universities in China. The teachers make the teaching plan according to the traditional
model of teaching, and teach the class according to the uniform teaching plan with the
isolated teaching methods such as multimedia. During instruction, the teaching
progress is made according to the students in the medium of the class without
considering the difference between students. So it is difficult to cultivate innovative
people that are demanded by society.
According to the characteristics that the teacher teaches based on subject and the
students study based on grade and major in the full-time universities in China, the F-R
Model of teaching pays more attention to the students’ meaningful construction under
the instruction of the constructivism theory. Also it establishes cognitive environment
based on the learning task and multi-level network knowledge system within school for
students to develop the learning ability under the instruction of the teacher.
discuss and discriminate mutually based on the “scene”. Thus students can achieve
their sensible construction and develop their high-level cognitive ability.
Each group introduces the difficulty and the attainment in the exploring process
briefly to the whole class. It can both facilitate other students to study, oppugn and
evaluate and activate the atmosphere in class. That the teacher concludes in time can
help students systematize the scattered knowledge they obtain effectively. And under
the hint and the introduction of the teachers, students should try to conclude and
summary by themselves.
That the assignment, the submission and the rectification of the homework are an
important alternant means to communicate the teaching of the teacher and the study of
the students. Through the practice, the students may consolidate and enhance the
comprehension and grasp of the learning content. During the practice and the review
after class, students may meet difficulties and need help and support at any moment.
The web-based teaching platform provides convenient accesses to solve these problems
and makes study more open without the limit of the space and time.
The Q/A system is divided into the synchronized Q/A and the asynchronous Q/A:
The teacher may ask the students synchronously and also may design the intelligent
teaching system to answer the common questions. The teacher can further obtain the
study state of the students while answering. It is good to the flexible teaching, too. At
the end of the study, the quality that students complete the task can be evaluated by “the
work” —the thesis or the project of graduation
5 Conclusion
Under the F-R model of teaching, the advanced instructional and educational thought is
taken to the class to instruct the whole instructional and educational process and
cultivate innovative person. In the actual teaching process, the teachers should appraise
every student’s study condition and adjust some links to meet the actual needs. Because
the F-R model of teaching integrates the classroom bearing the emotion exchange with
the free and open learning environment based on information technology, it both exerts
the teachers' domain role and emphasizes the students' main body status. So it forms the
harmonious atmosphere in the teaching process and enhances the study effect. Thus it
guarantees the cultivation of the students' emotion, attitude and values and shapes the
students' perfect personality
References
[1] Hui, Z., Xiangjun, C.: The interactive way and application of web-based instruction [J].
Educational Technology (2004)
[2] Kekang, H.: Constructivism —the theoretical foundation to innovate traditional teaching
[J], E-education Research (1997)
[3] Shizhen, D.: Modern Education Information Technology Basic [M]. Central China Normal
University Press, Wuhan (2000)
An Approach to a Visual Semantic Query for
Document Retrieval
1 Introduction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 316–323, 2008.
c Springer-Verlag Berlin Heidelberg 2008
An Approach to a Visual Semantic Query for Document Retrieval 317
in computer learning systems, may not have the technical skills to execute com-
plex command queries. Visual queries have been applied to ontologies mostly as
search and construction tools [5].
In this paper, we propose an approach to design an interface for document
retrieval based on their semantical information. Our approach consists of three
parts. First, we discuss the use of ontological information in the indexing of
documents. Second, we describe a method for translating the relative positions
of graphical objects into semantic query statements and a means of integrating
query objects with the query results in a common graphical area. The main pur-
pose of this study is to develop a prototype system which enhances the user’s
knowledge while he/she searches information. Using graphical features, such as
free object movements, queries can be executed dynamically, providing informa-
tion to the user as he/she interacts with the visual elements.
2 Related Works
Document retrieval methods using semantic web technologies such as ontologies
have been popular in recent years because the procedures for semantic web in-
volve textual analysis. These techniques are also used for information extraction
in document retrieval systems [6,7,8]. The system KEA [9] uses a method of
recovering documents through the use of ontological information. Initially it ap-
plies several textual procedures such as stemming, common word removal, and
document indexing. The indexes are created with the use of an ontology. KEA
was used in this research because of the good extracting results in comparison
with system-extracted and expert-extracted keyphrases [7].
Due to their hierarchical structure, it may be difficult to visualize the ontolo-
gies. Graphical tools have been developed for a better understanding of ontolo-
gies [10,11]. These studies use several graphical approaches in order to simplify
the amount of information to be displayed
Graphical interfaces for document retrieval have been specifically researched
[12,13]. Because retrieving documents can create extensive lists of results, spe-
cially in large repositories, it is necessary to apply visual aids to recover such
information. Although researches exist in the fields of document retrieval, visual
interfaces and ontology visualization, a combination of these technologies has
not been approached extensively.
3 Framework
Figure 1 presents the framework of our proposed system. To realize the described
approach, the following functions are needed: First, a function for extracting the
most relevant keyphrases from documents, using information from an ontology;
second, a function for executing queries over the ontology and the extracted
keyphrases; third, a function to transform the queries and results into visual
queries; and finally, a function to provide an interactive interface.
318 P. Villavicencio and T. Watanabe
Server Client
query
results
Keyphrase Query Query web services Graphical
extraction processor translator Query
query
query statement
requests
User graphical
Interface
File system
document extracted
ontology
repository keyphrases user
interactions
The framework is divided in two tiers, a web application server, and a desktop
client. The web application contains the function of keyphrase extraction and
query functions. The desktop client contains the interface functions. The com-
munication between the server and the client is done via web services. Although
the document repository and the ontology reside in the file system of the server,
it is not restricted only for this server. File systems in other machines may be
used as well, in order to create a more flexible structure.
The digital documents consist of texts stored in the file system. These are first
processed by the Keyphrase Extraction Algorithm (KEA)[3]. This extraction
method requires an ontology. The textual documents are processed using lexical
methods, and then analyzed with machine learning mechanisms, in order to
select the most relevant keyphrases. This method is semi-automatic due to the
fact that the machine learning requires a set of documents with manually selected
keyphrases in order to build a model. Later, this model is used to process other
documents.
In the graphical representation of a query, small circles stand for keyphrases and
ellipses for relationships. The combination of these items contained inside a larger
circle represent a query structure graphically. After a query execution, results
are also illustrated as circles located in the surroundings of the query circle.
Figure 2(a) shows the graphical representations and Figure 2(b) is a SPARQL
query statement.
result
subspace
prefix skos: <http://www.w3.org/2004/02/skos/core#>
relationship keyphrase
SELECT ?uri ?label WHERE
keyphrase {
?uri skos:prefLabel ?label
FILTER (?label ~ 'text')
query }
subspace
Fig. 2. (a) The graphical representation of a query statement. (b) A SPARQL query
statement.
The interface consists of elements that can be freely positioned inside the user
interface layout. The locations of the elements provide important information
to the system, since the query elements depend on their locations with respect
to others. At each user interaction, such as element focus, element drag, etc.
the system analyzes the manipulated elements in comparison with others. If the
new arrangement changes elements in query, a query statement is formed and
the information is sent as a query statement, results are immediately obtained
upon the SPARQL query execution. Since results constitute of keyphrases, the
320 P. Villavicencio and T. Watanabe
4 System Characteristics
User Interface
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
keyphr
any location. The results from a query are returned after analyzing the locations
of objects in the interface. Figure 3 shows the user interface with two queries.
In this case a user has created two queries with related keyphrases. Results are
displayed by the system in locations near to the keyphrases inside a query space.
Each keyphrase object contains the location and the term of keyphrase. This
information is processed by the query transformer in the following way: first,
the positions of all the keyphrase objects and the query spaces are collected and
sent to the query transformer. Second, with this information, each query space
is transformed into a statement, using the keyphrases as search elements. Third,
the query is executed over the ontology-document map stored in the keyphrase
extraction phase. The query result consists of semantically similar keyphrases to
moved keyphrase. Forth, the list of results is sent back to the query transformer,
which translates the results again into graphic objects for the interface. The
results are placed according to the position of the queried keyphrase and in
relation with the query space. The process of transforming the visual keyphrase
elements depends on their graphical positions. The query takes into account
the relative distance between each keyphrase, inside a query space. The query
transformer compares the distances between each keyphrase, if the distance is
less than a defined value a relationship is added to the query statement.
5 Prototyping System
The visual interactive interface consists of two spaces: the query space and the
result space, with a circular shape both will determine the area in which the
user can manipulate the query elements and the resulting elements. The result
space is located at the surroundings of query space.
a) b)
Fig. 4. (a) The results are given as a single query. (b) The keyphrase ”child” and
”disease” are separated, creating a separate query for each one.
322 P. Villavicencio and T. Watanabe
6 Conclusion
In search tools, the problem of having long lists of results can be solved by
considering the semantic relationships among keyphrases. With this option, the
results of a search would be given, not as primer attention to the documents, but
rather to keyphrases, these can be understood by the user, and also related. By
learning from these relationships, the user can select relevant documents that
match his/her search.
References
1. Beale, R.: Supporting serendipity: Using ambient intelligence to augment user ex-
ploration for data mining and web browsing. Int. J. Hum.-Comput. Stud. 65(5),
421–433 (2007)
2. Shah, U., Finin, T., Joshi, A.: Information retrieval on the semantic web. In: CIKM
2002: Proceedings of the eleventh international conference on Information and
knowledge management, pp. 461–468. ACM, New York (2002)
3. Jones, S., Paynter, G.W.: Automatic extraction of document keyphrases for use in
digital libraries: evaluation and applications. J. Am. Soc. Inf. Sci. Technol. 53(8),
653–677 (2002)
4. Kiryakov, A., Popov, B., Terziev, I., Manov, D., Ognyanoff, D.: Semantic annota-
tion, indexing, and retrieval. Web Semantics: Science, Services and Agents on the
World Wide Web 2(1), 49–79 (2004)
5. Wienhofen, L.W.M.: Using graphically represented ontologies for searching content
on the semantic web. iv 00, 801–806 (2004)
6. Bhogal, J., Macfarlane, A., Smith, P.: A review of ontology based query expansion.
Information Processing & Management 43(4), 866–886 (2007)
7. Jones, S., Paynter, G.W.: Human evaluation of kea, an automatic keyphrasing
system. In: JCDL 2001: Proceedings of the 1st ACM/IEEE-CS joint conference on
Digital libraries, pp. 148–156. ACM, New York (2001)
8. Khan, L., McLeod, D., Hovy, E.: Retrieval effectiveness of an ontology-based model
for information selection. The VLDB Journal 13(1), 71–85 (2004)
9. Witten, I.H., Paynter, G.W., Frank, E., Gutwin, C., Nevill-Manning, C.G.: Kea:
practical automatic keyphrase extraction. In: DL 1999: Proceedings of the fourth
ACM conference on Digital libraries, pp. 254–255. ACM, New York (1999)
10. Katifori, A., Halatsis, C., Lepouras, G., Vassilakis, C., Giannopoulou, E.: Ontology
visualization methods-a survey. ACM Comput. Surv. 39(4), 10 (2007)
11. Tzitzikas, Y., Hainaut, J.L.: On the visualization of large-sized ontologies. In: AVI
2006: Proceedings of the working conference on Advanced visual interfaces, pp.
99–102. ACM, New York (2006)
An Approach to a Visual Semantic Query for Document Retrieval 323
12. Jones, S., Staveley, M.S.: Phrasier: a system for interactive document retrieval
using keyphrases. In: SIGIR 1999: Proceedings of the 22nd annual international
ACM SIGIR conference on Research and development in information retrieval, pp.
160–167. ACM, New York (1999)
13. Crestani, F., Vegas, J., de la Fuente, P.: A graphical user interface for the retrieval of
hierarchically structured documents. Information Processing & Management 40(2),
269–289 (2004)
14. Corby, O., Dieng-Kuntz, R., Gandon, F., Faron-Zucker, C.: Searching the seman-
tic web: Approximate query processing based on ontologies. IEEE Intelligent Sys-
tems 21(01), 20–27 (2006)
Modification of Web Content According to the User
Requirements
Pavel Ocenasek
Abstract. The paper deals with the system for modification of web content
according to the user requirements. The system is based on the network proxy
server. The general idea is to employ translation rules (regular expressions) to
render web pages upon user requests stored in profiles. The concept of the
proposed proxy is general and the system can be used for various purposes. The
main targets of use are: providing secure content of web sites, translation of
non-accessible web pages into accessible form, etc.
1 Introduction
When the World Wide Web service was created and the markup language HTML
became its main pillar of strength, only some people could foresee that it becomes one
of the most valuable research or work instruments of wide society. Some of the best
qualities that this service offers are availability and immediate diffusion of
information published on the Internet. These characteristics are especially useful for
users with some types of disability. Moreover, they have seen how their access to
leisure, education, business or research activities has been improved.
2 Web Accessibility
To develop accessibility standards for Web sites and authoring tools, the W3C
Consortium (www.w3.org) [2] [7] adopted the Web Accessibility Initiative (WAI).
WAI guidelines group checkpoints into three levels of priority. Priority one includes
checkpoints that Web site administrators “must” implement. For example, users must
be able to avoid behavior that obscures the page content or disorients them. Flashing
content can cause seizures in people with photosensitive epilepsy or distract
cognitively impaired people. Distracting background images or sounds can affect
those with visual or hearing problems. Priorities two and three are checkpoints that
“should” or “may” be implemented [4] [6].
To avoid these problems, users must be able to filter WWW content or multimedia
presentations. However, structure and meta information is hard to recognize and to
filter. The main problems are:
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 324–327, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Modification of Web Content According to the User Requirements 325
3 System Concept
We have developed a new system, which will be useful for accessing web pages by
visually impaired users and translate these pages into the accessible form. The system
has been designed to make the web pages accessible independently [5] from the
presentation devices and technologies used.
The main idea of the system can be seen from the following figure:
Fig. 1. The principle of automatic translation system. The system can be used either as a
network proxy server (via proxy settings) or simple document proxy server (via URL prefix).
The system works as a proxy server for translating common internet pages into the
accessible form. The web accessibility is described by translation rules, that are
applied to the common pages.
The usage of our system is very easy. Before the first use, visually impaired user
creates a profile where impairment-specific requirements for the translation are
specified. Then the system is used via the standard web browser by specifying the
URL to translate in the form: http://www.bezbarierovy.net/www.yahoo.com . The
translation of the main page as well as all the linked pages that user visits from the
starting page is done automatically.
326 P. Ocenasek
4 Implementation
In general, the accessibility is performed according to the following instructions:
1. IMAGES – images could be easily switched off, resized or the color
depth/contrast can be changed according to the user-specific requirements.
2. LINKS – visitors to the web pages are looking for information, and the more
efficiently they can find it, the more valuable the site is to them. Most screen
readers have a shortcut command that will give users a list of all the links on a
page. This is a way to skim a page quickly.
3. COLOR – Consistent use of color can enhance the usability of your pages for
many people. We have to be sure that no information is conveyed solely through
the use of color.
4. TABLES – there are two simple things we can do to make tables more accessible
without changing their appearance. One is to use the summary attribute. This
attribute goes in the table tag along with the border, cell spacing and other
attributes. The other thing we can do is to use the scope attribute in the first cell in
each row and first cell in each column.
5. HEADINGS – those of us who are sighted use headings as a quick way to scan the
organization of a page. To create headings, many people use the font tag to make
larger text. However, most screen readers have a shortcut command that produces
a list of all the headings on a page created with the heading tag. If the page is well
organized and uses heading tags for headings, this can be a great way for visitors
using screen readers to skim the page.
There are many rules and specific translations that belong to these (and other)
categories. The detailed description is beyond the scope of this paper.
The proxy server can be used in two modes:
• Document proxy server, this mode is used when the impaired user enters
the URL address in the standard browser in the following form:
http://www.bezbarierovy.net/<URL_to_translate>. The system translates the
starting page and automatically follows all links into the recursive translation.
• Network proxy server mode serves on a specified TCP port and translates all the
content going through. The proxy server is activated by setting the proper address
and port in the browser settings (the Connection/Proxy parameter). Then the
common form of URL address is typed into the browser and the content is
automatically translated.
In both modes of use the proxy server is transparent and browser independent. The
translation is done according to the settings from the user profile.
The main contribution of this paper is the presentation of the system, which is
based on document-proxy techniques and translates web pages into the accessible
form upon specified translation rules. The main advantage of the presented system is
the universality of use and browser independency. Therefore, visually impaired users
can use this system from various places with access to the Internet, such as home
computers, libraries, school laboratories etc. Additionally, users can use their own
stored profiles to make the browsing and accessibility more specific to their
requirements.
Our next plan is to improve the user interface, the user-specific profiles and to
simplify the rules into regular expressions, which will be used for translation of web
content. We then will try to put these improvements into practical use.
Acknowledgement
The research has been supported by the Czech Ministry of Education in frame of the
Research Intention MSM 0021630528: Security-Oriented Research in Information
Technology, MSM 0021630503 MIKROSYN: New Trends in Microelectronic
Systems and Nanotechnologies, and by the Grant Agency of the Czech Republic
through the grant GACR 102/08/0429: Safety and security of networked embedded
system applications.
References
1. Isaak J.: Toward Equal Web Access for All. IT Pro 11-12 / ’00, IEEE 1520-9202/00, 49-51
(2000)
2. Kirchner, M.: Evaluation, Repair, and Transformation of Web Pages for Web Content
Accessibility. Review of Some Available Tools. In: Proceedings of the Fourth International
Workshop on Web Site Evolution WSE 2002, 0-7695-1804-4/02 (2002)
3. Liu, S., Ma, W., Schalow, D., Spruill, K.: Improving Web Access for Visually Impaired
Users. IT Pro 7-8 / ’04, IEEE 1520-9202/04, 28-33 (2004)
4. Macías, M., Sanchéz, F.: Improving Web Accessibility for Visually Handicapped People
Using KAI. In: Proceedings of the 3rd International Workshop on Web Site Evolution WSE
2001 0-7695-1399-9/01 (2001)
5. Ocenasek, P., Toufarova, J.: Web Accessibility for Visually Handicapped People. In:
INFORUM 2005: 11th Annual Conference on Professional Information Resources, Praha
(2005) ISSN 1801-2213
6. Whitelaw, K.: Why Make Websites Accessibile? And How? In: SIGUCCS 2003, San
Antonio, Texas, USA, pp. 259–261. ACM, New York (2003)
7. W3C World Wide Web Consortium. Cited 2008-01-30 (2006), http://www.w3.org
8. Ocenasek, P.: Automatic System for Making Web Content Accessible for Visually Impaired
Users. In: WSEAS Transactions on Computers Research, Athens, GR, vol. 1(2), pp. 325–
328 (2006) ISSN 1991-8755
9. Ocenasek, P.: Automatic System for Making Web Content Accessible for Visually Impaired
Users. In: Proceedings of the 6th International Conference on Applied Computer Science,
Puerto De La Cruz, ES, pp. 430–433 (2006) ISBN 960-8457-57-2
Virtual Environments with Content Sharing
Abstract. Content sharing over networked devices, beyond simple file sharing is
becoming a reality. Many devices are forming closer relationships with different
virtual worlds, such as World of Warcarft and Second Life. In one sense the gap
between the two is becoming increasingly more blurred. Consequently, this
opens up many new avenues for content sharing, not only between devices but
also between sophisticated virtual worlds. Given such interoperable platforms a
natural progression sees content that seamlessly resides within either. This will
open up new opportunities where third-party content providers and users alike
are able to create and share content over these new platforms. We aim to provide
a basis on which this vision can be realised where mechanisms have been
developed that facilitates the sharing of virtual world objects across different
virtual environment. The work has been tested using a working prototype that
allows digital content, to be shared and physical devices, such as mobile phones,
to be connected and their content to be shared.
1 Introduction
Many devices have developed rapidly to become multifunctional wonders. They
provide more functionality than they were originally designed to perform. The best
example is that of the mobile phone. They not only provide communication functions
such as making a phone call and sending text messages, but also work as a camera,
MP3 player and support web access. More and more devices are providing computing
capabilities, including increased networking functions enabling them to interact with
each other more easily. The multifunctional capabilities of devices have given birth to
exciting new application areas for networked appliances. Using the network, devices
can be controlled from anywhere in the world. In online games such as World of
Warcraft [1] and Second Life [2] players share a virtual environment in order to
communicate, do business, and develop digital objects, which not only involve
personal computer but also games consoles and mobile devices. Although real devices
are used and the user is the only physical entity, these players communicate over large
distance from different geographical locations from all over the world. Users can
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 328–342, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Virtual Environments with Content Sharing 329
generate and share content, and buy and sell it in virtual environment. The challenge is
to a achieve this within and across different heterogeneous virtual environments.
Many ways of distributing content may be possible; for example, between two
physical devices; physical devices and virtual environments; or between different
virtual environments and gaming platforms. Consider a scenario where a mobile user
could share an audio file with a Second Life player. One approach would be to allow the
user to simply ring the player’s virtual mobile phone and send the track. Another
example may be where a game user shares his or her assets and resources during a game
with other potentially different games. This could be done for free or for a small fee.
Here we could see a player in a massive multiplayer online game requesting and using
resources, such as a weapon, or life time from some other player, in order to remain in
the game longer. Here the user benefits from finding and using resources for free or at a
very low price. Conversely, gaming studios benefit because more and more players will
play or join the game because of the increased opportunity to extend game play and
even earn money brings.
In this paper we propose a distributed framework based on a service utilisation
framework which facilitates the sharing of content in virtual environments [3]. Using
this framework networked appliances are automatically created and connected with
associated avatars within a virtual environment. Our approach has many benefits,
which include the user’s ability to share, distribute and sell their content in virtual
environments and the ability to remove the physical constraints associated with real
world objects.
The remainder of the paper is structured as follows. In section II we introduce the
background and the related work. Section III provides an overview of our proposed
framework before describing a case study and a technical description of our approach
in Section IV In Section V we provide our conclusions and future work.
This has lead to shifts within the gaming industry, where increasing access to game
engines, software development kits and level editors has allowed games to be changed
more easily. This phenomenon – known as modding – marginally alleviates some of
the limitations discussed above [9-11]. Although modding provides a means of
adapting and evolving games, it is restricted to more technically savvy users, such as
software developers, rather than people who simply just play games. Furthermore,
mods are tied to specific games. For example, a mod developed for the unreal engine
will be incompatible with the quake engine. Some researchers suggest that distributed
technologies in conjunction with middleware may relieve many of these difficulties,
however it is generally accepted that more research is required to establish a suitable
architecture [12].
Modding is an activity that runs alongside mainstream games development, with
developers providing modding tools as a way to attract customers. In essence modding
is seen as a business strategy. Although not explicitly stated, incentives to mod games
are used as a means of generating free development for publishers, for example through
the use of modding competitions that act as a means of screening game enhancements
in order to include them in future releases. In most cases this is an unpaid source of
labour and gaming organisations carefully control how it is executed [9]. Through
competitions and gaming subscriptions for massively multiplayer online games, the
industry has a healthy flow of mod software. In support of this several game companies
adopt the principle of modding as a key strategy, where only a base solution is initially
provided. Any enhancement to the game thereafter is dependent on user modifications.
One example of this is BioWare’s Neverwinter Nights, which is heavily reliant on
gamer-created content [13]. Successful mods have been incorporated into subsequent
releases. Another example is Counterstrike, which is a modification for team play of
Valve Software’s Half-Life [13]. In this case modding can be seen as an important and
welcome source of innovation where commercial risks are not taken by the gaming
industry, but rely on the goodwill of the modders [10].
Whilst modding has been discussed from a game coding perspective, mods may also
exist as part of and within the game itself. Communities such as Second Life [14] are
heavily reliant on users shaping the virtual environment, extending the concept of
MUDs into realistically rendered virtual worlds [15]. Graphical objects of any
description can be developed and added to the virtual world, which can then be shared
or sold between avatars’ within that world. Modifications to the environment (e.g. land)
can be made and buildings can be constructed. This differs somewhat from
conventional modding in that all modifications take place within the virtual world.
However, there is no mechanism to allow the objects created in Second Life to be
shared and distributed amongst different online games and a better approach could be
used to expose these modifications so that they can be utilised universally.
The increasing popularity amongst multiplayer gaming platforms shows that they
are being used more than for just passing time. The virtual world’s platform is already
being utilised for business but it may also be used for community and financial analysis
of the gaming business itself. Among other uses we see them being used for specialized
training for armed forces or vocational training, medical consultation and
psychoanalysis, and for community and financial experimentation to analyse social
norms [16].
Virtual Environments with Content Sharing 331
The study shown in paper [17], plans to build a virtual world where virtual objects
visualize the information collected in wireless sensor networks by which virtual worlds
may allow peers to understand and estimate more easily the state of the real world
measured through sensor networks.
Many of us now have access to home technologies such as computers, game
consoles, and the internet which can help us to create content. There is however
limitations with current approaches where they fail to investigate how young people
participate in content creation or what tools they use and the extent of their
commitment. Such surveys are required to better identify the potential of content
creation [18].
The research carried out in [6] proposes to examine cognitive overload problems
among game players who always interact with the game world as well as with other
users [6]. Using Maple Story [19] as a case study; the authors have found different
results which show different types of cognitive overloads to come into sight during
game play which might cause serious problems to all players.
This paper [20] introduces Konark, a service discovery and delivery protocol for
ad-hoc, peer-to-peer networks in which the authors provide an infrastructure to set up
generic peer-topeer systems. It acquires advantages of basic networks for peer naming
and message routing. It uses entirely distributed, peer-to-peer techniques for resource
discovery which provides every peer with the ability to publicise and discover the
resources in ad-hoc networks [20].
There are many business opportunities and challenges involved in the virtual world
where millions of people from all over participate in the network to play online games.
This paper
[21] particularly discusses the principles and policies related to the social
implications of Second Life [22] which raises significant research questions. One of the
important questions, for example, is the payment issue either to the avatar or customer,
while another issue is taxing people who are earning money in the virtual business. As
can be seen there is considerable research interest around virtual environments with
each approach making some very interesting contributions. In the remainder of this
paper we discuss our proposed framework that builds on these advances to extend
current approaches where user generated content can be more freely shared across
different virtual environments.
3 Framework Overview
Our approach is based upon the service utilisation framework proposed in [3] and
depicted in Figure 1, we have designed a plug-in containing several services that allow
users to share their resources.
This component provides an interface between applications such as gaming
platforms, virtual worlds and network devices. The Visual Resource Manager, as is
illustrated in Figure 2 extends the functionalities provided by the service utilisation
framework.
332 M. Merabti et al.
also allows descriptions to be extended through the addition of new tags which makes it
efficient. Ontologies, a shared understanding of some domain, are used to promote
better understanding of relationship between the same concepts using different
terminologies.
When we receive an object the Virtual Resource Manager registers the resource with
the resource Lookup Service. Following this it extracts the meta-data of that object and
passes it to the Visualization Engine which in turn renders the 3D object into a graphical
shape. At the same time the Behaviour Matcher lookups up the behaviour of similar
objects in that environment. If it finds the behaviour then it assigns it to the object along
with the effects it supports when it is executed. Let us assume that a user sends a game
object to another game. The game objects behaviours should also be transferred from the
source environment to the destination environment so that the user can fully enjoy the
new object features such as graphical special effects or how it reacts to stimulus from the
game, such as being shot at.
The object consists of two layers; the first layer contains Meta data used to describe
the object as illustrated in Figure 3, which is an XML file containing different attributes
of that object as shown in Figure 8 below.
The second layer is the scripted behaviours. A mapping is performed between the
object and the game engine by extracting the meta data used to describe the object and
its 3D characteristics and the rewriting scripting engine used to find appropriate
behaviours the game engine can accommodate as detailed in the scripted behaviour of
the object.
Using the principles of service-oriented computing, components, such as game
consoles and mobile phones, implement a small footprint of code allowing functions,
such as audio and video, to be disseminated within the network. Using the framework
services, the components can link to the network using any communication protocol;
discover and/or publish and use framework and application services locally (provided
by the component itself) or remotely (provided by other components); carry out
semantic interoperability between different vocabularies used by component
manufacturers; automatically form communication links with other components in the
network; self-manage links with other components in the network; and self-manage
their operations based on composite and environmental changes. Application particular
334 M. Merabti et al.
services, on the other hand offer a means of dispersing and utilising component
functionality (such as audio and video), gaming engines, and player (AI behaviours)
and game objects (tree, car or avatar).
This is achieved using the service integration framework [13], implemented on
every component – be it a networked appliance or software module from the virtual
world. This is a peer-to-peer interface that can be mapped onto any middleware model.
Devices connect to the network as either specialised components or simple
components. A specialised component has the ability to provide services as well as to
propagate service requests within the network. A simple component by comparison has
more restricted abilities: it joins the network, propagates queries and invokes
discovered services. For example, this could be sensors in a network that provide
multimedia data for crowds or flocking. This enables any component irrespective of its
capabilities to effectively choose how it will interact within the network.
Using this architecture, we have designed a distributed service-oriented platform for
use with virtual environments and physical devices. This allows multimedia content to
be shared with the virtual environment from any physical multimedia producing device
such as mobile which we discuss in more detail below.
Whilst it is important to bear in mind the overall structure that a virtual environment
might take, it has been a goal of our work to deconstruct as far as possible the holistic
notion of a virtual environment into a set of autonomous, generalised and reusable
components. Whilst the development process of our framework necessarily entailed the
compartmentalisation of various aspects of a traditional game, the final result must
therefore be considered from the opposite perspective. Ultimately we aim to allow
gaming to exist as an ad hoc interaction between various networked components, the
entirety of which forms the virtual environment. None of these components in isolation
can be considered to be the virtual environment itself. Perhaps the closes to what might
be considered the heart of the virtual environment might be the rendering or physics
engines. However, these will only provide one of any number of interpretations of the
interactions that occur between components.
Various parts of Homura are declaratively specified in XML files in the root of a
Homura project, and this provides a link between the classes and concepts used in a
Homura project. These XML files come in useful when considering exporting to a
website, as the website can parse these files in a standard way, and act upon the data
contained within them. The IDE can also act upon, and manipulate these files to change
various parts of a Homura project. This is similar to how Eclipse works with the plugin
XML that sits in the root of each RCP project, allowing concepts to be linked to classes,
and functionality exposed to other plugins.
The IDE itself hooks into the running Homura engine while a Homura application or
game is running to provide various introspection and debugging facilities. For example,
the user is able to see details about the concepts which are in operation within the
application, as well as the current frame-rate through a statistics view. One feature that
is particularly important for the user to inspect is the scene graph, as this allows them to
find out why their graph is not correct and help them find the area of code which is
manipulating it incorrectly. They are also able to view the various properties of the
scene elements they can select, for example, the world and local translation of a node.
In order to provide the necessary hooks into the running engine, some parts of jME
have been modified, as modifying Homura alone may not be enough, or too inefficient
at certain levels. For instance, it may be difficult to tell when a scene graph has changed
if it is being modified programmatically through the user's own code. Other parts of the
engine can be probed at intervals to check their status.
Parts of the Homura IDE use Homura and jME, not just the games and applications
that the user creates. However, if Homura is providing a game interface where parts of
the API are accessed in a game-like context, the method of integrating parts of this
engine into the IDE will be less than ideal. At the moment, jMonkeyEngine has a
game-specific context. The idea is to unravel this into a hierarchy of non-game specific
classes and interfaces, with the game-specific classes and interfaces at the top of this,
with the notion of being able to run an 'application' and a 'game'. Then, only the
game-specific classes will have access to game-specific concepts. Therefore, it will be
necessary to keep the game-specific details separate from the application-specific ones.
5 Case Study
A case study has been conducted to demonstrate our approach that shows one way of
sharing content. We have developed our scenes using blender. For example, Figure 8
shows and XML representation of a gun. In order to load these XML files into the scene
graph of jME the function illustrated in Figure 5 is used.
jME doesn’t support loading any file formats directly. Rather it uses jME binary, its
own format. Different classes included in jME convert scripts are used to create the
jME binaries. First of all the binary converter and binary reader is loaded as it is
illustrated in Figure 5 from lines 4 to 7. As illustrated in line 10 and 12 the
OutputStream and the InputStream are used to send and read the appropriate contents.
The XML file is converted with the ByteArrayOutputStream and read with the
ByteArrayInputStream. This process allows us to transfer meta data from one
environment to another environment as shown in figure 6.
Using our framework we have implemented an application that links a mobile phone
to a corresponding avatar in the jME as depicted in Figure 7. Through this connection
we are able to us the physical phone and its avatar representation in the virtual world.
We can answer and make calls from the physical and the virtual and using the same
communication channels we are able to transfer user generated content between the
two. For example, we can pass a music track to the virtual world along with metadata
describing what it looks like and the behaviours it supports.
In Figure 7 we see a mobile phone and for each song passed from the physical device
to the virtual mobile phone a radio button is added and visually connected to the avatar
phone. When a song is selected the behaviours appears in the virtual world; in this case
we can see that by selecting Song1 we can execute one of its behaviours, i.e. exit, play,
and stop, by pressing the buttons located to the right of the virtual world screen. In the
following section we discuss the technical aspects in achieving this.
338 M. Merabti et al.
6 Technical Description
In the above case study a peer makes a request for a service, such as a game object
(e.g. gun), in the virtual lab where another peer has previously advertised its sharable
assets using JXTA advertising services. In Figure 8 we show, how we have
implemented the scenario in which the user requests a gun resource.
We pass the meta-data to jME, which in turn is used to render a 3D representation of
the gun in the scene. The gun object also contains the scripting behaviours it supports.
For example, Figure 9 illustrates, in part a simple script for the fire behaviour.
Javascript was used and where developed using the Rhino API from Mozilla [25],
which is used with the Java Scripting Framework [26] and the open-content repository
API provided by Captive Imagination [27].
Fig. 9. Sample behaviour Fig. 10. Rules used to create scripted behaviour
The Meta data and scripts, including the aforementioned tools where applied in the
same way to allow music to be shared between our mobile phone and its associated
avatar in the virtual lab. The goal here is to show how two very different types of
content can be shared. One associated with multimedia and the other associated with
conventional game playing objects. Perhaps these act as two extremes between which
many other possibilities are possible.
Both meta data for objects and the scripted behaviours are passed between different
environments using JXTA pipe and messaging objects in which all required
information is presented to extract and construct the associated object. Whilst, we
simply use the meta data to construct the objects, we run all scripting behaviours
through a set of rules, as discussed in the above section on the Behaviour Matcher.
Whilst objects may support behaviours in their source environment, it is not necessarily
the case in the target environment. Here the rules try to extract the behaviours the target
environment supports. The Behaviour Ontology acts as an interoperability mechanisms
between terminologies which we have implemented and serialised using the Web
Ontology Language (OWL) [28]. The rules where developed using Drools, where
Figure 9 shows in part a simple fire and part of the script for a behaviour being
generated.
In the mobile phone scenario we demonstrated how two users are able to share
multimedia content between physical mobile device and corresponding avatars in the
virtual world. We stream multimedia content from the physical mobile device to the
virtual mobile using the Java Media Framework (JMF) [29] and the Real-time
Transmission Protocol (RTP) [30].
340 M. Merabti et al.
RTP packets were wrapped in JXTA [23] messaging objects to abstract the IP
dependent format used for HTTP calls in RTP. This provides a unified addressing
scheme ensuring that all components are addressed in a uniform way. Frames were sent
from the physical mobile to the virtual environment using JXTA Pipes. Upon receiving
the JXTA packets, the RTP packets are extracted and processes by a custom data source
adapter developed for the purpose, which streams RTP data much as it is done
traditionally, after network connectivity , the avatar requests the list of song in which
he/she then chose a song from the list, after that the mobile start streaming the content
using RTP protocol in JXTA pipes; when it receives first stream JMF process the
stream and check supported codec, if it is supported then it will continue receiving
streams from mobile device in our case while playing the stream using JMF and a
plugin called Fobs4JMF [31] which supports most formats such as mp4 or 3gp. These
tools in conjunction with Skype allowed us to enable bi-directional communications
between the physical mobile phone and its virtual world counterpart.
content sharing services, the Homura engine and IDE and the Drools Rules and Java
Scripting Framework. This will be the focus of much ongoing work. Ultimately, the
success of a framework such as this relies on the development of exciting content that
can be used to build up gaming environments. Nonetheless we believe that a flexible
and distributed system such as this provides many opportunities for the advancement of
gaming, virtual environments and networked devices in the physical world into new
areas and in new ways.
References
[1] Ducheneaut, N., Yee, N., Nickell, E., Moore, R.: Building an MMO with Mass Appeal: A
look at Gameplay in World of Warcraft. Games and Culture - Journal of Interactive
Media 1(4) (2006)
[2] Herman, H., Coombe, R.J., Lewis, K.: Your Second Life? Cultural Studies 20(2-3) (2006)
[3] Merabti, M., Fergus, P., Abuelma’atti, O., Yu, H., Judice, C.: Managing Distributed
Networked Appliances in Home Networks. Proceedings of the IEEE Journal (2008)
[4] Seay, A.F., Jerome, W.J., Lee, K.S., Kraut, R.E.: Project massive: a study of online gaming
communities. In: CHI 2004 extended abstracts on Human factors in computing systems,
pp. 1421–1424. ACM Press, Vienna (2004)
[5] PlanetSide (2006) (Accessed 2007), http://planetside.station.sony.com/
[6] Star Wars Galaxies (2006) (Accessed 2007), http://starwarsgalaxies.station.sony.com/
[7] The Sims Online (2006) (Accessed 2007),
http://www.ea.com/official/thesims/thesimsonline/
[8] EVE Online (2006) (Accessed 2007), http://www.eve-online.com/
[9] Sotamma, O.: Have Fun Working with Our Product!: Critical Perspectives on Computer
Game Mod Competitions. In: International DiGRA Conference, Vancouver, Canada
(2005)
[10] Kucklich, J.: Precarious Playbour: Modders and the Digital Games Industry. International
Journal on Fibreculture 1(5) (2005)
[11] El-Nasr, M.S., Smith, B.K.: Learning Through Game Modding. ACM Computers in
Entertainment 4(1) (2006)
[12] Hsiao, T., Yuan, S.: Practical Middleware for Massively Multiplayer Online Games. IEEE
Internet Computing 9(5), 47–54 (2005)
[13] Computer Game Modding, Intermediality and Participatory Culture, F. University of
Tampere (2003) (Accessed: September 2003),
http://old.imv.au.dk/eng/academic/pdf_files/Sotamaa.pdf
[14] Yee, N.: The Unbearable Likeness of Being Digital: The Persistence of Nonverbal Social
Norms in Online Virtual Environments. The Journal on CyberPsychology and Behaviour
(to appear, 2006)
[15] Curtis, P., Nichols, D.A.: MUDs grow up: social virtual reality in the real world. In:
COMPCON 1994, pp. 193–200. IEEE Computer. Soc. Press, Los Alamitos (1994)
[16] Book, B.: Moving Beyond the Game: Social Virtual Worlds. Series Moving Beyond the
Game: Social Virtual Worlds (2004)
[17] Kwon, T., Choi, S.-M.: Deriving the Virtual World from Wireless Sensor Networks for
Interaction with Consumer Electronic Devices. Consumer Electronics. In: ICCE 2007
(2007)
[18] Hayes, E.: Game content next term creation and it proficiency: An exploratory study
(2007)
[19] Maple Story (Accessed 2007), http://maplestory.nexon.net/
342 M. Merabti et al.
[20] Desai, N., Verma, V., Helal, S.: Infrastructure for Peer-to-Peer Applications in Ad-Hoc
Networks (Accessed 2007), http://www.harris.cise.ufl.edu/projects/
publications/konark_p2p.pdf
[21] Papagiannidis, S., Bourlakis, M., Li, F.: Making real money in virtual worlds: MMORPGs
and emerging business opportunities, challenges and ethical implications in metaverses.
Technological Forecasting and Social Change (2007)
[22] Second Life (Accessed 2007), http://www.secondlife.com
[23] JXTA (Accessed 2007), https://jxta.dev.java.net
[24] Java Monkey Engine User Guide (Accessed 2007),
http://www.jmonkeyengine.com
[25] Rhino: JavaScript for Java. 2007, Mozilla.org (Accessed 2007),
http://www.mozilla.org/rhino/
[26] O’Connor, J.: Scripting Framework for Java (2006) (Accessed 2007),
http://eventhorizongames.com/wiki/
doku.php?id=articles:java_scripting_framework
[27] The open-content repository for games. Captive Imagination (Accessed 2007),
http://captiveimagination.com/svn/public/cigame/trunk/
[28] Berners-Lee, T., Hendler, J., Lassila, O.: The Semantic Web, vol. 284(5). Scientific
America (2001)
[29] Java Media Framework (JMF) (Accessed 2007),
http://java.sun.com/products/java-media/jmf/
[30] RTP: A Transport Protocol for Real-Time Applications (Accessed 2007),
http://www.ietf.org/rfc/rfc1889.txt
[31] RTP: A Transport Protocol for Real-Time Applications 1996, IETF (Accessed 2007),
http://www.ietf.org/rfc/rfc1889.txt
Hand Contour Tracking Using Condensation and
Partitioned Sampling
1 Introduction
Visual articulated hand tracking is an attractive area of research in the computer
vision community. It has a great potential in VR, AR, HCI and computer games.
However, the high degree of freedom of hand configuration, the self-occlusion of
fingers, and the kinematic singularities in the fingers’ articulated motion, make visual
articulated hand tracking become a difficult and challenging problem.
Because of the wide range of potential applications and the technical challenges,
visual hand tracking has attracted a great deal of researches onto it during the last
decade. In 1993, Rehg builds a system called DigitEyes [1] where a hand could be
tracked against a black background. In the system, a kinematic 3D hand model whose
initial configuration is known is used to represent human hand. The state of the model
is updated by solving a constrained nonlinear optimization problem. In addition, self-
occlusions of the fingers are handled using layered templates. Rehg’s work is seminal
since it establishes a classical approach for model based tracking of articulated hand,
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 343–352, 2008.
© Springer-Verlag Berlin Heidelberg 2008
344 D. Zhou, Y. Wang, and X. Chen
however, the adaptation of the kinematic 3D hand model to a new user is time-
consuming, and finger occlusions are only tracked off-line in the system, what’s
worse, the optimization process may get trapped in local minima which could result in
the tracking failed. In another system similar to Rehg’s, Kuch and Huang [2] simplify
the model adaptation to a new user with only three snapshots of the user’s hand in
three predefined configurations. Though making some improvements to the DigitEyes
system, their new system still has the local minima problem. In 2001, Stenger et al.
[3] construct an anatomically accurate hand model using truncated quadric, and an
unscented kalman filter is used to estimate the pose of the hand. In 2006 [4], Stenger
et al. improve their approach by first discretizing the state space and organizing it into
a hierarchy of hand templates, and then searching down the template hierarchy to
refine the fitting of the hand model to the hand in an image. Their method produces
good results, and it is capable of handling out-of-image-plane rotations, fast motion,
and automatic recovery of tracking. However, the system has large memory
requirements and does not work in real-time.
In 1998, Blake and Isard [5] establish another important kind of hand tracking
approach based on deformable 2D contours and particle filter. In the paper, they
introduce the Condensation algorithm and give a number of experiments to
demonstrate the robustness of their method. However, it’s inefficient to track an
articulated object which has a high dimension state space using Condensation alone.
In 1999, MacCormick and Blake [6] introduce a new technique called partition
sampling, which makes possible to track more than one object. A year later,
MacCormick and Isard [7] implement a vision based articulated hand tracker using
this technique. Their tracker is able to track position, rotation and scale of the user’s
hand while maintaining a pointing gesture. Based on Blake’s work, Martin Tosas [8]
implement a full articulated hand tracker which is then used for Virtual Touch Screen
in his thesis. He makes some technique extensions to Blake and MacCormick’s
methods.
Besides the approaches mentioned above, the following are some other kinds of
hand tracking methods. Nolker and Ritter’s algorithm [9] firstly find the fingertips of
a hand in a grey-scale image by means of a hierarchical neural network, and then
update a 3D hand model from the fingertips positions. Stefanov [10] combines Hough
transform features with behavior knowledge in order to guide and achieve robust hand
tracking. Kolsch and Turk [11] use a technique called flocks of features to track a
hand through rapid deformations against complicated background.
Our implementation is based on Martin’s work [8], while some changes are made:
the skin extracting method is simpler, the measurement operation is more effective,
the calculation of the weight for a complete particle is new and reasonable, and
instead of particle interpolation used in Martin Tosas [8], we use a simpler selection
strategy to form complete particles.
The rest of the paper is organized as follows. Section 2 introduces briefly a simple
but effective skin-color extracting method. Section 3 gives out a deformable hand
contour template and a non-linear state space which represents the movements of the
hand contour. In Section 4, we describe how to track in real-time the hand contour
throughout a video sequence. Section 5 shows some experimental results. And
Section 6 is a short summarization.
Hand Contour Tracking Using Condensation and Partitioned Sampling 345
Our tracker is skin-color based. Before tracking performs, we need to extract skin-
color from video. In this paper, we use a single 2D Gaussian on Cb-Cr space to model
skin-color. The mean vector and covariance matrix of the Gaussian model are
estimated in advance using some well-chosen training data. To extract skin color in a
24-bit image, some pre-process are performed at first, then each pixel is transformed
into YCbCr space, and the CbCr sub-vector is tested against the 2D Gaussian model
to decide whether the pixel belongs to skin color or not, only the pixels belong to skin
color kept. Though the model is simple, the result it produces seems good and
satisfies our application. Figure 1 shows the result of this method.
Fig. 1. Skin-color extracting using a single 2D Gaussian model. The left is a common image
(RGB), the right is the corresponding skin-color image.
constant length. This is because the thumb is assumed to flex only on the same plane
as the palm. Since we assume that the user always keeps his palm approximately
parallel to the image plane when moving in front of an ordinary camera, the hand
contour model seems suitable for our tracking task. Besides the 50 control points,
there are 6 pivots and a palm centre around which the whole hand rotates should be
chosen. Figure 2 shows the deformable hand contour template.
Fig. 2. Hand contour template showing its 50 control points, 6 pivots and 1 palm centre
Since we assume the user always keeps his palm approximately parallel to the
image plane, the contour of his hand can only perform three kinds of movements. The
first is rigid movement of the whole contour, including translation, rotation and
scaling, the second is rotation of each finger’s contour around their corresponding
pivot, and the third is flexion/extension of the contour of each finger except for the
thumb since we assume its two segments both keep a constant length. In conclusion,
the movements of the hand contour can be represented by the following 14
dimensions state vector:
χ = (x, y, α, λ, θL, lL, θR, lR, θM, lM, θI, lI, θTh1, θTh2). (2)
This state vector is partitioned into 7 parts as follows, the first part is the sub-vector
(x, y, α, λ), which is a non-linear representation of a Euclidean similarity transform
applied to the whole hand template. The second part is the sub-vector (θL, lL) which
represents the non-rigid movement of the little finger, θL means the little finger’s
angle with respect to the palm, and lL means the little finger’s length relative to its
original length in the hand template. The following three parts are (θR, lR), (θM, lM),
and (θI, lI), they all have the same explanation as the sub-vector (θL, lL), but for the
ring finger, the middle finger and the index finger respectively. The sixth part is the
parameter θTh1, which represents the angle of the first segment of the thumb with
respect to the palm. And the last part is θTh2, which represents the angle of the second
segment of the thumb with respect to the first segment of the thumb. It is noticeable
that the finger and thumb angles are 0 when they are in the template position, and that
the length of the fingers is relative to the template’s finger length. So it’s 1 when they
have the same length as in the template. In our implementation, some constraints as
follows are imposed on the finger lengths and finger angles, the minimum allowed
length is 0.3 and the maximum is 1.2, and the finger angles must satisfy the following
inequality: -0.20 ≤ θL ≤ 0.45, -0.25 ≤ θR ≤ 0.25, -0.25 ≤ θM ≤ 0.25, -0.45 ≤ θI ≤ 0.20, -
0.45 ≤ θTh1 ≤ 0.90, -0.10 ≤ θTh2 ≤ 0.25.
Hand Contour Tracking Using Condensation and Partitioned Sampling 347
pt ( zt | xt ) ∫ pt ( xt | xt −1 ) pt −1 ( xt −1 | Z t −1 )dxt −1
pt ( xt | Z t ) = . (4)
pt ( zt )
Since pt(zt) is generally a constant independent of xt for a given image, it can be
neglected so that only relative likelihoods need to be considered:
pt ( xt | Z t ) ∝ pt ( zt | xt ) ∫ pt ( xt | xt −1 ) pt −1 ( xt −1 | Z t −1 )dxt −1 . (5)
Formula (5) suggests that the conditional probability of the hand configuration at
time t can be approximated as a sum of the previous conditional probabilities of the
hand configuration multiplied by the dynamics, all weighted by the observation
density.
In Condensation, an important notation called weighted particle set is used to
represent a complicated probability distribution. A weighted particle set is a list of n
348 D. Zhou, Y. Wang, and X. Chen
pairs (xi, πi) (i = 1, ... n) drawn from a probability density, where xi belongs to a
∈
certain state space, and πi [0,1] is a weight on xi proportional to the value of the
density function at xi, with Σni=1πi = 1. Generally, any probability density can be
represented by a particle set, no matter it is unimodal or multimodal.
In the context of hand contour tracking, since the posterior pt(xt | Zt) is usually too
complex to be evaluated simply in closed form, we use a particle set to approximate
it. Each particle drawn from the posterior comprises a state vector xi which represents
a hypothesized contour configuration, and a weight πi which is the likelihood of the
hypothesized contour representing the real target object. For the tracking task, we
need to evolve in time the conditional density pt(xt | Zt), this is done in Condensation
by propagating a particle set for the posterior from one time-step to the next. The
propagation process in the Condensation algorithm consists of three operations called
resampling, prediction and measurement, the following is a briefly description about
them.
4.1.2 Measurement
After prediction, we obtain some new particles whose states are known, but do not
have a weight yet. In order to calculate their weight, which is the likelihood of a
particle (hypothesized contour) representing the real hand contour, the measurement
operation is applied.
Given a new particle whose state vector is known, its weight can be calculated by
the observation density pt(zt | xt). Though there are various forms of observation
density to be chosen, the one introduced by Blake and Isard [5] are simple and
effective so we adopt it in our implementation, but make some changes to it.
As do in Blake and Isard’s approach, several line segments normal to the contour
represented by a new particle at some well chosen measurement points are used to
calculate a weight for the new particle. These line segments, termed measurement
lines, are processed in order to find image features along them. Each of the found
features has a contribution towards the weight of the measured particle. To quantify
this contribution, we use a single Gaussian centred on the measurement point to
model the features along each measurement line, as Blake [5] introduced. Sometimes,
more than one detected features appear along a certain measurement line. For
efficiency, only the one nearest from the measurement point is considered.
Currently, to measure a hypothesized contour, we only need to measure along each
chosen normal line using a single Gaussian and multiply the results. By simplifying,
the weight for a measured particle can be computed using the following formula:
M
1
π ∝ p ( z t | x t ) ∝ exp{ − ∑ f ( z1 ( s m ) − r ( s m ); μ )} . (9)
i =1 2σ 2
Fig. 3. Measurement lines: The left shows Blake’s method, all measurement lines have the
same length. The right shows our method, different measurement lines have different length.
As proved in [7], more dimensions means more particles for Condensation, rendering
the Condensation filter ineffective. So in our tracker, where the state space is too large
to be dealt with a single Condensation filter, another technique called partitioned
sampling is adopted to improve the efficiency.
In section 3, the state space is partitioned into 7 parts similar to [8].In this section,
each of the 7 parts will have a corresponding sub-particle set, called Palm, L, R, M, I,
T1 and T2 sub-particle set. Based on this decomposition, the partitioned sampling is
performed hierarchically on the particle set to be evolved at each time-step, as
follows:
1. For the Palm particle set do:
1.1 Use each complete particle of the previous time-step to generate a number
(250) of new particles for Palm, proportional to the weight of the used
complete particle (Resample).
1.2 Apply dynamics to each particle in Palm and then measure them.
1.3 Select particles from Palm that constitute peaks of weight (top 10%).
2. For each of the finger particle sets, i.e. L, R, M, I, and T1 do:
2.1 Use each of the selected particles in Palm to generate a number (100) of new
particles in the finger particle set, proportional to the weight of the used
Palm particle (Resample).
2.2 Apply dynamics to each newly generated finger particle and measure them.
2.3 Group the newly predicted finger particle set based on which Palm particle
they are generated from, and then select the particles with the highest weight
from each group.
3. For the second thumb segment particle set, i.e. T2 do:
3.1 Use the select particles in T1 to generate a number (100) of new particles in the
T2, proportional to the weight of the used particle in T1 (Resample).
3.2 Apply dynamics to each particle in T2 and then measure them.
3.3 Group the new T2 particle set based on which T1 particle they are generated
from, and then select the particles with the highest weight in each group.
4. Form complete particles for the whole hand contour, which will be used in the
next time-step, and choose the one with the highest weight as the tracking result.
Hand Contour Tracking Using Condensation and Partitioned Sampling 351
A complete particle is a particle for the whole state vector. It contains 7 sub-
particles, one for each of the 7 parts of the state vector, where the L, R, M, I and T1
sub-particle are generated from the Palm sub-particle, and the T2 sub-particle is
generated from the T1 particle.
In Martin Tosas [8], after step 2.2 and 3.2, a technique called particle interpolation
is used so that it is convenient for forming complete particles in step 4. Instead of
particle interpolation, we simply choose the best finger particles from each sub-
particle set generated from the same “parent”. Obviously, this is more effective, and
experiments show that the result is satisfactory.
After step 4, we obtain several complete particles. To measure each of them as a
whole, the following formula is used:
W=(2WPalm /8)×(WL /8)×(WR /8)×(WM /8)×(WI /8)×(WT1/8×)×(WT2/8). (10)
where W is the weight for a complete particle, and WPalm, WL, WR, WM, WI, WT1, WT2
are weights for the seven sub-particles respectively. Since the Palm sub-particle set is
the first hierarchy to search, the generation of the rest sub-particle set is based on it,
we assign a larger proportion to the Palm particle seems reasonable, experiments also
indicate this is right.
5 Experiments Results
Our tracker is able to track in real-time (30 frames per second) an articulated hand
contour on a current common PC (P4 1.7G CPU and 512M memory) equipped with
an ordinary camera (the image size is 320×240), both rigid motion and non-rigid
motion can be tracked robustly, it performs well even in a clutter background. Figure
4 are some snapshots of our experiment.
6 Summarization
In this paper, a robust articulated hand contour tracker is introduced. The tracker
works in real-time against a background without special requirements, it is able to
352 D. Zhou, Y. Wang, and X. Chen
recover from miss-tracking automatically in some situations. The last three pictures in
figure 5 indicate that it’s not sensitive to interference, since it performs well even
another hand put in front of the tracked hand. As the hand contour is modeled by B-
spline curve, it is able to adapt most people’s hand. In conclusion, the tracker
performs well when tracking both rigid and non-rigid motion of anyone’s hand
contour.
For our tracker, the only used clue is skin-color. Other image features such as
edges and frequencies are useful but neglected in our implementation. The
configuration constraints between fingers are also useful information. In future work,
these clues may be considered to improve the performance.
References
1. Rehg, J., Kanade, T.: Digiteyes: Vision-based hand tracking. Technical Report CMU-CS-
93-220, Carnegie Mellon Univ. School of Comp. Sci (1993)
2. Kuch, J., Huang, T.: Vision-based hand modeling and tracking for virtual teleconferencing
and telecollaboration. In: Proc. IEEE int. Conf. Computer Vision, pp. 666–671 (1995)
3. Stenger, B., Mendonca, P., Cipolla, R.: Model-Based 3D Tracking of an Articulated Hand.
CVPR II, 310–315 (2001)
4. Stenger, B., Arasanathan, T., Torr, P., Cipolla, R.: Model-Based Hand Tracking Using a
Hierarchical Bayesian Filter. PAMI 28(9), 1372–1384 (2006)
5. Isard, M., Blake, A.: Condensation–conditional density propagation for visual tracking.
Int. J. Computer Vision 29, 5–28 (1998)
6. MacCormick, J., Blake, A.: A probability exclusion principle for tracking multiple objects.
In: Proc. 7th International Conf. Computer Vision, pp. 572–578 (1999)
7. MacCormick, J., Isard, M.: Partitioned sampling, articulated objects, and interface-quality
hand tracking. In: European Conf. Computer Vision (2000)
8. Tosas, M.: Visual Articulated Hand Tracking for Interactive Surfaces. PhD thesis,
University of Nottingham (2006)
9. Nolker, C., Ritter, H.: GREFIT: Visual Recognition of Hand Postures. In: Proc. of the
International Gesture Workshop, pp. 61–72 (1999)
10. Stefanov, N., Galata, A., Hubbold, R.: Real-time hand tracking with Variable-length
Markov Models of behaviour. In: IEEE Int. Workshop on Vision for Human-Computer
Interaction (V4HCI), in conjunction with CVPR (2005)
11. Kolsch, M., Turk, M.: Hand tracking with Flocks of Features. Computer Vision and
Pattern Recognition, CVPR 2, 20–25 (2005)
12. Blake, A., Isard, M.: Active Contours. Springer, Heidelberg (1998)
Integrating Gesture Recognition in Airplane Seats
for In-Flight Entertainment
Rick van de Westelaken, Jun Hu, Hao Liu, and Matthias Rauterberg
Abstract. In order to reduce both the psychological and physical stress in air
travel, sensors are integrated in airplane seats to detect the gestures as input for
in-flight entertainment systems. The content provided by the entertainment
systems helps to reduce the psychological stress, and the gesture recognition is
used as input for the content and hence stimulates people to move, which as a
result would reduce the physical stress as well.
1 Introduction
Today a large number of people use air travel as a way of transportation. This number
is increasing every year, as well as the number of long-haul flights. At the same time
the flight duration is increased because of better fuel efficient airplanes, which makes
intermediate landings unnecessary. The air travel market is highly competitive and
therefore airlines try to maximize the number of seats [1]. Often this results in a very
limited amount of seating space for passengers, especially in economy class [2]. In
this context the EU project “SEAT” (Smart tEchnologies for stress free Air Travel)
was set-up [3]. This project, sponsored by a European commission focuses on
improvements and solutions to increase comfort in air travel. The partners in this
project are Imperial College London, Acusttel, Aitex, Antecuir, Czech Technical
University, Wearable Computing Lab ETH Zürich, Inescop, Queen Mary University
of London, Starlab, Eindhoven University of Technology, Thales and DHS. The work
described here is part of the SEAT project and is executed in Eindhoven University of
Technology.
The combination of long flight duration, limited space and an unusual cabin
environment in terms of air pressure, humidity and continues noise causes physical
and psychological discomfort for a large group of passengers as well as the crew [4].
Most of the airlines offer in-flight entertainment to their passengers. This provides
mental distraction and might lead to reduction of psychological stress. To reduce
physical stress some airlines recommend passengers in-flight exercises[2, 5]. The
ways these exercises are presented vary from paper versions to complete instruction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 353–360, 2008.
© Springer-Verlag Berlin Heidelberg 2008
354 R. van de Westelaken et al.
videos and crew demonstrations. The main goal of these exercises is to stimulate the
blood flow and prevent health related issues like deep vein thrombosis (DVT),
stiffness and fatigue. The problem with these recommendations is that it depends on
the passenger whether the exercises are executed. Another way to prevent DVT is to
wear compression stocking to prevent formation of DVT [6]. Ordinary this is only
used by passengers with increased risk of DVT, which means that only a relative
small group of people use this means.
Physical stress can cause a number of health related problems, as shown by
different institutes like the World Health Organization [7] and the British Medical
Association [8]. Therefore this work focuses on how passengers can be motivated to
move during long-haul flights in order to reduce physical stress. The solution
discussed here is an airplane chair equipped with sensors to detect the body
movements and gestures of passengers. The detected gestures could than be used as
input for interactive applications in in-flight entertainment systems.
The idea of controlling interactive content by means of body movement is not new.
In contrary it is very popular at the moment. Examples of movement controlled game
platforms are the Nintendo Wii and Sony Eye-Toy. Although there is a big difference
between these platforms, these platforms assume that the user has enough space in
which they can move. This assumption does not hold in an economy class flight
environment. There the amount of space is very limited, which means that new
solutions had to be explored.
to their passengers which can be performed during the flight. Research proved that the
most effective exercises to stimulate circulation and thereby reduce discomfort,
fatigue, stiffness and DVT. Research has been done to evaluate recommended lower
leg exercises by airlines to investigate which exercises induce optimum calf muscle
pump activity[5], and these exercise provide important information as starting point
for a solution (Fig. 1).
3 Concepts
With the problem and the context as starting points, an idea generation session was
held with four industrial designers. The most promising ideas generated during this
session are shortly described below. A remarkable detail is that all ideas except one
are based on games because games were considered to be typical interactive content
that motivates people to take active roles [11].
1. Quiz game; Answering questions happens by body movement.
2. Drive game; with physical steering wheel and pedals.
3. Agility game; balancing digital objects by physical movement.
4. Copy game; try to copy movements of others as well as possible.
5. Movement of passengers will be triggered by changes somewhere in the cabin.
A benchmark based on a Pugh matrix (also called a decision matrix) has done to
find the most promising idea. A Pugh matrix is a way to assess ideas based on pre
selected criteria and matching weight factors. The result of this matrix should show
the most promising idea(s) according to the set criteria. The used assessment criteria
from most important to less important are:
1. Usability of basic principle; can the basic principle behind the idea be used for a
wide variety of games / exercises or is it only applicable for a small range of games
and exercises.
Integrating Gesture Recognition in Airplane Seats for In-Flight Entertainment 357
2. Feasibility within space; to what extend might the limited amount of space in the
plane’ chair lead to problems in executing the exercises.
3. Movement intensity; to what extend do passengers have to move concerning the
intensity and repetitions.
4. Disturbance of other passengers; Does the idea lead to disturbance or annoyance of
other passenger.
The benchmark showed an almost equal score for a few ideas. The similarity among
these ideas was that it contained digital content that is controlled by physical movement.
The difference on the other hand is the digital content used for every idea. This means
that a generic interaction platform could be a solid basis for a width range of
applications. The differences would be made by different content. With this assumption
the project continued with finding solutions on how to translate physical movements or
gestures to digital information that could be used as input for digital content.
Solutions to measure movements or gestures can be found in two directions: with
attached sensors or detached sensors.
Attached sensors
A set of first solutions shown in figure 3 are based on attached sensors (sensors that
need to be attached to the passenger). The attached sensors are therefore in most cases
loose objects. This introduces some feasibility problems, which can be summarized
by three keywords: Safety, effort and dependency.
Safety: Whether this are written rules or not, it is logical that loose objects will cause
serious safety issues in certain circumstance. These circumstances could be turbulence
or passengers who are diminished responsible that use it for other purposes.
Effort: In case that a technological solution is found in a loose (embodied) object the
chance on success smaller. The user has to put in extra effort in order to use the
system. For some users this threshold is too high for tem to use the system.
Dependency: Regular objects can be used to obtain data from the user. For example
the standard headphones could be equipped with sensors to measure a passenger’s
head and neck movement. The problem with this is that a percentage of the
passengers won’t use this headphone or will bring their own. This results in
uncertainty about usage of the objects, which could lead to failure of the system.
The usage of separate attached objects is not desired and therefore ideally the
solution should be fully integrated in the airplane.
Detached sensors
A number of technical solutions with detached sensors were considered. These
solutions are described below.
Video-based gesture recognition: This way of gesture recognition is not ideal in the
context of an economy class airplane. The biggest problem lies in the fact that the
amount of free space is very limited. When the camera(s) is placed too close to the
passenger there is the problem of not being able to cover enough of the passenger’s
body.
Sensors integrated in the floor: With sensors integrated in the plane’s floor it is
possible to measure the angle and distance of a passenger’s feet in relation to the
floor. With this data the user’s lower body position can be calculated. This could be
quite accurate because human anatomy and the usable movement space are known.
However it would become integrated in the plain which makes adaptations or repairs
very difficult and costly.
Sensors integrated in the chair: By a grid of pressure sensors integrated in an airplane
chair it is possible to measure a person’s weight distribution. When there are changes
in this weight distribution it is possible to derive a person’s position or gesture out of
this data.
The last suggestion seems to be the most promising solution. However the question
is how many sensors are needed to obtain a certain accuracy level and how many
gestures can be recognized. Since the main goal is to stimulate people to move,
motivating people to move is the more important than knowing exactly what
movements are made. The system might work while it can only recognize a few
different gestures. Therefore a system with a relative low accuracy might be
sufficient. However this assumption needs to be further validated.
important part because there is always contact between this surface and the
passengers. At the same time the weight changes in this surface are relative large
compared to weight changes in the back area. This does not mean that sensors in the
back surface are useless. This could be a good enrichment of the other data obtained
from the seating surface. The obtained sensor values are processed and by means of
pattern recognition the corresponding gestures are derived. When the gesture is
determined this information can be used as input for the in-flight entertainment
systems.
To validate the concept and to prove its effectiveness a prototype was build. Starting
point for this prototype is a chair made out of MDF with similar dimensions of an
average airplane chair. The seating surface of this chair is eventually equipped with a
grid of 28 FSR’s (Force Sensing Resistor). In between the FSR and the user a four
millimeter thick rubber plate is placed. On top of this rubber a round plate with a
diameter of 40 millimeter is placed. This top plate is used to increase the surface that
press on the FSR. The rubber plate is used to equally divide the weight on the FSR. The
grid of sensors is connected with a 4 to 16 line decoder to a microcontroller. The data is
send to a computer by means of a serial data connection. It takes about 50 ms. to read all
the sensor values and to send them to a computer. The sensor values are fed into a
neural network and the output of this neural network is the corresponding gesture.
For implementing the neural network a software package “JOONE” was used.
JOONE (Java Object Oriented Neural Engine) is a Java based application to simulate
neural networks and to train them [12, 13]. The neural network used for the prototype
is a three-layer sigmoid network that consists out of 28 input neurons. In-between the
input and output layer there is one hidden layer consisting out of 30 neurons. The
output layer, which has nine neurons represent the different gestures.
The prototype is able to recognize nine gestures. A first test application had been
created to test the principle of a gesture controlled game, where users have to balance
a ball on a tray by lifting their legs was created. This game showed that a capability of
only detecting three gestures (lifted right leg, lifted left leg, both legs down) already
worked quite well.
References
1. Quigley, C., et al.: Anthropometric Study to Update Minimum Aircraft Seating Standards,
EC1270, prepared for Joint Aviation Authorities. ICE Ergonomics Ltd (2001)
2. Hinninghofen, H., Enck, P.: Passenger well-being in airplanes. Auton Neurosci. 129(1-2),
80–85 (2006)
3. SEAT Project Consortium, SEAT Project (2006), http://www.seat-project.org
4. Hickman, B.J., Mehrer, R.: Stress and the effects of air transport on flight crews. Air
Medical Journal 20(6), 2–56 (2001)
5. O’Donovan, K.J., et al.: An investigation of recommended lower leg exercises for induced
calf muscle activity. In: Proceedings of the 24th IASTED international conference on
Biomedical engineering, 2006 of Conference, pp. 214–219. ACTA Press (2006)
6. Ball, K.: Deep vein thrombosis and airline travel–the deadly duo. AORN Journal 77(2),
346–354 (2003)
7. World Health Organisation, Chapter 2: Travel by Air: Health Considerations, International
travel and health: Situation as on 1 January (2005), http://www.who.int/ith
8. Dowdall, N., Evans, T.: The impact of flying on passenger health: a guide for healthcare
professionals, BMA policy report, Board of Science and Education, British Medical
Association (2004)
9. Alonso, M.B.: Affective Tangible Interaction; Towards Reducing Stress. In: Proc. HCI
Close 2U - 9th Sigchi.nl conference (2005)
10. Civil Aviation Authority, CAP 747: Mandatory Requirements for Airworthiness (2006),
http://www.caa.co.uk/docs/33/CAP747.PDF
11. Breuer, H.: Interaction Design for Flight Entertainment, bovacon (2006)
12. Jeff, T.H.: Introduction to Neural Networks with Java. Heaton Research, Inc. (2005)
13. Marrone, P.: The Joone Complete Guide (2007), http://www.joone.org
Designing Engaging Interaction with Contextual Patterns
for an Educational Game
Chien-Sing Lee
Abstract. This paper aims to address two problems. The first problem is how to
develop engaging (deep and meaningful) pedagogical patterns and still ensure
there is sufficient incremental cognitive complexity. The second problem deals
with how to connect pedagogical patterns to HCI and software engineering to
form a systemic interaction design framework. An educational game is used as an
example. Significance of the study lies in the development of a means to create
interaction designs around design-for-engagement requirements and the flexible
scaling and synergy of different frames of reference (pedagogy-application do-
main-HCI- software engineering) through instantiations from the synergised
patterns; reducing error and cost and encouraging new experimentations with
transfer of engaging learning.
1 Introduction
Instructional design strategies revolving around the iterative Analyze, Design, Develop,
Implement and Evaluate (ADDIE) model have been regarded as the conventional
framework that guide the design of systematic learning interaction. At a higher level of
design, interaction tasks are mapped to learning outcomes. Bloom’s revised taxonomy
(remembering, understanding, application, analysis, evaluation, creation)[1] and
Kirkpatrick’s four levels of evaluation (reactions, learning, transfer in behavior and
effects on business results) [2] are often cited as references that help this mapping. The
former is for teaching and learning and the latter for training purposes. However, deep
and meaningful interaction design often remains elusive. To address this problem, we
will first look at what constitutes deep and meaningful interaction.
Earlier work has classified interactions based on the learning outcome when stu-
dents interact with specific media such as hotspots or animations [3, 4]. However, [5]
believes that quality of interaction should consist of a greater number of interactions
involving deep and meaningful cognitive processing such as problem solving, deci-
sion-making and evaluation compared to those of lower levels of cognitive processing.
Another perspective to interaction design measures its success by the degree that
it enhances usability and consequently, user experience. Usability is assessed in terms
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 361–370, 2008.
© Springer-Verlag Berlin Heidelberg 2008
362 C.-S. Lee
of the degree of effectiveness, efficiency, safety, utility, ease in learning and ease in
remembering. If these core factors are met, then user experience is more likely to result
in interactions that are fun, helpful, satisfying, motivating and encourages creative
expressions [6]. Hence, deep and meaningful interaction extends from hu-
man-computer interaction issues such as how to create user-friendly means for keying
in input, navigation or search. Deep and meaningful interaction involves higher cog-
nitive processes mediated through high degree of usability and satisfying user experi-
ences that motivate and encourage creative expressions.
Two problems are addressed. First, successful practices in designing for engaging
(deep and meaningful) interaction need to be shared, adopted or adapted for further
innovations to occur. As such, it is essential to identify a means to guide development
of interaction design for engaging learning.
Second, research on pedagogical patterns and interaction design patterns have often
been carried out separately from each other as the focus of the former is often on
identifying recurring problems and successful solutions for teaching and learning,
whereas the latter is concerned with successful solutions to recurring usability and user
experience problems with technology as mediator in human-computer interactions.
Factored into any systems design, the software engineering aspect need to be added to
pedagogical and human-computer interaction considerations. Hence, in order to design
deep and meaningful interaction from a systems perspective, there is a need to syner-
gize pedagogy, human-computer interaction and software engineering to form a
common frame of reference so that the interdisciplinary patterns obtained through the
tri-partite synergy can become the basis for forming pedagogically oriented interdis-
ciplinary pattern languages.
2 Related Work
o Learners need to be shown the task that they are going to solve
o Learners need to explicitly identify differences from one stage of learning to
the other
• Learning is facilitated when existing knowledge is activated as a foundation for
new knowledge (learning by doing)
o Learners are asked to recall, relate, describe or apply knowledge from past
experience
o Learners are provided with relevant experience fundamental to the next task
o Learners are given the opportunity to demonstrate their grasp of knowledge
• Learning is facilitated when new knowledge is demonstrated to the learner
(careful sequencing of learning activities)
o Demonstration is consistent with the learning goal e.g. demonstrations of
procedures, visualizations of processes and modelling of behaviour
o Learners are given suitable guidance e.g. provision of relevant information,
multiple forms of representations (text, graphics, videos etc.), contrasts be-
tween demonstrations
• Learning is facilitated when new knowledge is applied by the learner (reuse)
o Appropriate feedback and coaching should be provided inclusive of identifi-
cation and correction of errors
o The problems to be solved should be varied but incremental in complexity.
• Learning is facilitated when new knowledge is integrated into the learner’s world
(reuse and sharing)
o Learners should be given the opportunity to demonstrate their new knowledge
o Learners are given opportunities to reflect, discuss and defend their opinion
o Learners are given opportunities to create and explore new ways to use their
new knowledge
The author takes engagement in Merrill’s First Principles as the core that directs
other principles. As such, in Table 2, engagement is presented in the first column,
mapped to corresponding principles, student-content interaction types and Bloom’s
taxonomy.
The process of interaction design involves 4 activities: identifying needs and estab-
lishing requirements, developing alternative designs that meet those requirements,
building nteractive versions of the designs so that they can becommunicated and as-
sessed and evaluating what is being built throughout the process[6]. Patterns encap-
sulate the name, ranking (stretching from tried-and-tested to new), illustrations, the
problem(s) addressed, forces (gap in design aspects that need to be improved), exam-
ples, solution(s), and diagram (summarized version of the illustration). Connecting
related patterns form a pattern language. Some examples of popular patterns are listed
in [9]’s design patterns.
From the perspective of usability engineering, [10] integrates application domain
pattern languages, human-computer interaction pattern languages and software engi-
neering pattern languages. The application domain pattern language deals with
Designing Engaging Interaction with Contextual Patterns for an Educational Game 365
Table 2. Mapping between First Principles, student-content interaction types and Bloom
The Pedagogical Patterns project [12] was formulated during the OOPSLA’95 con-
ference. Aimed at aiding novice instructors, who have good knowledge of their subject
but are not necessarily experienced in teaching the subject, the Pedagogical Patterns
provide general guidelines from successful teaching and learning strategies from
which instructors can adapt based on their own creativity to suit their students’ needs.
The pedagogical patterns however, are not presented in the same format as design
patterns.
In the section below, the application domain-HCI-software engineering framework
is extended to include pedagogy.
PAHS regards instructional design phases as seamlessly iterating between one and the
other with no strict application in the choice and order of pedagogical techniques in
order not to constrain creativity. What is essential is the observance of the need to
analyze, design, develop and evaluate iteratively with as much user feedback as pos-
sible at each phase. In PAHS, pedagogical patterns are derived from a cross-section of
useful guidelines such as those illustrated in Table 2 above. It is assumed that these
patterns as well as the application domain, HCI and software engineering patterns can
be retrieved from a database through a graphical user interface. An example of PAHS
pedagogical patterns for an educational game in the domain of water pollution is shown
in Table 3 below. In cases where elements in First Principles and tactics are similar,
they are merged.
PAHS includes user feedback and pedagogical patterns into Borcher’s [10] integrated
pattern languages framework within the user-centered interaction design [7] context.
As a recap, interaction design requires first identifying needs and establishing re-
quirements, second developing alternative designs that meet those requirements, third
building interactive versions of the designs so that they can be communicated and
assessed and fourth evaluating what is being built throughout the process.
368 C.-S. Lee
Supposing that the needs of the students is to learn the biological effects on sea life
due to oil spills in the sea and to identify alternative ways to clean up the oil spills
within the shortest time and least cost and least loss of sea lives, the following tasks are
carried out instantiated from Nielsen’s usability engineering framework:
Designing Engaging Interaction with Contextual Patterns for an Educational Game 369
4 Conclusion
Effective learning is the ultimate goal of any instructor. One of the main essences of
effective learning is the development of engaging learning. As such, this paper is con-
cerned with the design of pedagogically oriented engaging interaction design. The au-
thor has presented the development of a user-centered synergised and pedagogically
370 C.-S. Lee
References
[1] Anderson, L., Krathwohl, D. (eds.): A taxonomy for learning, teaching and assessing: A
revision of Bloom’s Taxonomy of Educational Objectives. Longman, New York (2001)
[2] Kirkpatrick, D.L.: Evaluating Training Programs: The Four Levels. Berrett- Koehler, San
Francisco (1994)
[3] Shortridge, A.: Interactive web-based instruction: What is it? And how can it be achieved?
Journal of Instructional Science and Technology 4(1) (March 2001)
[4] Uttendorfer, M.: Interactivity in an online course: Making it more than page turning. In:
Proceedings of World Conference on e-Learning in Corporate, Government, Healthcare
and Higher Education, pp. 147–149. AACE, Cheasapeake (2003)
[5] Dunlap, J.C., Sobel, D., Sands, D.I.: Designing for deep and meaningful student-to-content
interactions. TechTrends 5(4), 20–29 (2007)
[6] Kearsley, G., Shneiderman, B.: Engagement theory: A framework for technology- based
based teaching and learning. Educational Technology 38(5), 20–23 (1998)
[7] Preece, J., Rogers, Y., Sharp, H.: Interaction design: Beyond human-computer interaction.
John-Wiley, New York (2002)
[8] Merrill, M.D.: First principles of instruction Educational Technology. Research & De-
velopment 50(3), 43–59 (2002)
[9] Gamma, E., Helm, R., Johnson, R., Vlissides, J.M.: Design Patterns: Elements of Reusable
Object-Oriented Software. Addison-Wesley, New York (2000)
[10] Borchers, J.: A pattern approach to interaction design. Wiley & Sons, New York (2001)
[11] Nielsen, J.: Usability engineering. Morgan Kaufmann, San Francisco (1993)
[12] Pedagogical patterns, http://www.pedagogicalpatterns.org/
Design and Implement of Game Speech Interaction
Based on Speech Synthesis Technique
1 Introduction
With the rapid development of computer and network technology, electronic game has
become a sort of main entertainments. Electronic game is also considered “the ninth
art”, in the same position with music, movie and other traditional art. Furthermore, as
creation for business purpose, it has exceed other art because of the accurate simulating
of real life. Participation and interactive are unique features of electronic game.
Through text, image and sound, electronic game delivers emotional experiences to
players. With the accumulation of game time, emotional experiences accumulate either.
Sound plays an important role in game but always ignored.
Sound in game is classified into three categories: music, sound effects and speech.
Speech is the main element which used to connects player and game. Generally, it con-
tains short voice and dialogue. Using sound in game has purposes for significance and
emotion. Significance purpose means that use sound to deliver game information, such as
“under attack”. Emotion purpose means that use sound to deliver emotional experiences.
For example, speech in NPC dialogue is used to direct or adjust players’ feeling.
With help of sound, players will immerse into game much easier. Background music
is used to enhance ambience, while speech in CG is introducing the scene and scenario
to players. A lot of games prove that sound can not only clew game information, but
also strengthen the expressive force of game story.
Speech interaction contains speech syntheis and speech recognition. Speech recog-
nition technique is far from mature but speech syntheis has been widely applied.
Therefore, this study based on speech synthesis technology and apply it in game speech
interaction. The purpose is to strengthen the interaction between game and player.
2 Related Work
Emotioneering™ is the term David Freeman created and it’s his registered trade mark.
He is the author of "Creating Emotion in Games". David Freeman identifies 32
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 371–380, 2008.
© Springer-Verlag Berlin Heidelberg 2008
372 X. Wang and R. Yun
Compared with traditional record play method, speech synthesis takes less memory. It
could be more flexible and effective too. Because the file space is reduced, speech can
be used in more places and other languages’ version can be added. The game developer
Design and Implement of Game Speech Interaction 373
only need to prepare actor's lines in text and set the speech engine. Speech can be read
out automatically. There’s several requests of game speech:
3.2 Timely
Interactive is a main feature of game and the feedback of game is rather single now. Use
speech to feed back player is more attractive. Different kinds of game have different
request: RPG game always use feedback in NPC dialogue but RTS game use feedback
for controlled unit.
NPC dialogue is an important technique in RPG game. NPC deeping technique and
NPC interest technique in Emotioneering™ often use dialogue and movement as me-
dia. The NPC dialogue appears in text on screen isn’t intuitionistic enough and could
hardly give player emotional experience. We can set characteristic voices for main
NPC characters and use speech to give player necessary feedback, immediately.
374 X. Wang and R. Yun
Player must control units in RTS games fast and careful. For feeding back player’s
operation, slogan and sound effert should be emitted while unit is enabling. It’s very
necessary in RTS game. For example, nearly every unit in “Warcraft III: Frozen
Throne” has its own slogan. The content of slogan is commonly for inspiriting fight or
awaiting orders. If the unit is zoetic or controlled by life-form, slogan is a better choice
than sound effert only. Characteristic unit response can inpress player profoundly and
enhance the quality of feedback. Player acts general, god and other characters. Speech
feeback follows the operation can violently excite player and give them better emo-
tional experience.
3.3 Humanized
Traditionally, game will use some sound efferts with text on screen to cluw players. But
it’s too mechanical and lack of emotion. Humanized speech clew could arose player’s
emotion experience.
“Warcraft III: Frozen Throne” provides some speech cluw of assistant imformation,
such as lack of resource, building completing, under attack etc. It’s a common trait of
RTS game. Player’s game experience should be well advised, and also should accord
with game speed. Speech can embodies authenticity, construct ambience and increase
mental stimulation. And speech is also a powerful tool used to materialize interactive.
Speech can offer timely and effective imformation. It could help ears and eyes get
information respectively.
Players may want to set the system personally while playing, such as sound on/off,
save mode, screen size etc. System has a default setting and can not give player im-
mediate responses or suggestions. For example, the configuration of a player’s com-
puter can not afford high quality graphics effects. If there’s no cluw to warn player
when a “high” option is choosed, computer may crash or game may terminate. Speech
used here to give player appropriate cluw can make the system more friendly and
effective.
Design and Implement of Game Speech Interaction 375
Game
Speech Engine
Voice Database
Acoustics Module
Speech Output
Mapping Rules
Training Phase
Conversion Phase
Speech conversion
the feature of source speaker. With speech conversion technique, game developer can
set more characteristic voice easier.
Microsoft Speech SDK contains API for Text-to-Speech and API for Speech Recog-
nition. Through using API for Text-to-Speech, programmer can easily develop pow-
erful Text-to-Speech application. Speech object library encapsulates details of speech
synthesis engine and provides upper interfaces for application to access.
The SpVoice object brings the text-to-speech (TTS) engine capabilities to applications
using SAPI automation. An application can create numerous SpVoice objects, each
independent of and capable of interacting with the others. An SpVoice object, usually
referred to simply as a voice, is created with default property settings so that it is ready
to speak immediately.
z Attribute
a. Voice: Denote pronunciation style, system will use correlative voice database to
read. These default four styles are Microsoft simplified chinese, Microsoft mary, Mi-
crosoft mike and Microsoft sam.
b. Rate: The speed of speech . Value range is from -10 to 10.
c. Volume: The volume of speech. Value range is from 0 to 100.
378 X. Wang and R. Yun
z Method
a. Speak: Accomplish the assignment of transforming text information to speech
with the parameters setted. The method has two parameters as test and flags which used
to appoint reading text and PRI of mood.
b. Pause: Pause the reading courses in use of the object.
c. Resume: Resume the reading courses in use of the object.
5.2 An Instance
With help of Microsoft Speech SDK, programmer just need to access supported in-
terfaces and leave bottom work to speech synthesis engine.
Speech Application SDK must be installed before programming. The newest
version is SAPI 5.1 which have language packages of Chinese, Japanese and English.
After being installed, the component “Microsoft voice text” can be used convenient
to realize speech read. The main code of an instance programmed by Visual Basic 6.0
is below:
z Common Event
‘First, declare the main SAPI object we are using in this
sample. It is created inside Form_Load and released inside
Form_Unload.
Dim WithEvents Voice As SpVoice
‘m_speaking indicates whether a speak task is in progress
‘m_paused indicates whether Voice.Pause is called
Private m_bSpeaking As Boolean
Private m_bPaused As Boolean
Design and Implement of Game Speech Interaction 379
z Form_Load Event
‘Creates the voice object first
Set Voice = New SpVoice
‘Load the voices combo box
Dim Token As ISpeechObjectToken
For Each Token In Voice.GetVoices
VoiceCB.AddItem (Token.GetDescription())
Next
VoiceCB.ListIndex = 0
‘set rate and volume to the same as the Voice
RateSldr.Value = Voice.Rate
VolumeSldr.Value = Voice.Volume
modes in game. The probability of changing traditional sound files mode to speech
synthesis mode were also discussed.
Game asks for high quality speeches, so the key of application is natural and effec-
tive synthesized speech. How to improve the quality of synthesized speech will become
an important part in furture research. At the same time, other in-depth speech synthesis
technique, such as emotional speech synthesis, will be researched. The application of
speech in game will get more attention and play a more important role in game.
References
1. Freeman, D.: Creating Emotion in Games, New Riders Games (September 2003) ISBN:
1592730078
2. Fotinea, S.-L., Tambouratzis, G.: A Methodology for Creating a Segment Inventory for
Greek Time Domain Speech Synthesis. International Journal of Speech Technology (June
2005), doi:10.1007/s10772-005-2167-5
3. Sen, A., Samudravijaya, K.: Indian accent text-to-speech system for web browsing, Sadhana
(February 2002), doi: 10.1007/BF02703316
4. Evgrafova, K.: The Sound Database Formation for the Allophone-Based Model for Eng-lish
Concatenative Speech Synthesis. Text, Speech and Dialogue, doi:10.1007/11551874_28
5. Qin, L., Ling, Z.-H., Wu, Y.-J., Zhang, B.-F., Wang, R.-H.: HMM-Based Emo-tional Speech
Synthesis Using Average Emotion Model, Chinese Spoken Language Processing,
doi: 10.1007/11939993_27
6. Black, A., Campbell, N.: Optimising selection of units from speech databases for concate-
native synthesis. In: Proceedings of Eurospeech (1995)
7. Michael Macon, W., Mark Clements, A.: An Enhanced ABS/OLA Sinusoidal Model For
Waveform Synthesis In TTS. In: Proceedings of Eurospeech (September 1999)
8. Bailly, G., Campbell, N.: ISCA Special Session: Hot top ics in speech synthesis. In: Pro-
ceedings of the European Conference on Speech Communication and Technology (2003)
Two-Arm Haptic Force-Feedbacked Aid for the Shoulder
and Elbow Telerehabilitation
1 Introduction
Patients suffering from muscle or ligament diseases need training for their rehabilita-
tion after the medical diagnostic. This kind of treatment can be applied only in a hos-
pital or in a physiotherapist offce. As mentioned in [2], having a machine able to handle
the patients in (even rural) hospitals would then be very useful, but there is still a need
of a physiotherapist during the treatment. If the patients want to get the best rehabili-
tation, they need an expert physiotherapist for their injured limb. Unfortunately, these
physiotherapists are most of the times working in the metropolis because of their
popularity. It would thus be a strong asset if they could remotely control these machine
from their offce. With such a system, the physiotherapists would easily train to better
rehabilitate the patients almost at home.
In this paper, we propose a remotely controlled system for patient shoulders and
elbows rehabilitation and training of physiotherapists. We decide to work with the
Immersion1 Haptic Workstation TM instead of other VR haptic devices. In fact, it is a
two-armed-based system [11] that allows a well-balanced effort of the patients between
their right and left limbs. This avoids excessive torsions and efforts with the spinal
column. Moreover, this machine can be easily tele-operated by a smartphone via
Internet in contrary of other “home medical systems” like the Biodex2 or the Cybex3
that are not less cumbersome.
We first present a short overview of the related fields that lead us to this improved
system. Secondly, we describe the overall system and its implementation. The third
1
http://www.sensable.com
2
http://www.biodex.com/rehab/rehab.htm
3
http://ecybex.com
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 381–390, 2008.
© Springer-Verlag Berlin Heidelberg 2008
382 P. Salamin et al.
2 Related Works
Using machines as an aid for the rehabilitation is not a new concept. Researchers al-
ready thought about it in 1965 [13] but only the veterans were targeted till 1967[7]. In
the following years, researches tends to aid handicapped people to substitute their
“damaged” limbs [4].
Since the beginning of this millennium, rehabilitation became a very fashion topic.
An overview of the current machines aiding to rehabilitation can be found in [16].
Researchers now also work on solutions for incapacitated people in order to help them
to recover the use of their limbs that we split into two parts: the lower upper limbs.
These last ones can be divided in four main parts that have to be considered: shoulder,
elbow, wrist and fingers. Tsagarakis et al. [14] developed one of the first machines
aiding to people rehabilitation. Unfortunately, the applied forces were bounded to 2kgs,
which limited a lot the efficiency of the prototype for a complete treatment.
David Jack et al. focused on the fingers rehabilitation with the help of force-feedback
gloves [6]. Others researchers made complementary works and developed a 6DOF
machine for the shoulder, elbow and wrist rehabilitation for one hand at a time [3].
Nevertheless, all these devices require the presence of a doctor during the patient
healing. In 2005, Demiris et al. highlighted in [2] the interest to develop a machine that
would allow healing the patient at distance. They involved by this way the telereha-
bilitation in the scientific research for the patients’ well-being.
Based on these researches and with the help of a physiotherapist, we assumed that
the HW would be very efficient and promising to treat upper limbs injuries like the
shoulders and the elbows. We present our system in the following chapter.
As the goal of this application is the telerehabilitation, we want to allow the physio-
therapists to be completely mobile and independent from the patient location. A PDA
with an integrated webcam (e.g. a PDA-phone) seems thus to perfectly fit to our re-
quirements. By this way, the patient can see a video-streaming of the physiotherapist
sent by the PDA during its treatment. The physiotherapist is informed of the patient
arms location with a 3D interface created through the MVisio 3D graphics engine [12]
and the data sent by the HW. As you can see on the first picture from the left on the
Figure 1, the physiotherapist does not need to be with the patient or in his/her office to
begin the treatment.
Two-Arm Haptic Force-Feedbacked Aid 383
The patient is wearing a HMD while being seated in the HW as shown on the third
picture from the left on the Figure 1. The HW allows the therapist to apply different
forces in a range of 0 gram to 10k grams in every direction on each arm independently.
As we can see in [15], the shoulder can be injured in several ways. It is thus useful to be
able to apply accurate forces in order to help him/her to move his/her arm at the be-
ginning and then to apply a force against his/her movement. As the HW applies the
forces on the wrist and as we want the user to move his/her shoulder, we hold the pa-
tient arm straight with the help of a harness. The main advantage of using the HW –
even if it is very expensive – resides in the possibility to apply forces on both arms. By
this way, if we apply symmetrical forces on both arms, the patient will not try to twist
his/her trunk to execute the movement. The rehabilitation would then not lead to new
problems with the spinal column.
Moreover, as we are working with the HW, with the help of the paradigm developed
by Renaud Ott in [10], it is possible to compensate the gravity effect as if the patient
arm was in weightlessness. The patients can be treated from the beginning with mi-
cro-gravity to their complete recovery with forces around 10kgs. It is also very inter-
esting for the physiotherapist to be able to apply a constant and exact force (less than
one gram). This would avoid most of errors due to human factor (deviance of the ap-
plied force that is neither exact nor constant).
Finally, electrodes are applied on the patient skin to detect the muscles activity with
the help of an EMG. The provided information is sent to the physiotherapist who can
better evaluate and appreciate the force to apply on the patient and the evolution of
his/her rehabilitation.
The physiotherapist only needs to have a webcam and a PDA with a very simple in-
terface as you can see on the second picture from the left on the Figure 1. He/she can
see the current patient position represented by an avatar. The doctor can also easily
change the forces applied on the patient with the help of the stylus by indicating the
concerned wrist and the force direction and amplitude which depends on the line he/she
draws (values in grams written on the screen).
On the patient side of our application, we know by experience that VR systems are
quite invasive and can stress the user. The simulation would then be less efficient,
traumatizing and even harmful for the patient. This is why the user is immerged in a
virtual environment with a relaxing landscape like a beach or mountains, depending on
his/her preference as you can see on the right picture of the Figure 1. A sweet relaxing
music as background noise also contribute to strongly reduce the patient anxiety due to
the VR machines [8].
Moreover, the patient can see a realistic representation of his/her hands which fol-
lows in real-time the real position and orientation of the real ones. Notice that the force
is applied on the user wrist(s) is represented as for puppets: the concerned wrist is
caught by a wire. With this artifact, the patient can see (right picture of the Figure 1 in
which direction the force is applied and also its intensity (a second wider red wire
indicates it). All these artifacts improve the immersion of the patient who could be
seriously perturbed by the used hardware that we present in the following section.
384 P. Salamin et al.
Once done, some scenarios could be easily added to the simulation. For example, the
user could have to touch a virtual ball the physiotherapist would move around specific
places. It has already been proved that such a playful simulation improves rehabilitation
results [1][9].
Finally, during the simulation, the patient has a “Window to the World” which al-
lows him/her to see the physiotherapist and to listen to him/her during the session. As
shown in [5], it is very relaxing for the patient to have a multimodal link with the doctor
while being in the virtual environment.
In order to prove the efficiency of our system we made some tests that are presented
in the following section.
4 Experiments
We display a relaxing landscape during the complete session. Moreover, we first let the
patient in the environment with a sweet music in background during fifteen minutes in
order to compensate the stress possibly brought by the VR engines. Once the user
seems to be relaxed, we begin the treatment. We decide to call “neutral position” the
position in which a tester tighten ones’ arms close to one’s trunk on the vertical plane
without moving.
Obviously, depending on his/her recovery status, a help (or a constraint) is applied by
the HW when the patient tries to perform the asked movements shown on the Figure 2.
We have to remind that for every exercise, the patient moves both arms in a synchronous
and parallel way (or symmetrical when they are on the same plane, e.g. for the abduction
we present hereinafter) to avoid useless and dangerous efforts and torsions on the spinal
column.
Fig. 1. On the left, the physiotherapist (hardware and interface); on the right, the patient (hard-
ware and interface) during the rehabilitation exercises
Fig. 2. Exercises for the rehabilitation (from the left to the right): inflection, abduction, lateral
rotation, medial rotation for the shoulder and elbow inflection
working during this exercise. In the following section, you will see that we put elec-
trodes on the main-working muscles to check their activity during the tests.
In the abduction exercise, as you can see on the second pictures from the left of the
Figure 2, the patient must move his/her arms in a vertical plane on his/her sides (instead
of ahead). In this experiment, the most acting muscles are the middle deltoid and the
supraspinatus, whose behavior will be observed and commented in the following
section.
Concerning the rotation exercises, the patient arms start in position shown on the top
third and fourth pictures from the left in the Figure 2. For the lateral rotation, the
forearms raise up to the vertical. In this case, the main working muscles are the poste-
rior deltoid, the teres mino and infraspinatus which are working together (bottom
centered picture). For the medial rotation, the forearms fall down (almost) to the ver-
tical. The main detectable activated muscles are the anterior deltoid and the pectoralis
major but some others like the subscapularis (under the pectoralis major), the latissimus
dorsi and the teres major are also working but less perceptible with the EMG.
Concerning the elbow rehabilitation (right pictures in the Figure 2, we first assume
the patient keeps the arm in the “neutral position” (close to the body trunk). After this,
he/she must move vertically in a synchronous way the forearms from the bottom to the
top and vice versa. The most important muscles used to perform this exercise are the
biceps and the triceps on which we also put electrodes in order to obtain their activity.
We present in the next section typical results we obtained during our sessions with the
help of our system and an EMG.
5 Discussion of Results
As written before, in order to avoid bias due to the VR devices used for these experi-
ments, the user is first seating in the HW with sweet music and a relaxing landscape
386 P. Salamin et al.
during fifteen minutes. He/she performs then twice the exercise, e.g. “down – up –
down – up – down” for the shoulder inflection. We present hereinafter the results we
obtained but, in order to better understand the provided graphs, we first define two
terms concerning the muscles:
– effectors: These muscles are used to execute a movement. They give the possi-
bility to the patient to accomplish a movement. If, in our case, the shoulder does not
move, they do not act. Their EMG representation has a big difference when the
shoulder moves or not.
– stabilizers: These muscles are always in activity, even when the shoulder does not
move because they have to stabilize the joint to avoid (e.g. a dislocation due to the
gravity effect). They almost do not act to perform a movement. Their EMG represen-
tation seems to be almost constant.
For the first exercise we present, the inflection, the electrodes are located on the
anterior deltoid and the pectoralis major. They are respectively represented on the EMG
graphs by the red and the blue lines on the Figure 3. Notice that we can perfectly see the
role of those muscles: both are effector muscles and work during the whole movement
(up -stay -down) while the patient arm is not in the “neutral position”. We can also
notice that the main used muscle for the inflection is the anterior deltoid while the
pectoralis major only helps it for this action. It is also proved on the right graph of the
right picture of the Figure 3 because when a force is applied against the patient
movement, the pectoralis major acts during the complete movement.
Fig. 3. Inflection exercise for the shoulder: (left)with help, (right)countered to and (center) po-
sition of the electrodes (anterior deltoid on the bottom left and pectoralis major on the right)
Fig. 4. Abduction exercise for the shoulder: (left) with help, (right) countered to and (center)
position of the electrodes (supraspinatus on the bottom left and middle deltoid on the right)
Two-Arm Haptic Force-Feedbacked Aid 387
The abduction exercise mainly involves the activity of the middle deltoid (in red)
and the supraspinatus (in blue). Their activity during the experiment can be seen on the
Figure 4. Once again, we decided to check the activity of an effector (the deltoid) and a
stabilizer (the supraspinatus) to verify if the HW is really efficient when it helps or
counter to the patient movements. We can see on the Figure 4 that the muscles, during
the movement, are used a lot – even with the help provided by the HW. But we can also
remark that they are almost not used when the user is in the “neutral position”.
For the lateral rotation (Figure 5), the electrodes are located on the posterior deltoid
(blue line) and both teres minor and infraspinatus because they are linked(red line). As
those last ones are stabilizer muscles, they seem to be always active during the ex-
periment while the posterior deltoid can “have a rest” when the patient is in the “neutral
position”. In this case, the main difference of muscle activity appears for the stabilizer
ones because the applied countering forces are strong (right graph of the Figure 5).
Fig. 5. Lateral rotation exercise for the shoulder: (left) with help, (right) countered to and (center)
position of the electrodes (posterior deltoid on the bottom left, teres minor and infraspinatus on
the right)
Fig. 6. Medial rotation exercise for the shoulder: (left) with help, (right) countered to and (center)
position of the electrodes (anterior deltoid on the bottom left, pectoralis major and subscapularis
(hidden) on the right)
The medial rotation (Figure 6) involves lots of muscles. The anterior deltoid the
pectoralis major are the most active and interesting to analyze with an EMG. Among
the others, we can cite the subscapularis, the altissimus dorsi and the teres major. As we
can see it on the left and right pictures, both analyzed muscles are effector ones (only
the subscapuralis is a stabilizer). They thus almost follow the same curve and we can
see that the forces applied by the HW are also quite efficient for this exercise. Notice
that for this exercise, the applied counter forces are pointing ahead and up.
388 P. Salamin et al.
The last exercise we present in this paper concerns the elbow rehabilitation (Figure 7).
We ask the patient to perform elbow inflections and we measure the biceps (red line) and
triceps (blue line) activity. It is interesting to see that the triceps, which is normally an
effector muscle for the extension of the elbow, also acts in this exercise and almost
follow the biceps curve. Otherwise, the efficiency of the HW and the applied force is
obvious. A big difference between the muscle activity between the left (helping forces)
and the right graphs where a force is applied to the front and to the bottom can
be noticed.
Fig. 7. Inflection exercise for the elbow: (left) with help, (right) countered to and (center) position
of the electrodes (biceps on the bottom left and triceps on the right)
Finally, we prove the benefits of working with both arms at the same time. We have
made this assumption because when a physiotherapist asks a patient to do rehabilita-
tion exercises at home, he/she often propose to perform them with both hands. We
then check the activity of the errector spinea (shown in the center of the Figure 8) for
the abduction exercise we have already presented. We locate an electrode on the left
side (red line of the graph below) of the spinal column and another one on the right
side (blue line). In order to obtain the graph on the Figure 8, we first asked the patient
to stick up his/her arm and then to lower it. After this, he/she performs the same
movement with the other arm and finally with both together. As you can see below,
there is a very big difference between the right and left spinal muscles activity when
the patient only sticks up one arm. This leads to spinal column torsion and often to
backache. This graph prove then the importance to work with both arms during the
rehabilitation.
Two-Arm Haptic Force-Feedbacked Aid 389
6 Conclusion
The goal in this paper was to present an efficient aid for the shoulder and elbow tel-
erehabilitation. Our application full-fills the tele-operation side that provides obvious
advantages for the patients and the physiotherapists training. Sometimes an adaptation
time of five to ten minutes was needed to discover the VR material, but none of the
patients were really perturbed by them during the simulation. Furthermore, the obtained
results seem to prove the efficiency of our system for the patients during all the reha-
bilitation phase. We can e.g. see in the graphs that a very light force is needed to
perform the action when the patient starts his/her rehabilitation. And when he/she has
almost recovered all his/her faculties, the HW can apply strong enough forces on
his/her arms to finish correctly the rehabilitation. The EMG last graphs also support the
idea of the minimal spinal column torsions mandatory for the patient comfort. More-
over, the possibility to cure the patients at distance also really interested our physio-
therapist. This technology extends the coverage of this kind of therapy because the
patients can be treated in any hospital even if the therapist is not physically present.
Concerning the future works, it could be really interesting, e.g. for the lateral and
medial rotation, to provide a support for the elbow during the therapy. The physio-
therapist should also take into account an additional five to ten minutes to install the
patient in the HW. However this task can be performed by any member of the medical
staff.
Acknowledgments
References
1. Betker, A.L., Szturm, T., Moussavi, Z.K., Nett, C.: Game based interactive motivating tool
for rehabilitation movements: case study. In: ACST 2006: Proceedings of the 2nd IASTED
international conference on Advances in computer science and technology, pp. 157–162.
ACTA Press, Anaheim (2006)
2. Demiris, G., Shigaki, C.L., Schopp, L.H.: An evaluation framework for a rural home-based
telerehabilitation network. J. Med. Syst. 29(6), 595–603 (2005)
3. Furusho, J., Li, C., Hu, X., Shichi, N., Kikuchi, T., Inoue, A., Nakayama, K., Yamaguchi,
Y., Ryu, U.: Development of a 6-dof force display system using er actuators with
high-safety. In: VRCIA 2006: Proceedings of the 2006 ACM international conference on
Virtual reality con-tinuum and its applications, pp. 405–408. ACM Press, New York (2006)
4. Greenleaf, W.J.: Applying vr to physical medicine and rehabilitation. Commun.
ACM 40(8), 42–46 (1997)
5. Gutiérrez, M., Lemoine, P., Thalmann, D., Vexo, F.: Telerehabilitation: controlling haptic
virtual environments through handheld interfaces. In: VRST 2004: Proceedings of the ACM
symposium on Virtual reality software and technology, pp. 195–200. ACM Press, New
York (2004)
390 P. Salamin et al.
6. Jack, D., Boian, R., Merians, A., Adamovich, S.V., Tremaine, M., Recce, M., Burdea, G.C.,
Poizner, H.: A virtual reality-based exercise program for stroke rehabilitation. In: Assets
2000: Proceedings of the fourth international ACM conference on Assistive technologies,
pp. 56–63. ACM Press, New York (2000)
7. Jaffe, D.L.: An overview of programs and projects at the rehabilitation research and de-
velopment center. In: Assets 1994: Proceedings of the first annual ACM conference on
Assistive technologies, pp. 69–76. ACM Press, New York (1994)
8. Kallinen, K.: The effects of background music on using a pocket computer in a cafeteria:
im-mersion, emotional responses, and social richness of medium. In: CHI 2004: CHI 2004
extended abstracts on Human factors in computing systems, pp. 1227–1230. ACM Press,
New York (2004)
9. Lathan, C., Vice, J.M., Tracey, M., Plaisant, C., Druin, A., Edward, K., Montemayor, J.:
Therapeutic play with a storytelling robot. In: CHI 2001: CHI 2001 extended abstracts on
Human factors in computing systems, pp. 27–28. ACM, New York (2001)
10. Ott, R., Gutierrez, M., Thalmann, D., Vexo, F.: Improving user comfort in haptic virtual
environments through gravity compensation. In: WHC 2005: Proceedings of the First Joint
Eurohaptics Conference and Symposium on Haptic Interfaces for Virtual Environment and
Teleoperator Systems, pp. 401–409. IEEE Computer Society Press, Washington, DC (2005)
11. Ott, R., Perrot, V.D., Thalmann, D., Vexo, F.: Mhaptic: a haptic manipulation library for
generic virtual environments. cw 0, 338–345 (2007)
12. Peternier, A., Thalmann, D., Vexo, F.: Mental Vision: A Computer Graphics Teaching
Platform. In: Pan, Z., Aylett, R.S., Diener, H., Jin, X., Göbel, S., Li, L. (eds.) Edutainment
2006. LNCS, vol. 3942, pp. 223–232. Springer, Heidelberg (2006)
13. Rose, G., Woodbury, M.A., Fergusson, E.S.: Recent study of teaching: machine aids in
rehabilitation therapy. In: Proceedings of the 1965 20th national conference, p. 7. ACM
Press, New York (1965), Turner, L.R. (Chairman)
14. Tsagarakis, N.G., Caldwell, D.G.: Development and control of a “soft-actuated” exo-
skeleton for use in physiotherapy and training. Auton. Robots 15(1), 21–33 (2003)
15. Uhthoff, D.H.K.: Shoulder injury and disability. Technical report, Orthopaedic Surgeon
(2002)
16. Yeh, S.-C., Rizzo, A., Zhu, W., Stewart, J., McLaughlin, M., Cohen, I., Jung, Y., Peng, W.:
An integrated system: virtual reality, haptics and modern sensing technique (vhs) for
post-stroke rehabilitation. In: VRST 2005: Proceedings of the ACM symposium on Virtual
reality software and technology, pp. 59–62. ACM, New York (2005)
Vision Based Pose Recognition in Video Game
1 Introduction
The recent video gaming consoles operate with the motion sensor technology for im-
mersive game playing experience appealing a wide range of game users. The
depth-sensing camera is also highly applicable but need extra cost for the end user [10].
Although the vision-based interface is considered as an effective way to capture user
inputs without any additional burden such as a data glove, it lacks its robustness in
illumination changes and small perturbations such as slightly moving objects in back-
ground. In addition, the minimum computation cost is also required for a game appli-
cation in real-time [6].
The background subtraction methods for segmenting foreground regions are at-
tempted using color or intensity thresholding. To model the variances of background, C.
Stauer used mixture of Gaussian on changes of pixel intensity taking several frames to
converge each Gaussian [8]. But it needs few frames at initialization step. Recently, a
few methods of temporal such as Optical flow and color co-occurrence were introduced
for object segmentation in motion [9], [13], but it is infeasible for real-time processing.
In contrast to tracking based algorithm that localizes regions or points [6], [8], [11],
in this paper, we analyze a whole image using local histograms despite their previous
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 391–400, 2008.
© Springer-Verlag Berlin Heidelberg 2008
392 D.H. Jang, X.H. Jin, and T.Y. Kim
location in a frame-by frame basis so that the unexpected convergence to local minima
can be avoided. Histogram-based object tracking is practiced in many previous re-
searches [2], [11], [14]. A color histogram is easy to compute and is partially robust
against small perturbations since it represents color distribution of the area while
lacking its spatial information [4], [5]. But, color histogram-based approach fails under
illumination variance caused by self-shadow casting foreground object or automatic
bright adjustment of camera device. To overcome this problem, using only the color
information is a possible solution like Hue channel in HSV color space [1], but it also
fails under low brightness because the Hue value isn’t defined.
Orientation histogram has been identified as a potential of using histogram-based
descriptors for visual tracking. It shows strong performance against illumination
changes which cannot be accomplished by color or intensity-based histogram. Moreover,
we adopt the local representation of image that describe the background and the fore-
ground by splitting each image into small cells which gives coarse spatial information.
For fast computation, the use of the integral histogram [3] and efficient comparing
method [1] are exploited and improved to work with orientation-based local histogram.
The extracted foreground cells are used as an input to the neural network where the
predefined poses are already trained. Including recognition module, we propose a vi-
sion-based interface system whose main purpose is to be seamlessly attached to the
existing game engine under given game context and corresponding input map. We have
implemented real-time game with vision interface system to show functional efficiency
of proposed system for controlling virtual environment. The overall control flow is
shown in Figure 1.
The following section explains how to calculate the local orientation histograms
with various. Section 3 introduces the vision interface system including the neural
network for pose recognition. Section 4 is experiment result with real-time 3D game
implementation and the final section concludes the paper.
The necessary steps to compute gradient orientation histogram in cell are as follows.
Firstly, the gradient of I is computed at each point ( x, y ) , given
dy = I ( x, y + 1) − I ( x, y − 1) and dx = I ( x + 1, y ) − I ( x − 1, y ) , then the magnitude and the
orientation is calculated as follows.
m( x, y ) = sqrt (dx 2 + dy 2 ) , and θ ( x, y ) = arctan( dy / dx ) (1)
Secondly, θ is quantized into N bins. The running sum of each bin is computed
separately. In order to reduce the effect of noise, the contribution of each point in
θ ( x, y ) to the corresponding bin is weighted by its magnitude m( x, y ) . We also apply
Gaussian kernel on 1D histogram for the θ ( x, y ) to contribute several bins according to
Gaussian weights since limiting the corresponding histogram bin of pixel ( x, y ) to a
single bin is a major reason of quantization errors. The example of gradient orientation
histogram is shown in Figure 2. In the next section, we describe how the calculation of
histogram can be accelerated by integral histogram and multi-scaled search algorithm.
As proposed in [1], the histogram based multi-scaled search algorithm requires multi-
ple extractions of histograms from multiple rectangular cells. The tool enabling this to
be done in real time is the integral histogram described in [3].
The integral histogram method is an extension of the integral image data structure
described in [7]. The integral image holds the sum of all the pixels contained in the
rectangular region at the point ( x, y ) in the image. This image allows computing the
sum for the pixels on arbitrary rectangular regions by considering the 4 integral image
values at the corners of the region i.e it is computed in constant time independent of the
size of the region. The integral histogram at each position ( x, y ) is calculated with
wavefront scan propagation that 3 neighbor integral histograms and the current orien-
tation value are used. It is done at each frame and consumes the most of computation
time and the Figure 3 gives detailed explanation of scanning process.
Then by accessing these integral histograms we can immediately compute the local
histogram in a given region with integral data at four corners as follows.
H l = I (right , bottom) − I ( right , top) − I (left , bottom) + I (left , top ) (2)
Given the local histograms of the background and the foreground, we can determine
which areas of the current frame contain foreground objects by computing distance of
394 D.H. Jang, X.H. Jin, and T.Y. Kim
Fig. 2. The test image (left) and its local orientation histogram (right), In histogram plotting, the
line direction represents its bin position (angle) and its length is magnitude.
Fig. 3. Integral data at I ( x, y ) contains the cumulative orientation values from the starting
point (0,0) to I ( x, y ) in rectangular region. The scanning must be done before histogram ex-
traction step. In the wavefront scanning, H ( x, y ) can be calculated from 3 memory access
from H ( x − 1, y − 1) , H ( x, y − 1) and H ( x − 1, y ) and add Q ( I ( x − 1, y )) .
two histograms. To compare the current local histogram Hlcur with the reference local
histogram Hlref at the same cell l, we first normalize the histograms and apply Bhat-
tacharyya distance measurement.
Rather than subtracting cell by cell, we adopt multi-scale search algorithm to skip
large region where no foreground objects come in [1]. The key algorithm is that it
recursively check histogram distance between foreground and background until the
search level reaches to maximum level. Note that there is no additional histogram
quantization process for searching at each level. With the help of the pre-scanned in-
tegral histogram, multiple extractions of histogram over different scaled regions are
done in constant time. This approach mainly has two advantages. The first is that it
Vision Based Pose Recognition in Video Game 395
suppresses the false positives due to various noises in the background. Another benefit
of using the multi-scaled approach is the superior computation speed. It automatically
skips the large background area and go deeper level to find foreground cells while cell
by cell comparison always consume v×h computation time.
Figure 4 shows the efficiency of multi-scaled search approach. Note that the
searched level is highlighted in red rectangles. The large local area implies that it skips
the sub region searches which results in suppression of the false positives and com-
putation advantage compared to the cell by cell comparison method. Also the figure
shows that the orientation histogram has a feature of illumination invariance.
Fig. 4. Foreground cell extraction using multi-scaled search algorithm of test image with light
on/off. The maximum search level is 4, the level threshold Tl is 0.03.
3 Vision-Game Interface
The extracted foreground cells give enough features to be recognized as a distinctive
pose to be transferred to game system as an input event. We consider gaming envi-
ronment that the user sits close to the monitor and the camera, then the absolute position
of foreground cells from the user’s upper body are used as inputs for the simple ex-
perimentation. If the captured signal falls in a predetermined set, the recognition
module compares the current foreground cells with the each of the stored templates and
either selects the category of the closest match or interpolates between templates using
Euclidian distance or similar measurement. Another obvious method is to collect an
396 D.H. Jang, X.H. Jin, and T.Y. Kim
example database of poses and use it to train an artificial neural network. This avoids
having to define each gesture in detail. It is hoped that the network will find the im-
portant features and abstract them. In addition, the networks are often robust to noise
and work well for different individuals.
Fig. 5. (a) Selected poses. (b) Neural network architecture for input map #2.
4 Experiments
We have built the vision game interface library written in C++. The main idea of this
paper is implemented with OpenCV vision library. We have developed a simple game
using Torque3D to show the usability of the proposed system in the real-time proc-
essing situation. The game application is free flying simulation over a simple terrain
map with first person view. The camera view position drifts freely until the neural
outputs are produced. Before the game starts, the simple main menu comes out for
selecting different map where the system use different input map from game.
playing scene. In the game playing scene, when foreground objects detected, the
neural classifier produces one of predefined classes which converted to keyboard input
and the result is transmitted to the game system. With the classified pose pattern, the
user can turn left/right or accel/stop in the game playing scene.
398 D.H. Jang, X.H. Jin, and T.Y. Kim
The camera grabs 320×240 RGB images at 60fps and the testing PCs are equipped
with Intel dualcore 2.0G processor, 1G memory and Radeon X1600 graphic card for
both the game and recognition processes. We measured 30 fps with 320×240 sized
integral histogram process. Moreover, with resized 160×120 image, we get 60 fps, i.e.
sufficient for application to 3D gaming environments. The main computational load is
scanning process of integral histogram where we expect that SIMD processing can help
significant performance improvement.
We prepared the ground truth data from captured movie clip. The 560 frames of
image are captured and the expected pose is manually marked at each frame. We com-
pared the ground truth data to recognition result and Table 1 shows the recognition result.
The poses from 1 to 3 are used for menu traveling in the menu screen and others are used
for controlling the air-plane view. Note that the frames that have non-determined poses
are excluded from counting. The overall detection ratio is more than 90% so that
the proposed vision-based interface system enables the game user to actually control
the game without difficulty. Figure 6 shows the both screens from the camera and the
game scene.
5 Conclusions
In this paper we extract foreground cells using local oriental histogram comparison and
the extracted information is used as inputs to the trained neural network for pose rec-
ognition. For robust and fast computation of local histogram we perform Gaussian
kernel to 1D orientation histogram and utilize the computational efficiency of integral
histogram. In addition, multi-scaled search algorithm proved to be tolerant to camera
noise and slight moving object including illumination changes in background. The
effective way of processing input events from the vision based module is proposed and
implemented as a vision interface library and the simple 3D game is given as a test bed
to prove its efficiency.
Acknowledgments
This research was supported by the ITRC (Information Technology Research Center,
MIC) program and Seoul R&BD program, Korea.
References
1. Jang, D.H., Chai, Y.J., Jin, X.H., Kim, T.Y.: Realtime Coarse Pose Recognition using a
Local Integral Histogram. In: International Conference on Convergence Information
Technology, November 21-23, pp. 1982–1987 (2007)
2. Mason, M., Duric, Z.: Using histograms to detect and track objects in color video. In: Ap-
plied Imagery Pattern Recognition Workshop, AIPR 2001 30th, IEEE, October 10-12, pp.
154–159 (2001)
3. Porkili, F.: Integral histogram: A fast way to extract histograms in cartesian spaces. In: Proc.
IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) (2005)
4. Noriega, P., Bascle, B., Bernier, O.: Local kernel color histograms for background sub-
traction. VISAPP, 1st edn., pp. 213–219. INSTICC Press (2006)
5. Noriega, P., Bernier, O.: Real Time Illumination Invariant Background Sub-traction Using
Local Kernel Histograms. In: British Machine Vision Association (BMVC ) (2006)
6. Bradski, G.: Real time face and object tracking as a component of a perceptual user inter-
face. In: Proc. IEEE WACV, pp. 214–219 (1998)
7. Viola, P., Jones, M.: Robust real time object detection. In: IEEE ICCV Workshop on Sta-
tistical and Computational Theories of Vision (2001)
400 D.H. Jang, X.H. Jin, and T.Y. Kim
8. Stauer, Grimson, W.E.L.: Adaptive background mixture models for real-time tracking. In:
Computer Vision and Pattern Recognition Fort Collins, Colorado, June 1999, pp. 246–225
(1999)
9. Wixson, L.: Detecting salient motion by accumulating directionary-consistent flow. IEEE
Trans. Pattern Anal. Machine Intell. 22, 774–780 (2000)
10. Yahav, G., Iddan, G.J., Mandelbaum, D.: 3D imaging Camera for Gaming Application. In:
Consumer Electronics 2007. ICCE 2007. Digest of Technical Papers (2007)
11. Marimon, D., Ebrahimi, T.: Orientation histogram-based matching for region tracking. In:
IEEE Eight International Workshop on Image Analysis for Multimedia Interactive Services
(WIAMIS 2007) (2007)
12. Viola, P., Jones, M.: Robust real time object detection. In: IEEE ICCV Workshop on Sta-
tistical and Computational Theories of Vision (2001)
13. Li, L., Huang, W.M., Gu, I.Y.H., Tian, Q.: Foreground object detection in changing back-
ground based on color co-occurrence statistics. In: Proc. IEEE Workshop Applications of
Computer Vision, December 2002, pp. 269–274 (2002)
14. Freeman, W., Tanaka, K., Ohta, J., Kyuma, K.: Computer vision for computer games. In:
Int’l Workshop on Automatic Face- and Gesture-Recognition (1996)
Memotice Board: A Notice Board with Spatio-temporal
Memory
Abstract. This paper describes the design and development of a novel digital
notice board which allows non experienced users to easily interact with digital
information. In particular, the system allows the user to receive and handle media
elements (pictures, text messages, videos). Instead of employing the file system
to interact with information, the user interface promotes a kind of interaction
which relies on spatial and temporal memory, which we believe to be more
adequate.
1 Motivation
The area of technology-enhanced learning deals, normally, with the tasks and objects
more directly involved in the learning process (contents, exercises, evaluation, etc).
However, in a conventional physical education centre, there are other objects and tasks
which, although not directly related to the learning process, enhance the student ex-
perience. An example is a notice board in an education centre which represents a social
space where students’ interests meet. The notice board is not only useful because of its
main functionality, but also because of its social side. We research new approaches to
include this kind of objects and tasks in the technology-enhanced learning process in an
appropriate way, as we think that these objects can also benefit from novel technol-
ogy-enhanced approaches.
In particular, in this paper we describe the design and development of Memotice
Board, a digital notice board with spatio-temporal memory. The design of this system is
based on previous work we carried out in the framework of the EC funded IST project
ICING (Innovative CIties of the Next Generation) which, among other things, explores
new ways of communication and interaction. In this context, needs for communication
mechanisms and social awareness were found in several communities (elderly people,
women association, families, etc.) through user studies. In order to fulfill these re-
quirements we designed DINDOW (DIgital wiNDOW), a system which allows the user
to receive, handle and send media elements (pictures, text messages, videos) in a very
simple way. Instead of employing the file system to interact with information, the user
interface promotes a kind of interaction which relies on spatial and temporal memory,
which we believe to be more adequate for our users. Thus, DINDOW is our original
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 401–409, 2008.
© Springer-Verlag Berlin Heidelberg 2008
402 J. Ibáñez et al.
2 Related Work
Awareness systems, according to Markopoulos et al. [10], can be defined as systems
whose purpose is to help connected individuals or groups to maintain a peripheral
awareness of the activities and situation of each other. The area of awareness system
is a flourishing field of research, and interesting systems have been proposed, in the
last years, for both workplaces [6][9] and social/family life [8][11][12][5]. In this
sense, Memotice Board is an awareness system which on the one hand can be used as
a peripheral display (it can be used as a notice board displaying pictures, news, etc)
and on the other hand can be actively employed (the students can interact with the
interface and the system administrator can manage the collection of multimedia
elements).
It is well known that managing disparate data through traditional hierarchical
storage and access interfaces is frustrating for users [3], specially for non experienced
ones. As a consequence, different approaches and metaphors have been proposed to
replace the desktop metaphor and its related hierarchical file system. A specially
interesting alternative was proposed by Fertig et al. in Lifestreams [7], which uses a
timeline as the major organizational metaphor for managing file spaces. It provides a
single time-oriented stream of electronic information, and supports searching,
filtering and summarization. Rekimoto extended this uni-dimensional idea to a
bi-dimensional interface in TimeScape [14], which combines the spatial information
management of the desktop metaphor and time travelling. Our system, Memotice
Board, extends these ideas by allowing time-travelling in any subregion of the user
interface, and adding the ability to formulate natural and intuitive spatio-temporal
queries.
3.1 Architecture
The overall architecture of the system is shown in figure 1. It is composed of two kinds
of agents (administration agent and public agent), a database and a file system.
Memotice Board: A Notice Board with Spatio-temporal Memory 403
The administration agent is the main agent, which is employed by the administrator
to manage the complete system. This agent includes the communication manager that is
responsible for receiving and sending information. It consists of a series of modules
implementing di_erent communication modalities. The current version contains a
module implementing communication over the XMPP protocol. Thus, users from the
Internet (students, ex-students, teachers, etc) can send new elements (text messages,
images or videos) to the Memotice Board by using an XMPP client (that is, an instant
messenger client). By using the administration agent, the administrator can access all
the functionalities of the system.
The information stored by the system can be classified into two categories: multi-
media elements and metadata. The multimedia elements are the images and videos
received by the system from other users. These elements are automatically stored in the
file system. The user does not need to know where these elements are physically lo-
cated nor how they are named. The metadata are semantic annotations about the mul-
timedia elements. These annotations include, for instance, the date when the element
was received, its width and height, its location on the screen, its location on the file
404 J. Ibáñez et al.
system, etc. The metadata are stored in a database. In particular, in the current version
the metadata are stored in eXist [13], an open source native XML database. Thus, this
metadata is accessed through queries programmed in the XQuery language [16].
There can be any number of public agents. Each public agent is responsible for
controlling a particular digital interactive notice board which is exposed in a public
space in the education centre so the students can interact with it. Usually, a public agent
allows the user only to access some of the system functionalities (a subset of the func-
tionalities provided by the administration agent).
The user interface (of both kinds of agents) incorporates novel interaction mecha-
nisms, programmed with SwingStates. The user interface also employs a physics li-
brary to simulate the behaviour of group movement. The combination of both the novel
interaction mechanisms implemented in the user interface and the metadata stored in
the database, allows the user to naturally and intuitively interact with the multimedia
elements.
As stated by Beaudouin-Lafon [2], the only way to significantly improve user inter-
faces is to shift the research focus from designing interfaces to designing interaction.
This requires, among other things, powerful interaction models, beyond the usual ones.
Moreover, as we were prototyping following an iterative process to design the system,
we required advanced software libraries which ease the rapid development of new
kinds of interaction. The typical libraries (like Java Swing) based on a set of well
known graphical widgets are not appropriate, as they are oriented to create new user
interfaces, not new interaction mechanisms. After surveying the state of the art, we
finally decided to employ SwingStates [1]. SwingStates is a library that adds state
machines to the Java Swing user interface toolkit. Unlike traditional approaches, which
use callbacks or listeners to define interaction, state machines provide a powerful
control structure and localize all of the interaction code in one place.
The user interface of the public agent is a subset of the user interface of the ad-
ministration agent. Thus, we will start by describing the latter one, and we will later
specify which parts of it compose the former one. Figure 2 shows the user interface of
the administration agent. As shown in the figure, the user interface is composed of three
main regions (A, B and C). Region A is a scrollable space where the multimedia ele-
ments (pictures, videos, texts) are automatically added when they are received by the
system. In particular, when a new element arrives, it appears on the right of this area,
while the other elements in this space shift to the left accordingly. Thus, all the ele-
ments in region A are temporally ordered. An element X was received earlier than an
element Y if X is visually located to the left of Y . The user can scroll through the
elements of this space. Thus, he can access any received element.
When a new element is received, not only does the element appear on the right of
region A, but also is the user warned by a particular sound. Moreover, the new element
is enhanced with a colourful frame which identifies it as a new incoming element.
The user can group elements in folders but always keeping the temporal order. Thus,
the folders are created in a simple way. The user only decides the extreme elements of
the folder, that is, the oldest and newest element, and the folder is automatically created.
Memotice Board: A Notice Board with Spatio-temporal Memory 405
When several elements arrive together in the same message, they are automatically
associated as a group. Then, these elements will behave as a group. For instance, the
user can move the whole group by just moving one of the elements. The smooth
movement of the group is simulated by employing the proper algorithms from a physics
programming library.
Note that pictures, videos and text messages are clearly differentiated by the way
they are represented. Pictures are displayed with the Polaroid aesthetic, while videos
are presented with a film look, and text messages are shown in coloured PostIts such
that the colour identifies the kind of sender (that is, there is a particular colour per each
sender: student, ex-student, professor, education centre, etc).
Region B is an area which can be seen as a kind of advanced notice board. The user
can move multimedia elements from region A to this area. When that happens, the
elements which are moved to region B still remain visible in region A, although in
ghost mode (with a degree of transparency). Elements moved from A to B, appear
bigger in B than in A. However, relative size among elements is respected. That is, an
element X will appear bigger than an element Y in region B if X as well appears bigger
in region A.
The elements in region B will keep visible until the user decides to remove them. In
this region, the multimedia elements can be freely moved, rotated and scaled. Moreover,
video elements can be played and text annotation can be added to pictures.
Region B has a kind of spatio-temporal memory. For each element, its state-changes
(location, size, rotation angle, time of change,...) are annotated in the database. Thus,
by employing this memory it is possible to time-travel in this region. By travelling to
the past, this region evolves by showing previous states of the region in an inverse time
order. Elements appear and disappear, at a configurable rate, at the locations they were
located at previous times, giving the user the impression that the region is travelling to
past. Moreover, it is possible to time-travel in a subregion of region B. The user can
mark a particular subregion by drawing its border on the screen. As a result, a hole
appears in the just marked subregion. Then the user can time travel only in this subre-
gion (see figure 4). Thus, he can access, for instance, an ad about a flat for renting he
remembers that was located on the center of the right side by marking this area and
time-travelling to past. Time-travels to the future (from past times) are also allowed.
Both, the direction of the time-travel (forward or backward) and its speed are control-
lable through a visual widget.
Natural spatial queries can be formulated in region B. To start the query, the user
marks a subregion by drawing it on the screen. Then he can specify search criteria in an
intuitive way. For instance, he can specify the sender of the element he is searching by
dragging the picture of the sender from region C to the just marked subregion. He can
also specify the approximate size of the element being searched by just drawing it. Thus,
queries are constructed by using simple gestures. In the mentioned example, the system
will find and show elements which were located in the marked subregion and fulfill the
specified search criteria (sender and approximate size).
Region C is a space that contains a set of pictures representing the different kinds of
contacts (students, ex-students, teachers, education centre, etc). These pictures can be
used as clues to filter (for instance to look for elements sent by a particular type of
contact).
Memotice Board: A Notice Board with Spatio-temporal Memory 407
As mentioned before, the user interface of the public agent is a subset of the user
interface of the administration agent. The administrator is responsible for defining the
capabilities of each public user interface. Usually, it consists only of the region B, that
is, the proper notice board (see figure 5). Thus, students can interact with the published
elements by time-travelling and performing spatial queries. However, they cannot
remove any element from the notice board (this is the administrator’s responsibility).
Acknowledgments
The authors thank both Leticia Lipp for generously proofreading and Caroline Appert
for her quick responses to our doubts when using SwingStates. This work has been
partially funded by the European Union IST program through the project ”ICING:
Intelligent Cities for the Next Generation”.
References
[1] Appert, C., Beaudouin-Lafon, M.: Swingstates: adding state machines to the swing toolkit.
In: UIST 2006: Proceedings of the 19th annual ACM symposium on User interface soft-
ware and technology, pp. 319–322. ACM Press, New York (2006)
[2] Beaudouin-Lafon, M.: Designing interaction, not interfaces. In: AVI 2004: Proceedings of
the working conference on Advanced visual interfaces, pp. 15–22. ACM Press, New York
(2004)
[3] Cook, T.: It’s 10 o’clock: do you know where your data are? Technology Review 98(1),
48–53 (1995)
[4] de Greef, P., IJsselsteijn, W.A.: Social presence in a home tele-application. CyberPsy-
chology and Behavior 4, 307–315 (2001)
[5] Dey, A.K., de Guzman, E.: From awareness to connectedness: the design and deployment
of presence displays. In: CHI 2006: Proceedings of the SIGCHI conference on Human
Factors in computing systems, pp. 899–908. ACM Press, New York (2006)
[6] Dourish, P., Bly, S.: Portholes: supporting awareness in a distributed work group. In: CHI
1992: Proceedings of the SIGCHI conference on Human factors in computing systems, pp.
541–547. ACM Press, New York (1992)
[7] Fertig, S., Freeman, E., Gelernter, D.: Lifestreams: an alternative to the desktop metaphor.
In: CHI 1996: Conference companion on Human factors in computing systems, pp.
410–411. ACM Press, New York (1996)
[8] Hindus, D., Mainwaring, S.D., Leduc, N., Hagstrm, A.E., Bayley, O.: Casablanca: de-
signing social communication devices for the home. In: CHI 2001: Proceedings of the
SIGCHI conference on Human factors in computing systems, pp. 325–332. ACM Press,
New York (2001)
[9] Jancke, G., Grudin, J., Gupta, A.: Presenting to local and remote audiences: design and use
of the telep system. In: CHI 2000: Proceedings of the SIGCHI conference on Human
factors in computing systems, pp. 384–391. ACM Press, New York (2000)
Memotice Board: A Notice Board with Spatio-temporal Memory 409
[10] Markopoulos, P., de Ruyter, B., Mackay, W.E.: Awareness systems: known results, the-
ory, concepts and future challenges. In: CHI 2005: CHI 2005 extended abstracts on Human
factors in computing systems, pp. 2128–2129. ACM Press, New York (2005)
[11] Markopoulos, P., IJsselsteijn, W., Huijnen, C., Romijn, O., Philopoulos, A.: Supporting
social presence through asynchronous awareness systems. In: Being There -Concepts,
Effects and Measurements of User Presence in Synthetic Environments, pp. 261–278. IOS
Press, Amsterdam (2003)
[12] Markopoulos, P., Romero, N., van Baren, J., IJsselsteijn, W., de Ruyter, B., Farshchian,
B.: Keeping in touch with the family: home and away with the astra awareness system. In:
CHI 2004: CHI 2004 extended abstracts on Human factors in computing systems, pp.
1351–1354. ACM Press, New York (2004)
[13] Meier, W.: eXist: An Open Source Native XML Database. In: Chaudhri, A.B., Jeckle, M.,
Rahm, E., Unland, R. (eds.) NODe-WS 2002. LNCS, vol. 2593, pp. 169–183. Springer,
Heidelberg (2003)
[14] Rekimoto, J.: Time-machine computing: A time-centric approach for the information en-
vironment. In: ACM Symposium on User Interface Software and Technology, pp. 45–54
(1999)
[15] van Baren, J., IJsselsteijn, W.A., Romero, N., Markopoulos, P., de Ruyter, B.: Affective
benefits in communication: The development and field-testing of a new questionnaire
measure. In: PRESENCE 2003, 6th Annual International Workshop on Presence, Aalborg,
Denmark (October 2003)
[16] Walmsley, P.: XQuery. O’Reilly Media, Inc., Sebastopol (2007)
Mobile Cultural Heritage:
The Case Study of Locri
Abstract. The goal of this project is to study the use of mobile tech-
nologies equipped with global positioning systems as an information aid
for archaeological visits. In this study we will focus in the study of the
technologies used to implement these systems. To this end we analyze
an archaeological site where this systems have been tested. In this ex-
periment we have applied state of the art technologies in virtual and
augmented reality to implement a system that allows users to access the
site using their mobile devices. We conclude that the use of this kind of
technologies is an effective tool to promote the archeo-geographical value
of the site.
1 Introduction
Advances in mobile technologies are enjoyed by an increasing percentage of the
population. This is due mainly to lower prices and to the technologization of life
and work style standards of the population [1].
Most of the current communication processes are based on the use of mobile
devices. Some of the most used are tablet pc, pocket pc, smart-phone, PDA
(Personal Digital Assistant), and iPod. These technologies provide several web
tools like, search engines, virtual communities and e-advertising among others.
Adapting the power of these technologies to the field of cultural heritage,
allows the broadcast of local heritage to a worldwide level. Innovative uses of
technology can stimulate curiosity and interest in users, satisfy their informa-
tion needs and ultimately allow the creation of a digital heritage [2] [3]. These
devices can guide users in virtual or real world spaces. Virtually reconstructed
environments take advantage of information rich databases providing the users
with historical, cultural, and geographical data. In these environments the user
can better explore in an augmented reality space. This system empowers the
user giving a knowledge rich environment that facilitates learning [4] [5] [6].
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 410–420, 2008.
c Springer-Verlag Berlin Heidelberg 2008
Mobile Cultural Heritage: The Case Study of Locri 411
This type of systems can mix real and virtual worlds, allowing the combination
of the geographical location with the exact historical or cultural information. An
added tool that identifies the geographical position [8] allows the system to have a
combined view of a culturally interesting artefact with its virtual reconstruction
(3D model). This 3D model represents its original shape. The user can also benefit
by using the user friendly interface of the device, to view other multimedia data
related to the artefact, for example the reconstruction of the virtual reproduction
of the original environment and the historical source [9].
The aim of this paper it to highlight the possibility to apply these technologies
to regions such as Calabria, which is rich in cultural and archaeological resources
not always exploited. In particular we study the system at the archaeological
park of Locri and study how the tools we describe can improve and value the
enjoyment of the place.
3 System Architecture
The system we present is a program that enriches the exploration of open spaces
with additional data. The system provides real time visualization, on a mobile
device, of a 3D reconstruction of the environment. This environment also gives
navigation capabilities using its GPS data system.
To achieve this task we designed a new graphical engine for the mobile device.
The system is built on top of a set of new graphic libraries developed in collabo-
ration with the E-Guide S.R.L. which were called Q3 libraries. These C++ APIs
(Application Program Interfaces) are divided in two parts:
These two set of APIs are linked together by these other libraries:
Q3Toolkit: is the glue between the 2D layer and the 3D layer and the OS
(Operating System);
The API is built on top of OpenGL ES and OpenVG [10] libraries that are
the standard de facto in mobile environments (Many CPU manufacturers sup-
port natively these libraries on their products). A converter allows the import
of a COLLADA [11] file or a Google Earth file (Google Earth 4 files are com-
pressed COLLADA files, with textures and other information) and save them as
a compressed format specifically designed for.
The system processes GPS data to obtain the user position and move the vir-
tual environment along with the user’s movements. It is also possible to connect
a GPS with an integrated compass in order to know the user’s orientation. If the
compass data is missing the user has the possibility to move the view using the
joypad of the device. The meshes position is stored in xml format together with
other information such as a text and multimedia contents so the user can click
on any object of the world and read the description, watch images, and so on.
Since most of today’s mobile devices do not have FPU (Floating-Points Unit)
the system hasn’t the possibility to be compiled using fixed-points and a special
library for fixed-point algebra was developed.
The GUI has many graphical effects like shading, anti-aliasing and is planned
to be used using a touch screen. The 3D engine can show any textured mesh and
support multiple light effects.
Fig. 2. The zone of Centocamere, in the archaeological site of Locri (image from Google
Earth)
Inside the archaeological park of Locri, tourist can visit 3 zones: Centocamere
(Fig. 2), Museum (Fig. 3), Theatre (Fig. 4).
In these zones there are few routes that allow tourists to access the most
interesting findings. Archaeologists have excavated the ancient ruins especially in
the zone of Centocamere, where the ruins of the ancient city centre are located.
These are characterized by houses, and workshops where clay ceramics were
manufactured and sold.
For example, in Fig. 5 we show the map of the zone of Centocamere in the
park, where we highlight the possible routes:
414 G. Cutrı́, G. Naccarato, and E. Pantano
Fig. 3. Museum zone, in the archaeological site of Locri (image from Google Earth)
Fig. 4. Theatre zone, in the archaeological site of Locri (image from Google Earth)
Fig. 5. A map with the possible routes of zone of Centocamere in the archaeological
park of Locri
STEP 1: User accesses to archaeological site and decides to rent the partic-
ular mobile device or to use his own (in this case he has to download on his
device all the useful information, like maps, photos or other data);
418 G. Cutrı́, G. Naccarato, and E. Pantano
STEP 2. The user starts the application and chooses the language (Fig. 8)
and route (a fixed route from the list or personalize his own) (Fig. 9);
5 Conclusion
We illustrate how the archaeological park of Locri, and is big extension, can
be enjoyed in a more effective and efficient way by using this new system. This
system allows users to understand, learn and appreciate also parts that don’t
exist anymore, artefacts, which were destroyed by weather or man. The user
can experience an immersive and more interesting experience, especially for that
part of population which is less interested in the archaeological heritage but
more sensitive to the use of new technologies [19].
A similar system could be applied to other archaeological sites with the same
success.
6 Future Work
Next generation of mobile devices will have more powerful CPUs and many will
have a GPU also, which means that there will be no problem rendering very
Mobile Cultural Heritage: The Case Study of Locri 419
complex meshes with a low frame rate. These new kind of devices will allow to
render even more realistic scenes.
But the future is not only based on new powerful hardware. AR (Augmented
Reality) systems will play an important rule on the current scenario. Even today
a lot of mobile phones and PDAs have a camera inside which already permits to
embed real image data in order to overlap the 3D reconstruction on top of the
reality. People can experience new HCI (Human Computer Interaction) that will
permit a more interaction with the environment and many new exciting features.
The next step will be the integration of maps and a multi-modal guidance en-
gine which will permits the user to be guided throw a city using various transport
services (bus, train, taxi, etc.). These technologies are currently under develop-
ment in collaboration with the E-Guide S.R.L.
Furthermore, the system we presented will be tested and evaluated through
a quantitative analysis with consumers in the archaeological park of Locri.
References
1. Reitano, A., Pantano, E., Feraco, A.: Comunicazione digitale e gestione del terri-
torio (in press, 2007)
2. Parry, R.: Digital heritage and the rise of theory in museum computing. Museum
management and Curatorship 20, 333–348 (2005)
3. Lin, Y., Xu, C., Pan, Z., Pan, Y.: Semantic modeling for ancient architecture of
digital heritage. Computers & Graphics 30, 800–814 (2006)
4. Bilotta, E., Pantano, P., Rinaudo, S., Servidio, R.C., Talarico, A.: Use of a 3D
Graphical User Interface in Microelectronics Learning and Simulation of an In-
dustrial Application. In: Proc. 5th Eurographics Italian Chater Conference, pp.
217–224 (2007)
5. Pan, Z., Cheok, A.D., Yang, H., Zhu, J., Shi, J.: Virtual reality and mixed reality
for virtual learning environments. Computers & Graphics 30, 20–28 (2006)
6. Cai, Y., Lu, B., Zheng, J., Li, L.: Immersive protein gaming for bio edutainment.
Simulation & gaming 37(4), 466–475 (2006)
7. Bellotti, F., Berta, R., De Gloria, A., Margarine, M.: MADE: developing edutain-
ment applications on mobile computers. Computer & Graphics 27, 617–634 (2003)
8. Burrogh, P.A.: Principles of geographical information systems for land resource
assessment. Clarendon Press, Oxford (1986)
9. Gleue, T., Dhne, P.: Design and Implementation of a Mobile Device for Outdoor
Augmented Reality in the ARCHEOGUIDE Project. Virtual Reality. In: Archae-
ology, and Cultural Heritage International Symposium (VAST 2001) (2001)
10. OpenGL ES and OpevVG (2007), http://www.khronos.org/
11. COLLADA (2007), http://www.khronos.org/
12. Costamagna, L., Sabbione, C.: Una cittá in Magna Grecia Locri Epizefiri. Guida
Archeologica. Laruffa Editore. Reggio Calabria (1990)
13. Serafino, C.: Locri antica e il suo territorio. Il Portichetto. Aga- Cuneo (1991)
14. Bertacchini, P.A., Dell’Accio, A., Giambó, S., Naccarato, G., Pantano, P.: WebGIS
and tourist personalized itineraries for exploitation of calabrian cultural and ar-
chaeological heritage. In: Proc. 2nd International Conference On Remote Sensing
in Archaeology (2006)
420 G. Cutrı́, G. Naccarato, and E. Pantano
15. Bertacchini, P.A., Dell’Accio, A., Mallamaci, L., Pantano, E.: Benefits of Innovative
Technologies for territorial Communication: the Case of Study Virtual Museum
Net of Magna Graecia. In: Proc. 5th Eurographics Italian Chater Conference, pp.
181–185 (2007)
16. Bertacchini, P.A., Reitano, A., Di Bianco, E., Pantano, E.: Knoledge media design
and museum communication. In: Proc. 3rd International Conference of Museology
(2006)
17. Vlahakis, V., Ioannidis, N., Karigiannis, J., Tsotros, M., Gounaris, M.: Virtual
Reality and Information Technology for Archaeological site promotion. In: Proc.
5th International Conference on Business Information Systems (BIS 2002) (2002)
18. Mignonneau, L., Sommerei, C.: Designing emotional, metaphoric, natural and intu-
itive interfaces for interactive art, edutainment and mobile communications. Com-
puter & Graphics 29, 837–851 (2005)
19. Mason, D.D.M., McCarthy, C.: The feeling of exclusion: Young peoples’ perceptions
of art galleries. Museum Management and Curatorship 21, 20–31 (2006)
Study of Game Scheme for Elementary Historical
Education
1 Introduction
With the expansion of the game industry, game had a great impact on the society
especially for young people. Some primary and middle school students wallowed in
online games and neglected their studies. Other purpose besides entertainment of
computer games is increasingly concerned. The United States first proposed the con-
cept of "Serious Game", computer games can be used in educational, training, simula-
tion and other areas etc.
In many European countries, educational and puzzle games is developed earlier and
gained a number of specific practical applications. In 1984, the famous American
Electronic Arts (EA) was issued the "Seven Cities of Gold". One of the EA founder
Trip Hawkins combined the two words "Education" and "Entertainment" together and
create the new terminology "Edutainment"1. On the other hand, a number of websites
for online education games abroad is very extensive. Relative to the maturity of edu-
cation game abroad, there are few domestic enterprises engaged in education game.
Shanda Interactive Entertainment Limited (SNDA) developed the first domestic
youth-oriented education game which named "Learn from Lei Feng". This game has
been widely concerned among researchers.
Most of the education game is developed for primary students, the majority subjects
of education game are focused on mathematics, language courses, and many few
on science, history, geography, physics, and other courses. This paper directed at
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 421–426, 2008.
© Springer-Verlag Berlin Heidelberg 2008
422 H. Wu and X. Wang
elementary historical education, discussed the factors we should consider during the
design of education game, and research the hybrid mode of game and education. Fi-
nally, a case of education game using hybrid mode is proposed. Students can learn
historical knowledge by completing the task of the game.
Help System: Students may encounter some problems in the course of the game, such
as the lack of understanding of the rules of the game, the knowledge embedded in the
game and so on. They will not go on playing if they suspended because of the difficulties
encountered in the game, and we can not achieve the purpose of education, So the game
need to provide help system, Help information can provide from NPC, partners etc.
Students can enhance and grasp knowledge of history through help system.
Information Feedback: Playing process is also a learning process, it is important to
provide a feedback for improving the effectiveness and quality of learning. So, if we
add the feedback system in education game, the student can understand their own
shortcomings, it will help them to create confidence.
It is the ultimate goal of the education game to integrate education naturally into the
network games (especially MMORPG-online games). For historical education, it is
relatively easily to map historical knowledge to game through some elements of the
games such as tasks, skills, stags and so on. At present, education game can be divided
into small Flash game and major RPG games, RPG game has complete story, unified
modeling figures, but it is more difficult to develop, and it need a long life cycle and
high costs. Flash game has no unified story and characters shape, it is more easy to
develop, and it need low cycle and low cost. In general, flash game is suitable for young
children, and primary and middle school students are more inclined large RPG game.
Knowledge can be merged naturally into the game is limited. So the second hybrid
mode of education and game is knowledge race by online war. We provide a
large-capacity database. The form of online war game can be responder, quiz, chal-
lenge champion and so on. The result of online war can also have an impact on cur-
rencies, decorations, equipment items, the value of experience, game time of player etc.
This mode is suitable for encyclopedia-knowledge which is interesting and widely
adaptability, it can also link to the database of Entrance Examination and become an
online learning platform of practice tests. Online war games can be set up in different
districts for the local database and solve the problem of different versions of textbooks.
We can provide the interface, so the local teachers can editor the local database ac-
cording to local textbooks.
The third mode is combined with educational websites. We can put historical knowl-
edge into the websites which can include the following modules: teacher talks, network
courses, practice database, encyclopedia knowledge and so on. Educational websites
424 H. Wu and X. Wang
and online games have certain inherent link by using a unified website and game ac-
count. Learning in the educational websites will affect the game money, decorations,
equipment items, the value of experience etc.
Website and the game may not have direct link, it can be only as a reference to the
contents of the study, a reference to the answer for the game, a reference to complete
tasks.
No matter which hybrid mode we choose, primary and secondary school students
can learn academic knowledge, encyclopedia in the course of the game if we use proper
treatment methods in the game scheme. We can also join the collaborating and explore
learning mode into the game and affect ethical behavior of students on network. But the
hybrid mode of game and education should keep to be researched before it is accepted
by school students and their parents.
5 Conclusions
The hybrid mode of education and game should be widely researched before it became
an accepted educational means. This paper aimed at elementary historical education,
and designed a case of education game using proposed hybrid mode. Students are
absorbed with virtual role-play in the magic school and learn historical knowledge by
completing the task. Meanwhile, the game also provides information feedback system,
and helps a student understanding their lack of the study.
Acknowledgments
This work is supported in part by the grand foundation of science and technology de-
partment of Zhejiang province under grant No.2005C13021, Zhejiang natural science
426 H. Wu and X. Wang
References
[1] http://game.86516.com/Game/news/chanye/130154474.htm
[2] Becker, K.: Teaching with games: The minesweeper and asteroid experience. Journal of
Circuits, Systems and Computers 17(2), 23–33 (2001)
[3] Coyne, R.: Mindless repetition: Learning from computer games. Design Studies 24(3),
199–212 (2003)
[4] Durkin, K., Barber, B.: Not so doomed: Computer game play and positive adolescent de-
velopment. Applied Developmental Psychology 23, 373–392 (2002)
[5] Squire, K.: The Games-to-Teach Research Team. Design Principles of Next-Generation
Gaming for Education. Educational Technology 43(5), 17–23 (2003)
[6] Virvou, M., Katsionis, G., Manos, K.: Combining software games with education: Evalua-
tion of its educational effectiveness. Educational Technology & Society 8(2), 54–65 (2005)
Integration of Game Elements with Role Play in
Collaborative Learning — A Case Study of Quasi-GBL in
Chinese Higher Education
Abstract. The reason that most undergraduate students in China spend a lot
more time playing computer games rather than learning is not that learning is
too hard, but it is boring, while playing games is fun. How to provide appropri-
ate and interesting opportunities that can engage students and improve the
learning process? This paper discusses an innovatory education paradigm quasi-
GBL, which integrates game elements with role play in collaborative learning.
Case study of quasi-GBL in an undergraduate course “Software Engineering”
reveals that quasi-GBL is successful in achieving experiential and fun learning,
offering a variety of knowledge presentations, and creating opportunities for
knowledge application.
1 Introduction
In recent years, college teachers in China have been harassed by the too much en-
gagement of undergraduate students in computer games. According to the 7th Online
Game Research Report of China, 43% of the game players are from the age of 19 to
25, 56% of the players are undergraduate students, and the average time spent in play-
ing games is 3 to 6 hours per day[1]. On the other hand, questionnaire in a university in
China shows that 50.75% undergraduate students miss classes frequently and those
undergraduate classes that over 80% of the enrolled students attend account for only
58.61% of the total[2]. The following questions are thus proposed: Why do under-
graduate students engage so much in games rather than in learning? How to design
effective learning opportunities to respond to this challenge? According to Prensky,
what we are waiting for is the great teacher-designers to step forth, the people with
the vision to harness games in the name of fun learning[3].
Analysis of game playing experiences indicates that games may offer supports in
several aspects of the learning process: learners are encouraged to combine knowl-
edge from different areas to choose a solution or to make a decision at a certain point,
learners can test how the outcome of the game changes based on their decisions
and actions, learners are encouraged to contact other team members and discuss and
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 427–435, 2008.
© Springer-Verlag Berlin Heidelberg 2008
428 Z. Han and Z. Zhang
negotiate subsequent steps, thus improving, among other things, their social skills[4].
Games carry enormous potential to create immersive, experiential learning environ-
ments, draw students into a project and enhance their capabilities in information
processing, decision making, knowledge application, problem solving, and group
collaboration, which are still lacking with Chinese undergraduate students. Ignoring
the educational power of games dismisses a potential valuable learning tool[5].
In this paper, the authors take initiative in facilitating and improving the learning
process by integrating game elements with collaborative learning, and foster an inno-
vatory education paradigm of quasi-GBL (quasi game based learning). It is ‘quasi’
because the pedagogy is not purely game based yet at the same time bears features of
both games and collaborative learning. Quasi-GBL, offering immersive experience, is
both engaging and effective for a broad spectrum of Chinese students. The authors
apply quasi-GBL pedagogy to a hybrid undergraduate course ‘Software Engineering’
in Nankai University in China, which students report to be ‘both fun and rewarding’.
This paper first discusses learning objectives and characteristics of the course
‘Software Engineering’. Game Based Learning (GBL) is then examined and the obsta-
cles to its widespread application into formal education are discussed. It is then re-
vealed, through a case study, that quasi-GBL, an innovatory alternative to 100% GBL
and more compatible with formal education, offers effective learning environments in
the instruction of the undergraduate course ‘Software Engineering’. The study con-
cludes that time has come for games to be integrated with education and the key is to
consider how best games can be used, for which quasi-GBL is a good example.
analysis of a case of software development cycle and facilitate discussions. Case study
integrated instruction engages students more by involving them in observation and
exploration of real cases. However, according to Chris Dede, an ideal learning environ-
ment allows students to alternate between being “inside” an environment and being an
outsider looking in[9], and case study only offers the latter. Active and effective learning
in the course “Software Engineering” can only be facilitated when students are inside
immersive environments into which knowledge and skills in the subject are built and
which represent distributed professionalism in various roles. Quasi-GBL provides such
learning environments.
Games, especially computer games, are an important part of leisure lives for young
people and are becoming a part of culture as well. The most popular computer games
include role play games, real-time strategy games, shooting and fighting games, ad-
venture games, action games, puzzle games, and chess games. Games are engaging
for their features like interactivity [11], rules and goal[12], challenge[13], curiosity and
control[14], etc., which hold educational potential if managed properly. In particular,
games have a high learning value in certain educational domains that stress group
communication and collaboration, decision making, and self exploration.
There has been a longstanding rift between games and more “worthy” activities,
such as learning. It is only in recent years that people are interested in asking whether
games can offer better learning environments and how to manage them. GBL (Game
Based Learning) looks into ways to integrate games into learning, which may enhance
students’ intrinsic motivation, create immersive experiences, involve social interac-
tions and collaboration. However, with positive results of quite a few studies on ap-
plication of games to teaching and learning, university teachers still find it difficult to
integrate games into their instruction, mainly because game based learning does not
fit in with formal university education and there are few educational games available
for university students. Thus an innovatory alternative to GBL—quasi-GBL, is pro-
posed by the authors, which may exert the educational potentials of GBL and also be
applied to formal undergraduate education in China.
Quasi-GBL (Quasi Game Based Learning), a term coined by the authors, refers to the
instructional method that integrates game elements with role play in collaborative
430 Z. Han and Z. Zhang
learning. It can be applied to face-to-face (F2F) classes, where quasi-GBL takes place
in a classroom or to hybrid classes, where it relies partly on a virtual learning envi-
ronment (VLE). In this study, quasi-GBL is used in the undergraduate hybrid course
“Software Engineering” at Nankai University in China. The course is hybrid in the
sense that some teaching and learning is done in the classroom while students can also
communicate with each other, share learning materials, and engage in collaborative
production on an online learning platform.
Role play finds its application in the course as it allows students to ‘be’ certain
roles that would otherwise be inaccessible to them, thus experiencing the ways a cer-
tain type of role thinks about and solves problems. Another characteristic of role play
is that students have to work collaboratively with others and practice social and com-
munication skills. Role play holds special educational value in the course “Software
Engineering” because it distributes expertise in the software development cycle
among the roles, requiring students to learn the skills and collaborate with others in a
team. Yet the prevailing role play model in China, often used in language classes and
highlighting the stage performance of players in different roles, does not fit in well
with tasks in the course “Software Engineering”, which takes much more time and
involves scheme proposing, trials, arguments, decision making, problem solving in
group collaboration, while display of the product is only a part of the prolonged proc-
ess. In order to engage students in the complicated and prolonged tasks and retain
their interest and motivation in the process, game elements are integrated with role
play in collaborative learning.
Seven basic elements are identified in games, including goal, rule, competition,
challenge, fantasy, safety, and entertainment[15]. In quasi-GBL, these elements pene-
trate into role play and manifest themselves in forms of real problems, individual
tasks and group collaboration, scores, puzzles, awards, and replays. In the course
“Software Engineering”, students work in small groups simulating software develop-
ment teams or companies, each playing a certain role. The teacher acts as a client to
each team and starts the role play by proposing several “Easy level” tasks for groups
to choose from. Groups will also be given puzzles with various awards at each of the
four stages in the software development cycle: Requirement analysis, Design, Imple-
mentation, and Deployment. Successful solving of a puzzle can win extra scores for
the group. At the end of the task each group presents the product and related docu-
ments to the class, which will be assessed by the class and the teacher with an eye to
the original requirement. Besides, each student submits a report to the teacher evaluat-
ing his group members’ and his own performance in the task, which will also be
counted in the assessment. Then another round of role play starts with the students
changing roles or groups and the teacher, that is, the client, proposing “Normal level”
tasks. The process of quasi-GBL application to the course “Software Engineering” is
shown in Figure 1.
Quasi-GBL adopts a holistic approach to learners’ assessment: class assessment of
the group product, teacher’s assessment of the group product, and students’ evaluation
of group partners and themselves all count in it. A case study of quasi-GBL in the
undergraduate course “Software Engineering” is presented in the following part.
Integration of Game Elements with Role Play in Collaborative Learning 431
4 Case Study
The undergraduate required course “Software Engineering”, lasting for one semester
at Nankai University, is a hybrid one in which there is a 3-hour face-to-face session
every week and a virtual learning environment supporting online communications and
discussions about tasks and problems, and sharing of learning materials. 150 under-
graduate computer majored students in their third year, from the age of 19 to 22,
enrolled in the course and are divided into three classes. Quasi-GBL is applied in one
of the three classes for a semester by one of the authors as the course coordinator and
lecturer.
At the beginning of the semester students were asked to form nine groups with five
members in each and a project manager was selected in each group. The teacher, as
client, started the role play by putting forth five 6-week “Easy level” software devel-
opment projects for groups to choose from. Take Group 5 for example. The project
chosen by Group 5 said, “Please develop an online purchasing system for our com-
pany in 6 weeks. For further details as to its functions and other requirements, please
contact the client”. Project Manager 5 then organized a group meeting and guided in
role assignment within the group: three were software engineers and the other was
quality assurance. According to what they had learned in face-to-face sessions, Group
5 contacted the client (the teacher) and conducted requirement analysis in the first
week, through one face-to-face meeting and three emails. At the end of the second
week, the client put forward three puzzles for the group: Puzzle 1 was an additional
432 Z. Han and Z. Zhang
speed requirement for the system with an award of 30 points to the total score of 100
points for the product; Puzzle 2 was a compatibility requirement for the system with
an award of 10 points; Puzzle 3 was an additional interface requirement with an
award of 20 points. Project Manager 5 organized another group meeting and there
was a heated discussion on how to cope with the three puzzles in the existing devel-
opment scheme. The project manager insisted on trying to solve all the puzzles while
the three software engineers thought it impossible. Finally the project manager was
isolated and only Puzzle 1 was included in the modified development scheme, which
was later successfully solved. Group 5 went into the stage of system design from the
beginning of the third week, which took much longer time than expected, and it was
in the middle of the fifth week when they started coding. It turned out that at the end
of the sixth week when the system was distributed there was a function of “Rating the
purchaser by its purchasing record” missing. In the process other puzzles were given
and another puzzle with 20 points’ award was solved. In the seventh week for soft-
ware presentation, Group 5 obtained an average score of 122 (85 for the product and
37 plus for the puzzles) from the class and 124 (83 for the product and 41 for the
puzzles) from the teacher. Besides, the group members’ performance evaluation re-
ports were submitted, from which it was found that the project manager failed in or-
ganizing the team work and was almost isolated in the whole process, while one of
the software engineers not only contributed much more than the others and solved the
two puzzles all by himself but also did a lot of management work for the group, which
was taken into consideration in the assessment of each member in the group. Then
another round of role play started with the client proposing five 9-week “Normal
level” projects at the beginning of the eighth week and each group changed roles and
chose a project. The “Normal level” projects were more difficult than “Easy level”
ones yet the role play went on in a similar style, which was an opportunity for stu-
dents to “replay” in another role.
At the end of the semester, focus groups interviews were conducted to evaluate the
effectiveness of quasi-GBL pedagogy. Patton suggests that focus groups interviews
are essential in the evaluation process[16] and selection of interviewees is essential for
the rigor of the evaluation. According to Stewart and Shamdasani, the group must
consist of representative members of the larger population[17]. In this case, two group
interviews were conducted: one with 6 students in the class and the other with 9 stu-
dents. The authors first sub-grouped the class with reference to their participation and
activeness in the course and then carried out convenience sampling in each of the sub-
groups. 6 from the more active participation sub-group and 9 from the less active one
were selected as interviewees. Each focus group interview lasted about an hour and 5
to 7 open-ended questions were asked about the students’ experiences in the course
with quasi-GBL. The first author acted as interviewer and the interviews were re-
corded, transcribed, and then analyzed by the two authors. Analysis of the interview
transcripts reveals that students are mainly impressed by two features of quasi-GBL:
experience and fun.
According to the interviews, students believe that they have learned more from ex-
periences than from listening to the teacher in classrooms or discussing with their
classmates in case studies alone. First, they have experiences in collaboration. In
quasi-GBL, students need to assign tasks among the group, communicate with part-
ners face-to-face or online, practice team administration and team work, and try to
Integration of Game Elements with Role Play in Collaborative Learning 433
solve conflicts and problems, from which they can learn a lot about techniques and
skills in collaboration. Second, they have experiences in the real process of a software
development cycle. Immersive experience in the course projects has helped students
to better understand what the tasks and focus are in each stage of the process, how the
development work is organized and implemented, and how complicated the process
might be. Third, they have a better understanding of the diverse roles involved in
software lifecycle, including required competencies and duties for each role. More
important, some students have identified their potentials, weaknesses, and preference
through quasi-GBL, which is a crucial part of holistic development.
Aside from experience, fun is a peculiar feature of quasi-GBL, which engages stu-
dents and makes them more active in learning. According to the interviews, fun ele-
ments come mainly from collaboration, role play, and puzzle solving. Most students
enjoy social interactions with group partners and draw much fun from sharing of
ideas, information, problems, and success. Role play also offers fun because it helps
students out of the boring memorization and they can complete learning through prac-
tice and experience, in which there are a variety of real decisions to make, problems
to solve, and responsibilities to take. Puzzle solving, with challenges and the possible
award, is reported to be the most fun in quasi-GBL. As one student said in the inter-
view, “Puzzles are really engaging and I can spend hours thinking about it and trying
to solve it, because it is sort of a test on how capable I am.”
4.3 Discussions
With puzzles added to the role play, students report more fun and motivation in
learning, as one student said, “I am more active in learning than before and every
member in our group feels the same. We like puzzles, and it’s challenging and
fascinating to solve them by ourselves.” In solving the puzzles, students access re-
sources, apply related knowledge they have learned, and engage in collaboration and
communications. They willingly spend a lot of time on puzzles, just like when they
are playing a game.
Score element in the role play is another incentive to students. After the seventh
week when their products were scored by the class and the teacher, students had a
higher sense of achievement and social belonging, as one student wrote in his per-
formance report, “We were excited on the day when presenting our group product to
the class and more excited to know how many points we’d got for it. It’s different
from knowing my score of the exam in other courses, as this was the joy that I could
share with my partners at what we’d been working on for the last several weeks. And
the score also gave us motivation to work harder and achieve higher marks in the next
project”. Some have complained about Chinese students’ caring about their marks in
exams more than anything else, yet score element in the role play has re-directed their
attention from memorizing information for the exam to making a better product and
solving the puzzles, in which process real learning happens while they are analyzing,
arguing, and trying. Scoring other group’s products is also a significant experience for
students. They can learn a lot from looking at other groups’ products, as one student
said in the interview, “Learning how others solve problems that I have little idea of is
a wonderful experience, which gives me a lot of insights and inspiration”. Besides,
434 Z. Han and Z. Zhang
the responsibility of assessing other groups’ products makes students more active in
learning as they feel they have their say in learning.
A problem students have in team work is how to deal with different opinions and
ideas without impeding the completion of tasks. In Group 5 there was so heated ar-
gument between the project manager and the system engineers that later on the project
manager was isolated and the system engineers communicated with the client direct
and completed the task by themselves. In the performance report, the system engi-
neers complained about the project manager’s “not listening to others, arbitrary, and
not cooperative”. The interesting part of the story is that in the “Normal level” project
when the former project manager became a system engineer, he seemed to have
learned something about team work as he wrote in his performance report about his
own alteration, “I’ve understood how my partners felt in the first project because now
I’m a developer. They need my suggestions, not orders”. Team spirit and capability in
collaborating with others are part of essential qualities for people in general, and more
in software development industry as the division of labor in the software development
cycle is becoming ever delicate. Experience in the role play can help students to iden-
tify their strengths and weaknesses in team work, which is the starting point to im-
provement and self development.
A special feature of the collaboration in the case is that it is hybrid, that is, both
face-to-face and online. Online collaboration is different from F2F collaboration, for
lack of real time feedback, gestures, expressions, and body language, just to name a
few. It is interesting, thus, to notice that in the course, most communications online
were via chatting systems while emails or online forums are seldom used between
students. When asked for the reasons, students said that they’d tried to use online
forums yet found them to be so “inefficient”. With software development becoming
more and more globalized, the ability to work collaboratively online with partners is a
necessity for future professionals. Immersion in such an experience is an important
benefit for the students.
5 Conclusions
Game based learning has not found many applications in schools or universities,
universities in particular, because it does not fit in well with curriculum of formal
education and schools are skeptic about whether it brings more benefits than harms.
Quasi-GBL, an alternative to GBL, which integrates game elements with role play in
collaborative learning, is an innovatory education paradigm in realizing “fun learn-
ing” and brings broad learning benefits to students. In the undergraduate course
‘Software Engineering’ studied in this paper, it is found that students enjoy the expe-
riential learning process, and more important, they are engaged in learning more,
because learning becomes more fun. Students have also benefited in a broader spec-
trum through quasi-GBL, in aspects of collaboration and team work both face-to-face
and online, communications with people, utilization of information for problem solv-
ing, planning and decision making, not by memorization but through experience.
Quasi-GBL can also be applied to other courses in higher education, whether a face-
to-face course or a hybrid one, and more studies need to be carried out in such field.
Integration of Game Elements with Role Play in Collaborative Learning 435
References
1. iResearch Consulting Group.: The 7th Online Game Research Report in China. iResearch
Consulting Group, 9-12 (2007)
2. Wen, K.: Countermeasure research and cause analysis of the existing college students’
study-weary phenomenon. Journal of Tianjin University of Technology and Education 12,
58–61 (2006)
3. Prensky, M.: Digital Game-Based Learning. McGraw-Hill, New York (2001)
4. Pivec, M., Dziabenko, O., Schinnerl, I.: Aspects of game-based learning. In: I-KNOW 03,
the Third International Conference on Knowledge Management, Austria, pp. 178–187
(2003)
5. Oblinger, D.G.: Games and Learning. Educause Quarterly 3, 5–7 (2006)
6. Wang, X., Wang, J.: Case Study in Software Engineering Instruction. Computer & Infor-
mation Technology 6, 114–118 (2006)
7. Wu, H.: Application of Case Study in Software Engineering Instruction. Journal of Yichun
University 4, 98–99 (2007)
8. Ye, J., et al.: Research into Case study in Software Engineering Instruction. IT Educa-
tion 7, 19–21 (2007)
9. Dede, C.: Planning for Neo-Millennial Learning Styles: Implications for Investment in
Technology and Faculty. In: G.D., Oblinger, J.L. (eds.) Educating the Net Generation,
EDUCAUSE, Boulder, Colo (2005)
10. Kirriemuir, J., McFarlane, A.: Literature Review in Games and Learning. A Report of
NESTA Futurelab, http://www.nestafuturelab.org/research/reviews/
08_01.htm
11. Tornton, G.C., Cleveland, J.N.: Developing managerial talent through simulation. Ameri-
can Psychologist 45, 190–199 (1990)
12. Johnston, R.T., Felix, W.: Learning from video games. Computer in the Schools 9, 199–
233 (1993)
13. Baranauskas, M., Neto, N., Borges, M.: Learning at work through a multi-user synchro-
nous simulation game. In: Proceeding of the PEG 1999 Conference, pp. 137–144. Univer-
sity of Exeter, Exeter, UK (1999)
14. Malone, T.W.: What makes computer games fun? Byte 6(12), 258–277 (1981)
15. Alessi, S.M., Trollip, S.R.: Multimedia for Learning: Methods and Development. Allyn
and Bacon, NY (2001)
16. Patton, M.Q.: Qualitative evaluation and research methods, 2nd edn. Sage, London (1990)
17. Stewart, D.W., Shamdasani, P.N.: Focus groups: Theory and practice. Sage, London
(1990)
A Case of 3D Educational Game Design
and Implementation
Center for Research on EduGame, Nanjing Normal University, Nanjing, 210097, China
nj_huizi@sohu.com
Abstract. Games are increasingly used for educational purpose. Player can
obtain experiences through playing game. Game is a safe and efficient way for
security education. This paper introduces a game design: Escapre from Fire. a
design procedure and system structure are outlined. Related programming tech-
nologies are discussed.
2 Related Work
2.1 911: First Responsible
Some fire games have been developed. In 911: First Responders Fig. 1 , the( )
player acts as a fire commander in a fictional rescue and catastrophe management
organization. The player can command a number of vehicles and staff to deal with all
the accidents. For example, while the firemen are fighting with the fire, the doctors
are required to cure the wounded. The command and the firemen can accumulate their
reputations to become the top fire department [3].
( )
Fire Captain: Bay Area Infer Fig. 2 is a fast-rhythm game. Based on two true fire
accidents. One is about fire of Holand firework factory in 1993, and the other is
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 436–441, 2008.
© Springer-Verlag Berlin Heidelberg 2008
A Case of 3D Educational Game Design and Implementation 437
Auckland forest fire in 1996. The role of the player acts as a fireman to deal with all
of the emergences [4].
Fire happens continually. There are 6~7 millions fires every year in the earth, more
than 70% of which are house fire. In china, nearly 20000 millions RMB per year is
lost because of fire and more than 2,000 people die of fires. We are developing a 3D
fire game Escape from fire to teach children and the teenage how to escape from fire.
Different from 911 and Fire Captain, the player act as victim in the fire spot, not
the fireman. From educational perspective, we summarized some features of this
game, as the following shows.
Realistic style. It means the results form the player’s operations accord with facts.
The game is based on real fire cases, such as Tokyo Star 56 Mansion Fire in 2001, the
Sinkiang Kelamayi Fire in 1994 , China and others from Discovery Program.
Time limit is another key character of this game. The fire will spread quickly and
the player is given limmited time to finish the task.
Feedback. The players may fail during playing. Once they fail, the information
about the wrong operation will come out.
438 H. Shi, Y. Li, and H. You
Form game perspective, challenge, fantasty, curiosity and are the main features. [4]
Unlike some eductional game, in which the knowdge is designed as questions and the
only chanlenge for the play is to answer questions. This game is well structed , the
player has to make decidions and take actions to sovle problems. Thus, the player
construct their knowledge by exploring[6]. That is learning by doing, not by answer-
ing questions.
How to combine education in game without lose of fun? It is a main difficulty for us,
maybe for all of those educational game designers. Although there is no unique guideline
for educational game design, all of them begin with instructional analysis, shown as fig. 3.
Events
Content Player’s activities
Analysis Play’s background
Instructors
Interface design
Players
Document
cares about what makes a computer application fun to operate. Based on the result of
content analysis, game’s story, characters, style, rules and programming tools etc. will
be discussed in this step. This game is a strategy-based game, in which the player acts
as a boy Dingding. He and his friend Xiaoqiang who is a clever but timid boy will
encounter different fire. They have to put out the fire or escape from the fire. The
OGRE game engine is decided.
4 Engine Analysis
ORGE is an open source game engine. It is mature, steady, credible, flexible and it
can produce most of the effects that other popular engines can do, such as object,
HDR, knaggy texturing etc,. Different from other engines, scene graph is separate
from content objects (Fig. 4). Thus , it is simple to produce engine-based realtime 3D
knaggy animation.
All of this geometry and these rendering properties are made available to the scene
graph via MovableObject. Instead of subclassing the movable object from a scene
node, it is attached to the scene node[ 8]. It is not necessary to set up the relationships
among the scene nodes, and know about the details of the renderble objects. Expected
functions can be carried out through operating on the scene graph interface. The scene
graph interface can even change completely and the content classes would not be
affected in the least.
How to produce real time animation in OGRE. First, we should built 3D models
,
and animation in 3D modeling software(Maya 3Dmax, etc,.). Second, these 3D
models and animations are exported as *.Mesh and *.skeleton files with proprietary
exporing tool of OGRE. Then, these *.Mesh and *.skeleton files are attacted to the
scene nodes. Thus, the realtime animation can be created in OGRE.
440 H. Shi, Y. Li, and H. You
The following code is about how to add a node in mSceneMgr and attach a existing
model called jaique to the node. The final rendering is shown as Fig. 5
Then, we set the playing model as spline interpolation, shadow effect as “true”, ob-
ject’s emergencing coordiate, object’s circumvolving angle and axis in CreateSkele-
tonAnimation(). Then, animation information in jaiqua.skeleton is added to the
nodels. Sneak actions form jaiqua.skeleton is given to jaiqua model.
void CreateSkeletonAnimation()
{
Animation::setDefaultInterpolationMode(Animation::IM_SP
LINE);
Entity *ent = mSceneMgr->createEntity("M1",
"jaiqua.mesh");
ent->setCastShadows( true );
SceneNode* BJmovie = mSceneMgr->getRootSceneNode()-
>createChildSceneNode("R1",Vector3(3,0,-3));
BJmovie->attachObject(ent);
A Case of 3D Educational Game Design and Implementation 441
BJmovie->yaw(Degree(-90));
mAnimState = ent->getAnimationState("Sneak");
mAnimState->setEnabled(true);
}
5 Future Work
The games are still on the tentative stage and need the further improvements in tech-
nology(modeling techniques, interaction technology etc.) and designin. We are to
improve the system so that it can provide three alterative game levels----from level 1
to level 3. The player can select game level by their own. For those experiened play-
ers, level 3 may be suitable, being most complex and challenging. For children and
some girl players, level 1 may be suitable, bejing simple and less challenging.. In
addition, we will try these games in schools and do some research so that we can
evaluate and modify the games from player perspectives. It is urgent for us to attach
importance to security education. We are planning to develop a series of games on
security education for children, such as traffic and electricity
References
1. Robertsona, J., Howells, C.: Computer game design: Opportunities for successful learning.
Computers & Education 50(2), 559–578 (2008)
2. Garris, R., Ahlers, R., Driskell, J.: Games, motivation and learning: A research and practice
model. Simulation and Gaming 33(4), 441–467 (2002)
3. http://www.atari.com/us/games/911_first_responders/pc
4. http://www.firecaptain-game.com
5. Malone, T.W.: What makes things fun to learn? Heuristics for designing instructional com-
puter games. In: Proceedings of: 3rd ACM SIGSMALL symposium and the first SIGPC
symposium on small systems, pp. 162–169 (1980)
6. Liang, T.: A 3D Escape-from-Fire System. Journal of Shenyang Agricultural University
(Social Sciences Edition) 9(5), 779–782 (2007)
7. Ebner, M., Holzinger, A.: Successful implementation of user-centered game based learning
in higher education: An example from civil engineering. Computers & Education 49(3),
873–890 (2007)
8. Junker, G.: ProOGRE 3D Programming, p.39 (2006) ISBN-13: 978-1-59059-710-1
Mathematical Education Game Based on Augmented
Reality
Abstract. The computer game industry has been grown rapidly due to the
development of graphics and communication industry. The computer games
have been taking on various forms with its development. Currently, the com-
puter games are not only used for amusement but also used for other purposes.
The computer games that try to provide users more than fun are called serious
games. Many researchers have developed serious games for various fields such
as education, medical treatment. Education especially stands out among those
fields. This paper suggests a mathematical education game developed using
Augmented Reality. It is a board game for kindergarten and elementary stu-
dents. Augmented Reality is used to extend user's experience and to increase
the usability of the system.
1 Introduction
Computer games have been gained a lot of attention nowadays. Computer games
have become one part of the culture. Currently, computer games are not only de-
veloped for simple amusement. Some developers try to obtain something more than
just simple fun. These games are developed to train users the specific skills while
enjoying them. These types of games are called serious games [2]. Serious games are
also called functional games or social impact games, and their focus is to provide
users with extra benefit and fun.
Especially educational games are gained huge attention from developers and users.
There are many supports from schools, corporations, and institutes, and many re-
searchers develop educational games in the various fields [1].
Educational games that combined learning and playing are very practical, and they
become new entertainment contents. Already there are various ways to present contents
that deal with educational games. The reason these educational games are growing is
not because they are games that combine enjoyment and education, but because they
are recognized as having beneficial contents. Edutainment is already being used in
variety of fields for educational purposes, such as educational games, livelihood
training, and politics education.
The purpose of this paper is to introduce an educational game using Augmented
Reality (AR) technologies. AR is used to extend user’s experience and the usability of
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 442–450, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Mathematical Education Game Based on Augmented Reality 443
the game. The board creation tool is also provided to increase users’ interest by helping
them to create their own contents.
The organization of this paper is composed as stated. Section 2 introduces related
research areas. Section 3 explains the technical elements and practical usage plan of the
proposed educational game. Section 4 concludes with a description of future plans.
2 Related Research
Research regarding educational game has been done actively, so many new educational
games are developed. Here we only deal with online educational games and AR-based
educational games, which are mostly related to the proposed game.
ALEPH Project Game is a good example of an educational game (Figure 1) [3]. Here,
the matrix of English words moves its position, and players must guess the correct word
using 12 different attacking options. This game requires strategic techniques from
players, and players can beat the opponents by using turn attack. Through this game’s
process, players can train educational effect by practice English words repeatedly.
(a) (b)
Fig. 1. ALEPH Project (a) Fight scene (b) It is a scene that make word as alphabet
(a) (b)
Fig. 2. Power politics (a) Candidate selection (b) Manage election campaign
AR is a good environment for serious games. AR can maximize the player’s experience.
As a representative example, there is “Solar-System and Orbit Learning in Augmented
Reality System” (Figure 3) [4]. This system promotes knowledge of volcano eruptions,
relations of the solar system, information of Earth’s surface and others through a camera.
(a) (b)
(c)
Fig. 4. wlzQubesTM (a) Main scene (b) Selected item (c) Two cubes
The Cooking System [8] teaches cooking recipes, and AR Squash Game [6] com-
bines AR and a sports game. Geometry System [7] allows players to use the virtual
interface to control the size of the shape, move, and combine, separate, and so on..
Chemistry education System [9] combines and separates atoms chemically. These
systems are combined with education in a broad and variety of ways.
The scenario of the proposed game is created based on the book Ria’s Math Play [16]
(Figure 5). This book introduces the concept of addition and subtraction to young
children through playing a dice game of two players. The dice game uses three dice,
(two numbered dice and one operator dice) and the board. Players roll the dice and
calculate the outcome, and move a piece on the board according to the outcome. The
player who arrives at the finishing point first wins the game. Since the dice game re-
quires computation, parents or another helper has to play together with the children to
teach the mathematical concepts. Children can only play with one board, potentially
446 H.S. Lee and J.W. Lee
(a) (b)
Fig. 6. Computer game based on Ria’s Math Play (a) Initial scene (b) Show scene that derive
from value of three dice on the board
The proposed mathematical educational game targets young children. We assume the
proposed game will be played in the living room with the TV, the computer and the PC
Camera. The proposed game is played with three dice and the board made of markers
(Figure 7). The board is created using multiple markers, so the system can augment
virtual images even though a few markers are occluded by pieces of the game or
players.
Mathematical Education Game Based on Augmented Reality 447
The number of markers used for the board is 15, which is decided by the viewing
angle of the camera and the distance between the camera and the board. The size and
the number of the markers are decided for reliable tracking based on the setup in which
the game is played on.
A different board can be selected according to scenario and user’s preference among
the given boards, and player’s piece are augmented as 3D models on the selected board
(Figure 8).
(1)
448 H.S. Lee and J.W. Lee
(2)
The rotation matrix is presented as the XYZ method [17] in the Euler Angle ex-
pression. Angle distances between Z axis of each viewable marker of the dice and the
board are computed, and one with the smallest angle distance within the given thresh-
old is chosen as the Z axis of the top marker of the dice (Figure 10). Using this
information, the system computes the outcome and overlay the corresponding infor-
mation on the dice.
The system also provides a tool allowing users to create their own board and
player’s pieces to improve game’s usability. The board is created by combining ele-
ments of the tool. Users can place a square, circle, polygon, straight line, rectangle,
ellipse or other provided elements and can fill each component up with the given color
or insert their own image on each component and the background. For free drawings,
players can use a free line to draw their name or picture as they want (Figure 11). Using
this tool, users can create their own games.
Mathematical Education Game Based on Augmented Reality 449
Fig. 10. Relationship between the board coordinate (GMC) and the operator marker or numbered
marker coordinate (BMC)
(a) (b)
Fig. 11. Player defined board (a) Initial scene (b) An example
4 Conclusion
Game has been changed in variety of forms as it develops. Through this development, it
shows that it can be applied to new purpose instead of being just a simple game. Within
this, an educational game has been one of the categories that have been continually
developed because it provides enjoyment and education.
This paper proposes an educational game in the Augmented Reality environment. In
order to increase variety of computer games and learner’s participation, AR is consid-
ered as a suitable environment. This is not limited to the proposed mathematical board
game. If educational fields and AR technologies are brought together, learners can ex-
perience and learn while having fun; this can maximize the educational efficiency. The
proposed mathematical educational game increases the enjoyment of learner and the
educational effect, but the reliability requires improvement. The distance between
camera and marker and various lighting condition cause unstable operations. Moreover,
if the markers registered in the board are covered by the operator marker, the numbered
marker or player marker, and can not be seen through the camera; virtual objects are not
reliably overlaid. In order to use the system in real life, we have to solve these problems.
Acknowledgments
This work was sponsored and funded by Seoul R&BD Program.
450 H.S. Lee and J.W. Lee
References
[1] Lee, J.-b.: The present of serious game. Game industry journal, Korea, 10–14 (2007)
[2] Lee, S.-h., Jang, j.-w., Jun, c.-u.: The impact of serious games on the education of social
problems. Game industry journal, Korea, 88–104 (2007)
[3] Son, B.-g., Go, B.-s., Gye, B.-g.: Strategy puzzle game: ALEPH, 1–21, 47–60 (2005)
[4] Woods, E., Billinghurst, M., Looser, J., Aldridge, G., Brown, D., Garrie, B., Nelles, C.:
Augmenting the science centre and Museum Experience, Graphite, Singapore, 230–236
(2004)
[5] ZhiYing, S.Z., Cheok, A.D.: wlzQubesTM -A Novel Tangible Interface for Interactive
Stroytelling in Mixed Reality. In: ISMAR, Japan, pp. 9–10 (2007)
[6] Lee, S.-H., Choi, J.-S.: AR Squash Game. In: ISMAR, Japan, pp. 4–8 (2007)
[7] Do, V.T., Lee, J.W.: Geometry Education using Augmented Reality. In: Mixed Reality
Entertainment and Art Workshop2 in ISMAR, Japan, pp. 11–15 (2007)
[8] Jang, H.-b., Kim, J.-w., Lee, C.-w.: Augmented Reality Cooking System Using Tabletop
Display Interface. In: International Symposium on Ubiquitous VR, Korea, pp. 36–37
(2007)
[9] http://ieeexplore.ieee.org/iel5/9382/29792/01357659.pdf?tp=&isnumber=&arnumber=13
57659
[10] http://www.hitl.washington.edu/research/shared_space/download/
[11] http://www.powerpolitics.us/home.htm
[12] http://www.food-force.com
[13] http://www.nintendo.co.kr/www/soft/brain_01.php
[14] http://news.bbc.co.uk/1/hi/education/1879019.stm
[15] http://www.mxrcorp.com/
[16] http://www.hansmedia.com/
[17] http://www.euclideanspace.com/maths/geometry/rotations/euler/index.htm
Game-Based Learning Scenes Design for Individual
User in the Ubiquitous Learning Environment
1 Introduction
In recent year, ubiquitous learning extends e-learning from indoor to outdoor, moreover,
not like most of mobile learning applications which usually only provides the knowl-
edge of single domain in particular environment, the ubiquitous learning emphasizes to
offer learners interdisciplinary knowledge domains in the real world [1][4]. Mobile
learning provides both teachers and learners a new learning way in the e-learning field,
however, there is still an unsolved research issue, which is the flexible learning issue.
The learners' learning activities will be limited in the specific learning environment
and/or the specific domain knowledge arranged in advance. Ubiquitous learning not
only extends e-learning from indoor to outdoor but also extends mobile learning from
specific learning environment and specific knowledge domain to anyplace and multi-
discipline [5].
Ubiquitous learning also provides the learning activities which allow learners to
observe and touch the learning objects in the real world based on learners’ choices and
preferences. If learners want, they still can do suchlike solving problems, answering
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 451–462, 2008.
© Springer-Verlag Berlin Heidelberg 2008
452 S. Wu, M. Chang, and J.-S. Heh
questions, expressing and organizing questions, and even brainstorming, during their
learning processes [2][3].
Some researchers have thoughts that there are four characteristics of games could
enhance learning effects: (1) challenges; (2) fantasy; (3) curiosity; and, (4) control [6].
Regarding the first characteristic, the challenges, Malone has offered the instructional
activities which have a various difficulty levels to different learners according to
learners' abilities [7][8]. The different learners then will face different challenges.
Active learning makes learners learning better because they are doing what they feel
interesting and they feel that they are controlling everything. Games can encourage
learners to learn after classes. Learners are usually 'actively' to play their favorite games.
So, it might be work if we apply game concepts, scenes and scenes switching, to help
learners 'active' learn and have fun in the ubiquitous learning environment.
The research analyzes and constructs learning scenes to offer learners personalized
game-based learning services in a ubiquitous learning environment, for examples, a
museum and a zoo.
According to personalized context-aware knowledge structure, different learning
scenes can be extracted for different learners. Through personalized scenes construc-
tion, the learning service can extract the learner’s learning preferences and design the
learning path for individual learner.
Section 2 introduces the related works of ubiquitous learning environment,
game-based learning, different knowledge structures, and different learning strategies.
The relations and definitions between the elements of the learning scene are described
in Section 3. Section 4 uses a way to construct learning scenes and introduces how to
switch between the scenes. In Section 5, a real example of ubiquitous learning in the
museum with learning scenes is built to demonstrate the effects of this research. Finally,
Section 6 makes a conclusion and discusses the possible future works about individual
scenes layout and switching.
2 Scene Analysis
For storing knowledge and learning object information, this research defines "learning
scene" to record all learning locations and objects in the ubiquitous learning environ-
ment. Besides, the scene is also used to gather the information about what learning
objects exist in different learning locations.
A scene will cover many learning areas; each learning area is a physical place, for
example, an exhibition room in a museum. In each learning area, there is at least one or
more learning spot. A learning spot may cover one or many learning objects. Take a
museum for example, there are many exhibition rooms (the learning areas). Each room
has at least one learning spots. When the visitor stands at each learning spot, he/she can
see one or many artifacts (the learning objects).
There are two major elements in a scene as Figure 1 shows: (1) Learning spot (spot),
indicates a learning location which covered the learning objects in a personalized
knowledge structure; (2) Learning object (object), denotes possible learning objects for
the learning spot.
Game-Based Learning Scenes Design for Individual User 453
The personalized context-aware knowledge structure then can be used to build learning
scenes for learners. Because different learners with different preference will feel dif-
ferently even when they are looking the same learning objects in the real world, hence,
the suitable learning scenes to the learner can be built once the ubiquitous learning
system has the personalized knowledge structure. Moreover, different learning activi-
ties can be given to learners depends on their preferences and learning objects' char-
acteristics.
In Figure 2 there are two learners, Alex and Stis, have different viewpoints and/or
preferences about the artifacts in the museum. The top part of Figure 2 shows Stis'
personalized context-aware knowledge structure. Stis prefers the characteristic, "Dy-
nasty", rather than "Color", hence, his personalized knowledge structure is constructed
based on "Dynasty" characteristic. On the other hand, Alex has more interest in "Color",
therefore his personalized knowledge structure root is constructed based on "Color"
characteristic.
According to Wu's personalized knowledge structure [10], the scene subject for Stis
is "Dynasty". The bottom part of Figure 2 shows how the ubiquitous learning system
selects all learning objects related to the subject, "Dynasty", which Stis interests with to
form a learning scene for Stis. Similarly, Alex's learning scene can be also constructed.
As the bottom part of Figure 2 shows, there are many learning objects matched the
preference of a learner. The next problem is what learning object should the ubiquitous
learning system suggests the learner learning and observing first.
Figure 3 shows an example of two learning scenes cover three learning spots and
some learning objects. Assuming a learner stands at the learning spot2. At learning
spot2, the learner can find two learning objects, {LO2, LO3}. Each learning object has
its own characteristics about "Dynasty" subject. For example, the learning object LO2
may have characteristic suchlike "Items of daily use were specially produced for the
Ming imperial court". Moreover, the degree of realizing these story-like characteristics
454 S. Wu, M. Chang, and J.-S. Heh
represent the learner's degree of mastery in the specific characteristic. When the
overall probability of a learning scene achieved a threshold, the ubiquitous learning
system can say that the learner clears the stage or levels up, and ready to challenge
another stage or level (another learning scene).
This research develops a way to extract scenes from the personalized knowledge
structure. The whole scene construction flow involves 5 phases as Figure 4 shows:
Phase I: Learning materials analysis, the learning objects in real environment have a
lot of information (or also called characteristics), which might cover different domains
or places. In this phase, we need to analyze the learning objects and its characteristics in
the real world first. After this phase, the context-aware knowledge structure can be
built.
Phase II: Basic personalized context-aware knowledge structure construction, be-
cause the personalized context-aware knowledge structure should be created according
to the learners’ preferences and interests. The basic personalized context-aware
knowledge structure for individual learner can be constructed depends on which stories
the learners interest with.
Phase III: Personalized context-aware knowledge structure refinement, even if two
learners choose the same story in phase II, their interests still might be a little different.
In order to precisely refine the personalized context-aware knowledge structure, the
system asks the learner some advance questions about the characteristics of the learning
objects which are involved in the story the learner has chosen. The personalized context
knowledge structure is then refined according to the learner’s feedback.
Phase IV: Personalized context-aware knowledge structure generation, after re-
peating the phase II and III, the correspondent learning objects and characteristics the
learner might need and/or interest with are clear to the system. The learner’s person-
alized context-aware knowledge structure is then can be generated in this phase.
Phase V: Learning scenes construction, the ubiquitous learning system can extract
the learning objects that the learner may prefer according the personalized con-
text-aware knowledge structure. Because the personalized knowledge structure can
represent different knowledge domain and the learning objects locate at different lo-
cations, the ubiquitous learning system then construct the game-based learning scenes
based on the distinction features of the selected learning objects, such as "location" and
"domain", as Figure 4 shows.
This research uses rough set theory [9] to develop scene switching methodology. In the
rough set theory, data will be divided into three sets, including the Positive Set (POS),
the Negative Set (NEG), and one or more Boundary Sets (BNDs). This research defines
the positive set data as the learning objects which have more than one characteristic the
456 S. Wu, M. Chang, and J.-S. Heh
learner is interesting with, the learner has to observe the learning objects in the positive
set. On the contrary, the learning objects in the negative set don't have any character-
istic that the learner interest with. Unlike the positive set and the negative set, there are
many boundary sets in the learning scene. The learning objects in a boundary set also
have the characteristics that the learner interests with. The difference between the
positive set and the boundary set is that all of the learning objects in the same boundary
set have the same characteristics, and the learner only needs to observe one of the
learning objects for the same boundary set when he/she does learning activities in the
ubiquitous learning environment.
Figure 5 shows an example of using rough set to categorize learning objects in-
volved in learning spots. At top part of Figure 5, there are two learning scenes cover
three learning areas. Each learning scene has one or more learning spots, and each
learning spot contains one or more learning objects. In the middle part of Figure 5,
every learning object has many characteristics. The learning objects can be categorized
into three sets. Table 1 lists a summary of the sets and its learning objects.
From Table 1 the ubiquitous learning system knows that the learner has to observe
the learning object LOa. If the scene only includes four characteristics, including
Game-Based Learning Scenes Design for Individual User 457
"Anonymous", "Red", "Ch'ing", and "Blue", then the ubiquitous learning system can do
scene switching after the learner finished observing the LOa and LOb.
4 Complete Example
4.1 Game-Based Learning in Museum
By using the game concept to help learner learning in the ubiquitous learning envi-
ronment, the first thing is to define the learner profile suchlike "ID", "ability", and
458 S. Wu, M. Chang, and J.-S. Heh
"major preference". The major preference includes the characteristics in the personal-
ized context-aware knowledge structure suchlike "dynasty", "color", "function", and
"author", in Figure 7. Ability is the learning object characteristics which learner has
been observed. ID is the learner's name.
4.2 Scenario
This paper takes learning in a museum for example. In the museum, there are many
artifacts with different subjects and/or topics exhibiting in the different rooms. For
example, the Room 203 as Figure 6 shows, the "gray area" represents different subjects
and/or topics (learning spots) suchlike "Elegant Living", "In Search of the Ancients",
and "The Beauty of Nature"; "black circle" represents the artifacts (the learning objects)
in the room; the number is the artifacts number. There are eleven learning objects in
Room 203. This research considers a room as a learning area.
According the Phase I in Figure 4, Figure 7 shows the analysis results of learning
objects. Figure 7 represents how to use context-aware knowledge structure to store a
museum's learning objects and its characteristics.
The personalized context-aware knowledge structures and game-based learning
scenes are constructed based on with the learners' choices and their answers about the
questions of artifacts' characteristics as Figure 8 shows. The width or depth of person-
alize context-aware knowledge structure depends on how much the learners' interest
with and how exquisite they know.
Figure 9 shows the process of constructing the learning scenes after the learner
answered the questions. After the learner answered the questions, the ubiquitous
learning system can revise the personalized context-aware knowledge structure and
find the preference learning objects and interesting characteristics from the personal-
ized context-aware knowledge structure. Furthermore, using the personalized knowl-
edge structure to realize the learning preference objects' locations, defining the learning
spots, and constructing the learning scenes for individual learner as the bottom part of
Figure 9 shows.
Game-Based Learning Scenes Design for Individual User 459
With the learning scenes, this research uses the distances to plan the observation se-
quence for the learning objects in the positive set and boundary set that the learner may
interest with.
The ubiquitous learning system then first picks the learning objects in the POS set,
including "Ruby-red kuan-yin Tsun vase", "Two round conjoined containers in carved
red lacquer", "Tripitaka in Manchu", "Archives of the Grand Council", and "Archives
in Old Manchu". The learning path has to route around the POS set learning objects.
Figure 10 uses the "white ellipse" to represent the learning objects in the POS set, the
must be observed learning objects. The scene can be switched from the Function Scene
to the Document Scene after the learner has finished observing the white ellipse
460 S. Wu, M. Chang, and J.-S. Heh
learning objects. According to the distance, the ubiquitous learning system generates
the learning guidance path: start Æ"Ruby-red kuan-yin Tsun vase"Æ"Two round
conjoined containers in carved red lacquer"Æ"Archives in Old Manchu"Æ"Archives of
the Grand Council"Æ"Tripitaka in Manchu".
5 Conclusions
There are still some research issues could be discussed and done in order to improve the
scene switching and learning path generation. For example, as we mentioned, the
challenge issue. Currently, the game-based ubiquitous learning system uses probability
as the measurement of learner (player) abilities. The next step is how to provide an
automatically generated and non-interrupt adaptive test for such kind of ubiquitous
learning environment. Regarding the learning path generation issue, the current re-
search and system uses only distance between the learning objects. As we all know, it is
just the simplest solving way, in the future research, other factors suchlike learning
scenes should be taken into consideration.
References
[1] Chang, A., Chang, M.: Adaptive Mobile Navigation Paths for Elementary School Stu-
dents’ Remedial Studying. In: Proceedings of the IEEE International Conference on
Interactive Computer Aided Learning (ICL 2006), Villach, Austria, September 27-29
(2006)
462 S. Wu, M. Chang, and J.-S. Heh
[2] El-Bishouty, M.M., Ogata, H., Yano, Y.: Personalized Knowledge Awareness Map in
Computer Supported Ubiquitous Learning. In: Proceedings of the 6th IEEE International
Conference on Advanced Learning Technologies (ICALT 2006), Kerkrade, The Nether-
lands, July 2006, pp. 817–821 (2006)
[3] Felder, R.M., Brent, R.: Learning by Doing. Chemical Engineering Education 37(4),
282–283 (2003), Retrieved on January 15, 2008, from http://www4.ncsu.edu/
unity/lockers/users/f/felder/public/Columns/Active.pdf
[4] Hwang, G.-J.: Criteria and Strategies of Ubiquitous Learning. In: Proceedings of the IEEE
International Conference on Sensor Networks, Ubiquitous, and Trustworthy Computing
(SUTC 2006), Taichung, Taiwan, June 2006, pp. 72–77 (2006)
[5] Hall, T., Bannon, L.: Designing ubiquitous computing to enhance children’s learning in
museums. Journal of Computer Assisted Learning 22(4), 231–243 (2006)
[6] Lepper, M.R., Malone, T.W.: Intrinsic motivation and instructional effectiveness in com-
puter-based education. Aptitude, Learning, and Instruction, III: Conative and Affective
Process Analysis, pp. 255–286. Erlbaum, Hillsdale, New Jersey (1987)
[7] Malone, T.W.: Toward a theory of intrinsically motivating instruction. Cognitive Sci-
ence 5(4), 333–369 (1981)
[8] Malone, T.W., Lepper, M.R.: Making learning fun: A taxonomy of intrinsic motivations
for learning. Aptitude, Learning, and Instruction, III: Conative and Affective Process
Analysis, pp. 223–253. Lawrence Erlbaum Associates, Hillsdale (1987)
[9] Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic
Publishers, Norwell, MA
[10] Wu, S., Chang, A., Chang, M., Liu, T.-C., Heh, J.-S.: Identifying Personalized Con-
text-aware Knowledge Structure for Individual User in Ubiquitous Learning Environment.
In: Proceedings of the 5th IEEE International Conference on Wireless, Mobile, and
Ubiquitous Technologies in Education (WMUTE 2008), Beijing, China, March 23-26,
2008 (in press, 2008)
Learning Models for the Integration of Adaptive
Educational Games in Virtual Learning Environments
1 Introduction
Nowadays, adaptation to the needs of different learners and contexts is becoming an
increasingly important aspect of Virtual Learning Environments (VLE) [1, 2]. This is
a result of the need to reach users anywhere and anytime combined with the flexibility
of web technologies. Typical adaptation mechanisms build student profiles based on
learner preferences, portfolio, previous knowledge, educational objectives, and, in
some cases, even different learning styles [3, 4].
Another increasingly important aspect in the field of educational technologies is
the inclusion of digital games in educational environments. Works such as [5, 6]
identify the importance of motivation in the learning process and, from that base,
other authors discuss the benefits of using videogames to enhance the quality of the
learning experiences [7, 8]. Moreover, other authors defend that the characteristics
that make videogames attractive (immersion, short feedback cycles, scaffolded learn-
ing, perception of progress, etc.) are also key elements in any effective learning
process [9, 10].
Additionally, the characteristics of game-based learning suggest its potential bene-
fits when applied to adaptive online learning. Adaptation is a pervasive feature of
commercial videogames, since they are practically required to support different diffi-
culty levels in order to cater to the broadest possible audience. Moreover, unlike with
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 463–474, 2008.
© Springer-Verlag Berlin Heidelberg 2008
464 J. Torrente, P. Moreno-Ger, and B. Fernandez-Manjon
an HTML or PDF document, in a game we can monitor very closely the interaction
between the student and the game and use this information as part of our adaptation
cycles [11].
However, in spite of all the work created so far, there still is a need for effective
learning models that leverage the potential of games integrating them with the avail-
able learning materials and VLE. The use of educational videogames as content in
adaptive Virtual Learning Environments in a sensible way is not a trivial problem. To
begin with, we cannot force all the broad variety of students to learn from game-based
solutions. Some students may lack the proper equipment, some public computers may
restrict the types of content that can be accessed, and some students may simply re-
fuse to play the games seeing them as a waste of their time [12]. Even from an adapta-
tion perspective, some student profiles may require a very closely guided learning
experience as opposed to the exploratory freedom offered by most games. In such
cases, a conventional approach based on HTML, PDF and multimedia content could
be more appropriate. Finally, traditional web content is more accessible for people
with special accessibility needs, for example using text-to-speech tools for blind
people.
For these reasons, in the context of adaptive online learning, the employment of
educational games should not replace other approaches, but try to complement the
learning experience by integrating both alternatives in richer learning environments.
We propose a new model of integration, in which game-based content is blended and
complemented with traditional web contents as opposed to substituting them.
However, in order to apply such a model, it is necessary to address a number of
technical issues. In this work we describe how the <e-Adventure> educational game
platform [13]can be used to support this approach, facilitating the development of edu-
cational adventure games and their integration with Virtual Learning Environments.
This work is structured as follows: section 2 describes some of the issues regarding
adaptation and game-based learning that supply the base for this work. Then, on sec-
tion 3, we discuss the issues that emerge when substituting traditional content with
game-based content and propose a learning model that mixes both. On section 4, we
describe our <e-Adventure> platform, which can be used to implement that learning
model. Finally, some conclusions and future work are discussed in section 5.
different learning styles. For the scope of this work, we will only focus on the techni-
cal needs when it comes to adaptation in terms of a) different levels of initial knowl-
edge; and b) different learning styles.
Adaptation to different levels of initial knowledge requires finding out the current
level of each student in order to adapt the content accordingly. For instance, it could
be addressed by filtering basic content for those students with a certain level of
knowledge, so they could effectively focus their effort
Regarding the adaptation to different learning styles, we are aware that this is still a
controversial field. In spite of long empirical efforts to pin them down, the identifica-
tion of learning styles remains elusive [14]. However, most people with teaching
experience acknowledge intuitively that there are differences in how their students
learn [15]. For this work we will assume a very rough classification of student pro-
files, cataloguing them according to whether they are able to self-regulate their own
learning processes (preferring a free and exploratory approach), and those who prefer
close teacher control and guidance.
Fig. 1. A game pattern for adaptive game-based learning. The game can exhibit different be-
haviors to support different learning styles (Profile 1, 2…) and can omit certain levels that are
too basic for the level of initial knowledge. The game also gathers information from the interac-
tion during the in-game exam and uses it to modify its own behavior.
This pattern also contemplates using the game mechanics for assessment purposes,
monitoring the activity of the student during an in-game exam. Whilst the game is
played, a lot of interaction between the learner and the game is produced. From the
monitorization of all these interactions it is possible to infer data about the learners
that could be used to categorize them into one of the learning styles previously de-
fined. In some cases, if the results achieved by the student are insufficient, it is also
possible to reassess the profile of the student and run the game again with a different
profile.
As it has been stated in the previous section, educational videogames can be a vehicle
for the introduction of complex adaptation procedures in the learning experience,
enhancing motivation and providing an immersive domain to interact with. However,
game-based approaches would neither suit every student’s tastes nor be adequate in
all contexts and situations.
First, the technological complexity of videogames is an issue. Most games demand
high system requirements and some students may find their computers unable to exe-
cute them. Similarly, as videogames are complex pieces of software, its use is re-
stricted in most public and private systems beyond the personal field. Additionally,
there is an emerging trend towards the use of mobile devices (e.g. PDAs, mobile
phones, etc.), in what has been named m-Learning (mobile learning) [20] and mobile
learners usually will favour traditional web-based content.
Apart from the technological issues, sometimes the students themselves would de-
cide not to use these games. Games are usually time-consuming, as they require get-
ting familiarized with the environment (i.e. the domain of study) and sometimes even
learning how to use and interact with the universe devised in the game. Moreover,
some students are averse to videogames. A person with no experience playing com-
mercial videogames would find in educational videogames an extra and superfluous
challenge to waste their efforts on instead of an additional motivation, as witnessed
and reported in [12]. If addressing adaptation in terms of students’ preferences and
learning styles is desirable, we should avoid forcing those students who distrust
videogames to learn from game-based contents.
Another issue would be students that cannot interact with game-based contents. For
instance, blind people would find it impossible to learn through videogames, as visual
interaction cannot be replaced in games with a further use of the other senses. In these
cases, alternative learning itineraries, like HTML web pages that can be interpreted by
a text-to-speech tool, should be provided, following the current research aiming to
provide access for all to Information Systems [21].
Finally, game-based solutions are not recommended for those students that feel
overwhelmed by the freedom of exploration provided in games (even when such
freedom can be gauged and adapted transparently in the games). In those cases tradi-
tional content, in which the interaction during the learning experience is more rigid
seems to be more suitable.
On the one hand, the educational game is more costly to develop and it may be
complicated to ensure that guarantees the accomplishment of the same learning objec-
tives. This adds a new burden for the instructor or the person entrusted to produce and
maintain the learning content. On the other hand, the inclusion of the game in the
learning module would require reshaping the content and the pedagogical approach in
order to fit in a game-based delivery.
These issues could be addressed by supplying instructors (or content designers)
with mechanisms for the automatic integration of the existing content into the new
game-based content so that it can be easily accessed from the game, with no extra
effort for the instructors. In this way, both learning itineraries (traditional and game-
based) would incorporate the same information.
Taking into account the aforementioned aspects, we propose an adaptive model (Fig-
ure 2) with two adaptation layers. Firstly, the VLE decides whether a game-based or a
traditional HTML approach is more appropriate for the learner according to the pro-
file. This decision may depend on the requirements of the student (learning styles,
disabilities, etc.), the current context (a mobile device or a short consultation session)
or the student’s preferences. Secondly, when game-based content is chosen, a more
fine-grained adaptive mechanism is applied to adjust the game’s behaviour in the
terms discussed thus far.
These adaptive mechanisms are supported mainly by a test performed at the begin-
ning of the learning experience. This pre-test, checks whether the game-content is
suitable for the students and their initial level of knowledge. In some cases, the stu-
dent profile kept by the VLE will be sufficient to decide the shape of the content to be
delivered, making this test unnecessary. In other circumstances (e.g. when the student
is a novice using the system) a questionnaire could be given to the student in order to
find out their knowledge, preferences and perhaps some information about their
“learning taste”, so that their learning style can be inferred. The complexity of these
tests is in accordance with the adaptation mechanisms that we wish to apply. Most
models that attempt to capture learning styles include methodologies to infer the
learning styles of each student (like Vermunt’s Inventory of Learning Styles ques-
tionnaire [22]).
The adaptation within the game can fit different learning styles by displaying dif-
ferent game behaviours (such as biasing the behaviour towards guided or exploratory
styles) and also support different levels of initial knowledge by skipping those sec-
tions which appear too simple for a particular student.
After the learning experience has been completed, a second test (post-test) is per-
formed. The results of this test (that can be obtained within or outside the game) can
be used to refine the student’s profile in order to improve future adaptation decisions.
If a game-based content was the choice, an in-game exam could be the source of the
assessment, along with other data gathered through the monitorization of the interac-
tion of the student with the game. On the other hand, if conventional content was
chosen, a traditional online exam would provide the information.
Learning Models for the Integration of Adaptive Educational Games in VLE 469
Fig. 2. Adaptation model considering two different learning itineraries: HTML-based and
GAME-based. The adaptation is focused on two conditions: learning styles (different content
paths, different profiles in the game) and prior-knowledge (initial levels of the game can be
skipped).
Regarding the issues when integrating game-based content in the learning experi-
ences, we consider that both learning itineraries should not be disconnected. As
shown in figure 2, the itineraries are linked allowing the games to access the HTML
content as suggested in section 3.2.
The platform can be integrated with Virtual Learning Environments. When de-
ployed from a standards-compliant VLE, the implementation of the engine can query
the LMS for a set of properties that are used to adapt the game. The games are defined
so that the different values of those properties will change the initial state of the game
and consequently the game will be adapted, thus supporting the second adaptation
level of the general pattern described in section 3. The adaptive cycle is closed by in-
game tests, which are automatically assessed by the engine providing the necessary
feedback to readjust the student’s profile if necessary.
Another issue to address is how to effectively provide domain information to
learners while playing because this aspect is not usually present in commercial games.
Following the traits of the adventure game-genre [23], the interactive conversations
with the characters that the players encounter inside the game represent a good source
of information. However, interactive conversations with other characters are not al-
ways ideal. Some contents may require being delivered through alternative meta-
phors. When a large amount of data has to be delivered, conversations are not the
most “natural” channel. Long conversations will prompt the students to loose their
attention and focus. Moreover, embedding large amounts of information in conversa-
tions reduces its availability.
It is often desirable to make sure that some reference information is at the reach of
the student at every moment. A first solution would be to allow the students to consult
separate online materials containing the information, although this can be cumber-
some for the student who is forced to switch back and forth between the game and the
contents. For these reasons, <e-Adventure> includes the notion of in-game books.
Those books are available to the learner at any moment to be looked up, supporting
both text and images.
The books can be specified in <e-Adventure> by two different approaches: On the
one hand, the content of the books can be marked up along with the XML-storyboard,
as depicted in Figure 3. On the other hand, the definition of the contents of the books
can be detached from the game content by referring to a web page. In this manner, the
Fig. 3. A fragment of a marked-up in-game book in <e-Adventure>. The figure depicts how a
book is represented in the storyboard (left) and how it is visualized in the game (right), as an
actual book.
Learning Models for the Integration of Adaptive Educational Games in VLE 471
content of the book is retrieved from a URL where an XHTML document is located
and subsequently displayed in the game, as depicted in figure 4. A first advantage of
detaching the content of the books from the storyboard (where books are marked up)
is that the production and maintenance of book content is eased as the instructor can
edit and organize the information using HTML authoring tools. However, the main
advantage of this approach is how it supports the integration of existing web content
in <e-Adventure>. This can be leveraged to support the key issues from the adaptive
learning model described in section 3, allowing direct access to the web-based content
deployed on the VLE from inside the game.
Fig. 4. Reference information can be designed using web authoring tools as an XHTML docu-
ment (left). Then, <e-Adventure> renders this document into a better designed book (right).
The piece of XML (below) shows how the book is defined by just referring to its assets (back-
ground image and content web page).
5 Conclusions
In this work, we have argued the necessity of bringing together two increasing trends
in online education: adaptation (in order to suit the broader range of people, reaching
users anywhere and anytime) and the application of videogames to educational pur-
poses. Nonetheless, it has been remarked that traditional content should not be totally
replaced, as game-based solutions are not always the best choice for everyone at any-
time. Instead of that, both approaches should coexist in the Virtual Learning
Environments, combining the advantages of both approaches. In our opinion, the
achievement of the aforementioned goals will need the research to move towards the
development of learning models that integrate in a sensible way both key concepts
(adaptation and use of educational videogames).
472 J. Torrente, P. Moreno-Ger, and B. Fernandez-Manjon
Considering all these reasons, we have outlined an adaptive learning model sup-
porting a full adaptation cycle and the integration of game-based content with tradi-
tional alternatives. This model addresses adaptation in two layers: First, we introduce
the capacity to diversify the learning experience, supporting different itineraries (in-
cluding game-based and traditional content) in order to suit the broadest range of
learners and situations. The second adaptation stage takes place when game-content is
chosen, leveraging the characteristics of videogames to provide a much more fine-
grained adaptation mechanism. The first adaptation step is thus to decide what sort of
content is to be delivered according to the learner’s profile. Therefore, this layer not
only represents a first adaptation step, but also a general guideline for the integration
of educational videogames in existing VLEs: Traditional content should be kept and
offered as an alternative instead of replaced.
Another relevant concept discussed is the relevance of integrating the existing web
content inside the games. Some types of information do not translate easily into game
features, and simply work better in their textual form.. A first idea would be to allow
the students to consult the online materials while playing, although this can be cum-
bersome for the student who is forced to switch back and forth between the game and
the contents. Another approach would be to embed all the HTML content into the
game. This allows the student to use the reference materials from within the game.
However, duplicating the content is not a sustainable approach from a content adapta-
tion and maintenance perspective. The solution presented here is to link these contents
from the games, having the game engine render web content retrieved directly from
the VLE.
It is important to remark that the proposed model is a simplification devised to dis-
play the features that should be supported by the technology employed to perform the
integration of adaptive games in online education. For the sake of simplicity, the
model only displays adaptation driven in terms of learning styles and prior-
knowledge, although it is broad enough to support other adaptation approaches.
As described in section 4, this model can be implemented using the <e-Adventure>
educational game platform. The assessment and adaptation mechanisms offered by
the platform support the adaptive features described in the adaptive game model out-
lined in section 2.2. Additionally, since <e-Adventure> can be deployed in standards-
compliant Virtual Learning Environments, it can support the general adaptive cycle
described in section 3.3, facilitating the proposal of alternative game-based itineraries
without detaching the game experience from the rest of the learning process. Finally,
the capacity of the platform to render HTML-based Learning Objects inside the game
using a book metaphor solves the issues related to the integration of existing web
content inside the games in a sensible way.
As a final remark, it must be mentioned that the use of in-game books to provide
the learner with reference information, may be a double-edged weapon: an abuse of
in-game books is prompt to make the learning experience boring and give the narra-
tion a slow pace, instead of making it more motivating and attractive. When talking
about educational videogames we should always bear in mind that both fun and edu-
cational content should be present and balanced; otherwise we will have the “bene-
fits” of a bad learning experience (sometimes referred as “eduboring”) but at a higher
cost. We are trying to achieve such balances developing learning modules that
Learning Models for the Integration of Adaptive Educational Games in VLE 473
implement this model in the field of Computer Science teaching. The results obtained
will be useful to gauge a further detailed model and propose future lines of work.
References
[1] Burgos, D., Tattersall, C., Koper, E.J.R.: Representing adaptive and adaptable Units of
Learning. How to model personalized eLearning in IMS Learning Design. In: Fernández
Manjon, B., et al. (eds.) Computers and Education: E-learning - from Theory to Practice.
Springer, Heidelberg (2007)
[2] Sancho, P., Martínez-Ortiz, I., Fernández-Manjón, B., Moreno-Ger, P.: Development of a
Personalized e-Learning Experience Based on IMS Standard Technologies. In: Mendes,
A.J., Pereira, I., Costa, R. (eds.) Computers and Education: Towards Educational Change
and Innovation, pp. 73–82. Springer, London (2008)
[3] Brusilovsky, P.: Adaptive Educational Systems on the World-Wide-Web: A Review of
Available Technologies. In: WWW-Based Tutoring Workshop at 4th International Con-
ference on Intelligent Tutoring Systems (ITS 1998), San Antonio (1998)
[4] Hannafin, M.J., Land, S.M.: Student centered learning and interactive multimedia: status,
issued, and implication. Contemporary Eduaction 68(2), 94–99 (1997)
[5] Malone, T.W., Lepper, M.R.: Making learning fun: A taxonomy of intrinsic motivations
for learning. In: Malone, T.W., Lepper, M.R. (eds.) Aptitude, learning and instruction III:
Cognitive and affective process analysis, pp. 223–253. Lawrence Erlbaum, Hillsdale
(1987)
[6] Lepper, M.R., Cordova, D.I.: A desire to be taught: Instructional Consequences of Intrin-
sic Motivation. Motivation and Emotion 16, 187–208 (1992)
[7] Garris, R., Ahlers, R., Driskell, J.E.: Games, Motivation and Learning: A Research and
Practice Model. Simulation & Gaming 33(4), 441–467 (2002)
[8] Jenkins, H., Klopfer, E., Squire, K., Tan, P.: Entering the Education Arcade. ACM Com-
puters in Entertainment 1(1) (2003)
[9] Gee, J.P.: What video games have to teach us about learning and literacy, p. 225. Pal-
grave Macmillan, New York, Basingstoke (2003)
[10] Van Eck, R.: Building Artificially Intelligent Learning Games. In: Gibson, D., Aldrich,
C., Prensky, M. (eds.) Games and Simulations in Online Learning: Research and Devel-
opment Frameworks. Information Science Publishing, Hershey (2007)
[11] Martinez-Ortiz, I., Moreno-Ger, P., Sierra, J.L., Fernández-Manjón, B.: Production and
Deployment of Educational Videogames as Assessable Learning Objects. In: First Euro-
pean Conference on Technology Enhanced Learning (ECTEL 2006). LNCS. Springer,
Heidelberg (2006)
[12] Squire, K.: Changing the game: What happens when video games enter the classroom.
Innovate, Journal of Online Education, 1(6) (2005)
[13] Moreno-Ger, P., Martínez-Ortiz, I., Sierra, J.L., Fernández-Manjón, B.: A Content-
Centric Development Process Model. IEEE Computer 41(3), 24–30 (2008)
[14] Mayes, T., De Freitas, S.: Review of e-learning theories, frameworks and models (2004)
474 J. Torrente, P. Moreno-Ger, and B. Fernandez-Manjon
[15] Coffield, F., Moseley, D., Hall, E., Ecclestone, K.: Learning styles and pedagogy in post-
16 learning, Learning Skills Research Centre (2004)
[16] Aldrich, C.: Learning by Doing: A Comprehensive Guide to Simulations, Computer
Games, and Pedagogy in e-Learning and Other Educational Experiences. Pfeiffer, San
Francisco (2005)
[17] Kirriemur, J., McFarlane, A.: Literature review in games and learning., NESTA Future-
lab, 8 (2004)
[18] Mitchell, A., Savill-Smith, C.: The Use of Computer and Videogames for Learning: A
Review of the Literature. Learning and Skills Development Agency, Trowbridge, Wilt-
shire (2004)
[19] Hunicke, R., Chapman, V.: AI for Dynamic Difficulty Adjustment in Games. In: 19th
Nineteenth National Conference on Artificial Intelligence (AAAI 2004), AAAI Press,
San Jose, California (2004)
[20] Savill-Smith, C.: The use of palmtop computers for learning: a review of the literature.
British Journal of Educational Technology 36(3), 567–568 (2005)
[21] IMS Global Consortium. IMS Guidelines for Developing Accessible Learning Applica-
tions, Version 1.0 White Paper (2005) (cited March 2008),
http://www.imsglobal.org/accessibility/index.html
[22] Vermunt, J.: The regulation of constructive learning processes. British Journal of Educa-
tional Psychology 68(2), 149–171 (1998)
[23] Amory, A.: Building an Educational Adventure Game: Theory, Design and Lessons.
Journal of Interactive Learning Research 12(2/3), 249–263 (2001)
The Potential of Interactive Digital Storytelling for the
Creation of Educational Computer Games
1 Introduction
E-learning has been considered one of the most promising new markets during the last
years. Major activities covered the development of learning environments and stan-
dards for the exchange of learning material. From the point of view of learning con-
cepts, the resulting material adapted mostly ideas from behaviorism, resulting in a
more or less rigid structure of electronic learning environments, making it hard to
provide self-guidance to a learner. On the other hand, e-learning offers the chance to
deliberate about learning concepts in a new way.
Game-based learning (GBL) describes the application of computer games for
learning, or, as Prensky has put it, it “is precisely about fun and engagement, and the
coming together of serious learning and interactive entertainment into a newly emerg-
ing and highly exciting medium” [1]. Furthermore, Gee [2] has analyzed that com-
puter games are new media that let children and adults experience a learning effect
while enjoying themselves.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 475–486, 2008.
© Springer-Verlag Berlin Heidelberg 2008
476 S.A. Weiß and W. Müller
by elements such as challenge, fantasy, curiosity, and choice and control [6]. More-
over, achieving a flow experience [7] may also provide further extrinsic motivation
[8]. Furthermore, narrative in games may provide a cognitive framework for problem-
solving [9]. However, consolidated findings on the effectiveness of game-based learn-
ing are limited [10].
The fact that due to the high cost for developing good and complex computer
games, educational games do often not meet necessary quality standards in terms of
technology, complexity of the scenario, story, and didactic contents (see for instance
[11], [12]), represents a major problem in this context. As a consequence, commercial
of-the-shelf (COTS) computer games are frequently being applied to teaching and
learning. However, commercial games are in general not designed for applications in
learning, and content may be inaccurate or incomplete. Applying COTS games to
learning requires therefore careful analysis of the game’s strengths and weaknesses as
well as a matching of the game’s content to learning goals [13].
A number of commercial learning games exists targeted for specific learning sce-
narios. Yet, most of these follow behaviouristic approaches, (e.g., Mathematikus [14],
Physicus, Informaticus [15]). As such, they do not reflect today’s understanding on
how to design successful learning scenarios. There are also a number of simulation-
based approaches, e.g., Making History [16], or SimCity [17]. While simulation
systems represent a promising tool to further learning, these games typically do not
integrate easily into educational processes and also demand a careful application.
In the next chapter we will discuss the extent to which such commercial games
meet the demands of learning and which learning processes have to be considered.
Requirements that result from the combination of game play and learning
GBL has been around for quite a while now, and several products are available that
take on this approach. However, most of today’s commercial learning games can
hardly be considered convincing examples of successful game-based learning solu-
tions. In general, these applications fall into two classes: those stressing learning, and
those putting the focus on the game idea. While applications from the first class usu-
ally lack the targeted levels of fun and playfulness, applications from the second class
typically fail to integrate learning and learning elements successfully. In general, the
question on how to integrate learning successfully with elements of play and games is
unsolved.
Jantke stated several requirements for educational games that most of today’s GBL
application do not fulfill [11]. Among these the following two demands were not met
in most cases:
• Frustration caused by a conflict between game play and teaching material should
be avoided e.g. learning interactions may not interrupt the flow of the game play
and disturb the player’s immersion. These interactions of learning have to be sup-
portive instead and shall not hinder the player from reaching his goals.
• In general, the playing part and the learning part of the game should form a unit.
The most important criterion is that interactions of learning should appear as inher-
ent constituents of the game.
In the following we propose Interactive Digital Storytelling (IDS) to be utilized for
GBL. As well, we discuss how IDS techniques can be used to interweave elements of
game and learning and the requirements for such a system. We also discuss current
examples and how far learning can be integrated into IDS applications.
the basis for text-based interactions. The result is an emerging dialogue, influenced by
the users’ inputs and the bots’ databases of possible lines matching a text pattern
coming from either a user or another bot. The bots also take into account parametric
settings and scene descriptions provided by an author.
The experience of interacting with the platform of multiple chatbots shows that
there is high entertainment value through the fact that the course of the conversation
cannot be completely anticipated, not even by the writer of the dialogue patterns.
While there are still problems with non-sequitur verbal reactions to user input, people
mostly cope with it as within chats in their real life, and as a result, rather assume
strange character traits to the bots according to their appearance.
Fig. 1. Screenshot showing the main application window of Scenejo currently demonstrating
the Killer Phrase Game
a definition of a dramatic arc. This provides a link to the field of narrative and to
innovative approaches in the field of IDS.
Interaction
Interaction represents a further challenge in this field. Dialogue-based interaction
seems the most natural approach. A number of systems allow for dialogues between
users and virtual actors based on utterances (e.g., [22], [23]). Often, it is abstained
from applying speech recognition, reducing interaction to text input and output, some-
times enhanced by speech output and elements of non-verbal communication based
on animation. With the advancement of corresponding technologies, multimodal in-
teraction and tangible device are being applied more often (e.g., [39]). Besides the
general problem of speech and utterance recognition also the generation of utterances
for virtual actors appropriate for a narrative context and reflecting the characteristics
of a virtual character (knowledge, style, emotional state, etc.) remains a major chal-
lenge [40]. Both, enhanced input and output technologies are not only necessary to
provide a coherent immersive experience in the game, but also to support dialogues
on the necessary level of quality in the context of learning, where user input (e.g.,
questions, answers) have to be recognized in a reliable way.
5 Conclusions
The integration of Interactive Digital Storytelling with game-based learning has the
capability for creating effective educational computer games. We have to stress that
Interactive Digital Storytelling is about the development of a new genre. At present,
we do not know what good interactive stories are. Also, the experiences with IDS in
the field of Game-based learning are very limited. First experiments are however
promising. The Killer Phrase Game, a small and simple educational game on rhetoric
and communication could be implemented successfully based on the Scenejo frame-
work. So far, it has shown that there is a potential for designing successful games for
learning involving virtual actors based on digitally implemented agents. But there is a
need for further technical development and a lot more of convincing showcases have
to be built. However, these open tasks are preconditions for applying IDS success-
fully in the field of learning.
References
1. Prensky, M.: Digital Game-Based Learning. McGraw-Hill, New York (2001)
2. Gee, J.P.: What Video Games Have To Teach Us about Learning and Literacy. Palgrave
Macmillan, New York (2003)
3. Lepper, M.R., Chabay, R.W.: Intrinsic Motivation and instruction: conflicting Views on
the Role of Motivational Processes in Computer-Based Education. Educational Psycholo-
gist 20(4), 217–230 (1985)
4. Rieber, L.P.: Seriously considering play: Designing interactive learning environments
based on the blending of microworlds, simulations, and games. Educational Technology
Research & Development 44(2), 43–58 (1996)
5. Gee, J.P.: Situated language and learning: A critique of traditional schooling. Routledge,
London (2004)
6. Malone, T.W., Lepper, M.R.: Making learning fun: A taxonomy of intrinsic motivations
for learning. In: Snow, R.E., Farr, M.J. (eds.) Aptitude, Learning and Instruction III: Cona-
tive and Affective Process Analyses. Erlbaum, Hillsdale (1987)
7. Csikszentmihalyi, M.: Flow: The Psychology of Optimal Experience. Harper Perennial,
London (1990)
8. Bowman, R.F.: A “Pac-Man” theory of motivation: Tactical implications for classroom in-
struction. Educational Technology 22(9), 14–16 (1982)
The Potential of Interactive Digital Storytelling 485
9. Dickey, M.D.: Game design narrative for learning: Appropriating adventure game design
narrative devices and techniques for the design of interactive learning environments. Edu-
cational Technology Research and Development 54(3), 245–263 (2006)
10. Mishra, P., Foster, A.N.: The Claims of Games: A Comprehensive Review and Directions
for Future Research. In: Crawford, C., et al. (eds.) Proc. of Soc. for Information Technol-
ogy and Teacher Education International Conf., pp. 2227–2232 (2007)
11. Jantke, K.P.: Games that do not exist communication design beyond the current limits. In:
Proc. ACM Conference on Design of communication (2006)
12. Jenkins, H., Squire, K.: Harnessing the power of games in education Insight 1(3), 5–33
(2004)
13. Van Eck, R.: Digital Game-Based Learning: It’s Not Just the Digital Natives Who Are
Restless. EDUCAUSE Review 41(2), 16–30 (2006)
14. Westermann: Mathematikus (2008) (last visited: 28.02.08),
http://www.westermann.de
15. Klett, H.: Informaticus (2008) (last visited 28.02.08), http://www.braingame.de/
16. Muzzy Lane Software, Making History (2007) (last visited: 28.02.2008),
http://www.making-history.com
17. Wright, W.: SimCity (1989) (last visited 28.02.08), http://www.maxis.com/
18. Schank, R.C., Cleary, C.: Engines for Education. Lawrence Erlbaum Pub., Mahwah (1995)
19. Schank, R.C.: Tell Me A Story – Narrative and Intelligence. Northwestern Univ. Press,
Evanston, Illinois (1995)
20. Murray, J.H.: Hamlet on the Holodeck: The Future of Narrative in Cyberspace. The Free
Press (1997) ISBN 0-684-82723-9
21. Spierling, U., Iurgel, I.: Just Talking About Art - Creating Virtual Storytelling Experiences
in Mixed Reality. In: Balet, O., Subsol, G., Torguet, P. (eds.) ICVS 2003. LNCS,
vol. 2897, pp. 179–188. Springer, Heidelberg (2003)
22. Mateas, M., Stern, A.: Integrating Plot, Character and Natural Language Processing in the
Interactive Drama Façade. In: Proc. TIDSE 2003, Darmstadt, pp. 139–151 (2003)
23. Weiß, S.A., Müller, W., Spierling, U., Steimle, F.: Scenejo – An Interactive Storytelling
Platform. In: Proc. ICVS 2005, Strasbourg, France, pp. 77–80 (2005)
24. Aylett, R.S., Louchart, S., Dias, J., Paiva, A., Vala, M.: Fearnot! - an experiment in emer-
gent narrative. In: Panayiotopoulos, T., Gratch, J., Aylett, R.S., Ballin, D., Olivier, P., Rist,
T. (eds.) IVA 2005. LNCS (LNAI), vol. 3661, pp. 305–316. Springer, Heidelberg (2005)
25. Spierling, U.: “Killer Phrases”: Design steps for a digital game with conversational role
playing agents. In: Mayer, I., Mastik, H. (eds.) Proc. Isaga 2007, Eburon Pub. (2008)
26. Campbell, J.: The Hero with a Thousand Faces. Princeton University Press, Princeton
(1949)
27. Propp, V.: Morphology of the Folktale, 2nd edn. University of Texas Press (1968)
28. Spierling, U.: Interactive Digital Storytelling als eine Methode der Wissensvermittlung. In:
Eibl, et al. (eds.) Knowledge Media Design. Oldenbourg Verlag (2006)
29. Frasca, G.: Ludology Meets Narratology: similitudes and differences between (video)
games and narrative. Originally published in Finnish as Ludologia kohtaa narratologian in,
Parnasso, 3 (1999), English version online at http://www.ludology.org
30. Frasca, G.: Simulation 101: Simulation versus Representation (2001) (last visited:
28.02.08), http://www.ludology.org/articles/sim1/simulation101.html
31. Pearce, C.: Theory wars: An argument against arguments in the so-called ludol-
ogy/narratology debate. In: de Castell, S., Jenson, J. (eds.) Changing views: Worlds in
play, Digital Games Research Association Conference Proceedings, Vancouver, BC
(2005)
486 S.A. Weiß and W. Müller
32. Bizzocchi, J.: Games and Narrative: An Analytical Framework. In: Proc. CGSA 2006
(2006)
33. Crawford, C.: Chris Crawford on Interactive Storytelling; New Riders. Berkeley (2005)
34. Szilas, N.: Interactive Drama on Computer: Beyond Linear Narrative. In: AAAI Fall Nar-
rative Intelligence Symposium. AAAI 1999 Fall Symposium Series (1999)
35. Alexander, C.: A Pattern Language. Towns, Buildings, Construction. Oxford University
Press, New York (1977)
36. Barthes, R.: Introduction to the structural analysis of narratives. In: Sontag, S. (ed.) A
Barthes Reader. Hill & Wang (1966)
37. Spierling, U.: Adding Aspects of “Implicit Creation” to the Authoring Process in Interac-
tive Storytelling. In: Cavazza, M., Donikian, S. (eds.) Virtual Storytelling, Proc. ICVS
2007. LNCS. Springer, Heidelberg (2007) (Best Paper Award)
38. Naone, E.: Grandios gescheitert. Technology Review (January 2008) (last visited:
28.02.08), http://www.heise.de/tr/artikel/100944
39. Cavazza, M., Lugrin, J., Pizzi, D., Charles, F.: Madame Bovary on the Holodeck: Immer-
sive Interactive Storytelling. In: Proc. ACM MULTIMEDIA 2007, pp. 651–660 (2007)
40. Cavazza, M., Charles, F.: Dialogue Generation in Character-based Interactive Storytelling.
In: AIIDE 2005, pp. 21–26 (2005)
41. Charles, F., Mead, S.J., Cavazza, M.: Generating Dynamic Storylines through Characters.
Int. Journal of Intelligent Games & Simulation (IJIGS) 1 (March 2002)
Designing Virtual Players for Game Simulations in a
Pedagogical Environment: A Case Study
Jean-Marc Labat
1 Introduction
As de Jong proposed [Calie 04], learners are encouraged to construct their own
knowledge in realistic situations. This supposed an increasing of the learner’s activi-
ties. There are several ways to improve learner’s activities. On the one hand, the net-
works enhance communication among learners and also with the tutor whose role has
been acknowledged as highly important. On the other hand, it is important that,
whenever possible, the learner be engaged in problem-solving activity involving real-
istic situations. This indeed is the case when the tutoring system includes a problem
solver, an environment based on microworld, a virtual reality environment or a simu-
lator [Guéraud et al., 99].
In this paper, we focused on particular simulators which are business games. These
simulation activities are widely used in France at college level (particularly in engi-
neering and business schools) to train students in company management.
In section 1, we analyse briefly the different kinds of simulation and the level at
which students interact. Section 2 provides a general description of a particular busi-
ness game and of the architecture of the system. Section 3 presents the pedagogical
environment that the company, with which this project is developed, needs. Their
need was to introduce virtual players in order to increase the gameplay when there are
not enough human players. As we are in an commercial context, the aim of our indus-
trial partner is to provide a large set of business games on line in an ASP context.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 487–496, 2008.
© Springer-Verlag Berlin Heidelberg 2008
488 J.-M. Labat
Therefore, our goal is not to define virtual players for one particular game but rather
for a host of games. Section 4 presents a more or less generic methodology for de-
signing virtual players. We underline the epiphytic character of the pedagogical envi-
ronment developed, using the metaphor proposed by [Giroux et al., 96]. Lastly, we
present some results of experimentations and related works.
The game is composed of ten periods, each corresponding to a business year for
the company. In each new period, the playing team must make a general decision
about all the different parameters of the company before a deadline which is deter-
mined at the beginning of the simulation. Generally, the tutor allows fifteen days for
the decision process.
This decision is composed of production decisions (management of the production
equipment), commercial decisions (management of the commercial and marketing
parameters of the company), and financial decisions (budget and portfolio manage-
ment). During the game, it is possible for the teams to communicate with each other
and also with the tutor. In this way, the referee can advise the teams, and the teams
can exchange information relevant to subcontracting and tendering (take-over bids).
At the end of each period, the simulation engine run with all the decisions received
from the teams. Then the results are sent to all the teams together with the referee’s
comments. The results obtained give a broad outline of the company and its competi-
tors. These results belong to different categories: general results common to all the
companies (market share per product and company; average investment by company
in advertising, research, quality, organisation and maintenance; average price by
product for all companies…); accounting results which are specific to each company
(income statement; assessment; details of stock by product in volume and value);
production results specific to each company (management of the production equip-
ment: assignment of production units, detailed cost price; calculation of the produc-
tion cost of the products).
After the last period, a winner is designated on the basis of his score. Apart from
the first decision, there are three sources of information for each team: the historical
logfile of the past decisions, present results of the team and its competitors, and a
market study provided by the system. With this information, the player can define his
own strategy and plan a new decision.
Teams
Team support
Game website
Referee website
Database Simulation
engine
Data swaps
all the services of the game: participation in the contest, presentation of the results,
comments, e-mails and FAQ. The second site is accessible only to the referees.
Initially, the referee plays the role of the organiser: he defines the simulation ac-
cording to different scenarios, fixes the complementary parameters and manages the
simulation (data processing, virtual players …). Secondly, he plays the role of tutor:
he analyses the decisions and the results of the teams, gives them advice and also
answers the e-mailed questions. As [Crampes and Saussac, 99] wrote, "the choice of
the scenario and the technical and human organisation are fundamental". Concerning
the technology developed in HTML and ASP, the sites run on a SQL Server database
and a simulation engine developed using Visual Basic.
If the virtual player’s behaviour is too different from that of the human player’s, the
simulation will not be attractive enough; this will not only detract from the playability
of the game but will also, probably, divert the students from the subjacent objective
which is to acquire concepts of business management. It is thus essential that the
virtual players be accepted by the human players. Bearing this in mind, it is necessary
that they have the following characteristics:
1. They must play in a normal way, neither too well nor too poorly. This, however, is
not necessarily easy to arrange. If the games are too simple, for example if they are
algorithmic in nature, the virtual players can play perfectly. On the other hand, if
the games require real expertise, the virtual players may have aberrant behaviours,
in particular when they are designed along symbolic lines.
2. Their behaviour must be unpredictable so that, in identical situations, the human
players cannot anticipate their reactions. This is a priori easier to arrange (you just
need to introduce more randomness into the game). However, in most games (con-
sole games) the behaviour of the virtual players is completely stereotyped which
means that their human adversary can learn their reactions “by heart”. In addition
to making the game more random, it is possible to define several categories of vir-
tual players, as we propose in this paper (see below).
3. They must not "cheat". On the one hand, if the human players notice that the vir-
tual players are not playing fair, they will probably feel distaste for the game;
moreover, this cheating can hinder the training. Indeed, if the virtual players are ef-
ficient because they possess information which they are not supposed to have, the
learners may make false conjectures about the virtual player’s reasoning1. Conse-
quently, they must be no doubt about the information available to the virtual play-
ers. In this particular case, the virtual players have the same information as the
human players, but the expertise model they contain is the same as the model
which is implanted in the simulation. This is one of the fundamental ideas of this
1
This can be seen in computerized bridge games which use knowledge of the four deals to
play the cards. This leads to aberrant behavior (for instance, making a useless finesse because
the players know it will work).
Designing Virtual Players for Game Simulations in a Pedagogical Environment 491
project: collecting the expertise of a field is not difficult insofar as, by definition,
this expertise is in the simulation engine. To meet the different needs of a simula-
tion, the virtual players were designed as an expert system (SE). The expert-system
approach was the obvious choice as it can represent a true game rationality: i.e.,
adapt the decision to the economic situation in the game (company results, market
trends etc.), evolution of the decision-making (change of strategy or behaviour dur-
ing the game)… Thus the virtual player is not simply a calculus algorithm. He is a
module composed of a set of objects and rules, a module which sets the numerical
parameters of the virtual player’s decisions.
4. An epiphytic2 virtual player: In order to adapt to several simulations, the virtual
player must be able to integrate easily into the online game. The expert-system ap-
proach respects this criteria: the CLIPS module is integrated into the data-
processing environment of the decision through a DLL enabling the virtual player
to communicate with the database and also with the website used by the referee
(with this website, the referee can define the number of players, their behaviour
and some global economic variables about the game). The expert system operates
in the same way as the human player with his calculus sheet and his decision sheet.
It finds all the necessary information for decision-making (results of the previous
period and parameters of the current period), it processes it and sends the numeri-
cal data of the new decision to the database. Its running does not interfere with the
functioning of the game: the expert system must not modify the data processing of
the simulation. For human players and for the system, there is no difference at all
between the decisions taken by virtual or real players: The same procedure enters
the data into the database. The only difference comes from the fact that the proce-
dure receives data from a form for the human players and from the expert system
for virtual players. Thus, the virtual player can be described as an epiphytic system
[Giroux et al., 96].
It is well known that to explicit the expertise is the bottleneck for the development of
expert systems, especially when the goal is to model humans. But, in the case of
simulation, we are in a very favourable situation as the expertise of the domain is
represented within the simulation engine in a procedural manner. So, what we have to
do when defining virtual players is only to represent this expertise in a declarative
manner. Moreover, as human players are represented in the database by attributes, we
use the same attributes, with the same names, to describe virtual players in the expert
system, using the Clips templates. These two points are the key points when defining
the project. As we will see in section 4.3, we define a methodology in four stages.
4.3 Stages
Determine
Choosing
Sending Marketing
marketing
Marketing variables
parameters
Sending marketing
which are raised by the creation of the virtual player: In which environment does the
virtual player operate? What are the rules of the game? What are the strategies of the
game? How should he play? The aim is to build a general outline of a player’s deci-
sion-making based on two identification processes: identification of the various steps
of the decision (which makes it possible to define the way to be followed by the vir-
tual player to make his decision) and identification of the decision variables of each
step, which makes it possible to define the numerical parameters. Modelling was done
according to the UML method (see figure 2). The virtual player was constructed in
two modules in order to define a reliable game behaviour: the first module represents
the decision making of the player and the second the procedural aspect; i.e. how the
simulation engine computes the parameters.
2 Stage 2: Selecting in the database the necessary tables and the attributes to
represent the players and the expertise
The data generated by the decision-making of the human players are stored in a data-
base and are treated later by the simulation engine. The analysis of these data makes it
possible to determine the relevant tables and the right attributes representative of the
player’s decision and to transform them into an object structure within the expert
system. The results of the simulation engine after each period are also stored in the
database. The relevant attributes to model the expertise of the game also were trans-
formed into object structures within the expert system. This point is essential for the
reverse engineering methodology. From our experience, there exists a large similarity
between the Entity/Association model for the designing of database and the modelling
of objects in expert system. It is even simpler in expert system as there are, very often,
multivalued attributes.
which must belong Y(x): if Y(x)>Ymax then Y(x)=Ymax. When these parameters are
set, the cost price is automatically calculated by calling upon calculation functions
which require no decision.
5 Some Results
An interface was developed to accomplish preliminary tests. The performances of the
expert system were compared to the decisions made by the players whose strategies
were similar to the three behaviours defined above. The tests were carried out on three
types of periods, each of which has a specificity in the time-course of the game: the tests
of the period 1 type were chosen as they involve the initialisation of the simulation, the
tests of the period 2 type because the companies cannot yet produce all the products
(they are no surfboards), and the tests of the period 7 type because the companies which
have completed their research phase can produce surfboards. Initially, each player (cau-
tious, neutral and ambitious) was tested over these periods in the CLIPS environment.
After that, the test interface was used to test them in the simulation environment. The
results showed that the expert system took into account the various behaviours in a
random way: a cautious player never made the same decision in two consecutive epi-
sodes, and this was true for the two other types of players as well.
494 J.-M. Labat
The first contest with the virtual players was carried out with eight players: five
human players and three virtual players. We implement the three types of virtual
player: cautious, ambitious and neutral. Among the human players, there were the two
designers of the simulation game whose objectives were to test and evaluate the vir-
tual players (and the generated explanations). The last three players were students
who compete in normal conditions. Therefore, the two designers made voluntary
mistakes to see how the system reacts.
Another important point to notice is the fact that one contest is not necessarily rep-
resentative because there is a great variability in the game: first of all, the tutor
chooses the scenario (for example, he decides of the evolution of the market), sec-
ondly there is a random part in the choices made by the virtual players and in the
simulation itself, thirdly, the tutor chooses the criteria used to rank the players.
Keeping in mind these preliminary remarks, the results were very satisfactory. The
ranks of the virtual players were two, four and five. Moreover, the comments of the
designers of the game, who participated in the contest, were positive. They said that
virtual players had an average level: their decisions were good but they did not antici-
pate as humans would. It is exactly what we want, as the virtual players must be nei-
ther too good neither too bad. Since this first experimentation, current uses of the
simulation game confirm these first results: the virtual players are reasonably good
and the students cannot distinguish them from real players.
7 Conclusion
The development of virtual players and explanations based on reverse-engineering
technology using the expertise implanted in the simulation, seemed to be effective:
the virtual players are correct, even though they can be slightly improved. But, as the
company died in 2006, we have not tested the methodology with other simulation
games in order to measure the generic part of our methodology.
Finally, we are sure that simulation-based learning will be more and more developed
in the future as it allows to increase the learner‘s activities, using the dynamic capacities
of computers. We totally agree with [de Jong T., 04], constructivism is nowadays one of
the dominating learning theories, at least in Computer Based Learning context.
References
[Aimeur, E. & Frasson, C. 96] Analyzing a new learning strategy according to different knowl-
edge levels. Computers & Education 27(2), 115–127 (1996)
[Baylor A. & Kim Y. 05] Simulating Instructional Roles through Pedagogical Agents. Interna-
tional Journal of Artificial Intelligence in Education 15, 95–115 (2005)
496 J.-M. Labat
[Baylor, A. L. 00] Beyond butlers: Intelligent agents as mentors. Journal of Educational Com-
puting, Research 22(4), 373–382 (2000)
[Beaufils & Richoux, 03] Un schéma théorique pour situer les activités avec des logiciels de
simulation dans l’enseignement de la physique, Didaskalia. pp 9–38 (2003)
[Chan, T.-W. 91] Integration-Kid: A Learning Companion System. In: Mylopolous, J., Reiter,
R. (eds.) Proceedings of the Twelfth International Conference on Artificial Intelligence
(IJCAI-1991), vol. 2, pp. 1094–1099. Morgan Kaufmann Publishers Inc., San Francisco
(1991)
[Chan, T. W. & Baskin, A. B. 90] Learning companion systems. In: Frasson, C., Gauthier, G.
(eds.) Intelligent tutoring systems at the crossroads of artificial intelligence and education,
pp. 7–33. Ablex Publishing Corporation, NJ (1990)
[Chan, T. W. & Chou, C. Y. 97] Exploring the design of computer supports for reciprocal
tutoring systems. International Journal of Artificial Intelligence in Education 8, 1–29 (1997)
[Crampes M. & Saussac G, 99] Facteurs qualité et composantes de scenario pour la conception
de simulateurs pédagogiques à vocation comportementale. revue STE 6(1), 11–36 (1999)
[de Jong T., 04] Learning Complex Domains and Complex tasks, the Promise of Simulation
Based Trainig. In: CALIE 2004 conference, Grenoble, pp. 17–23 (2004)
[Dillenbourg, P. & Self, J. 92] People power: A human-computer collaborative learning system.
In: Frasson, C., McCalla, G.I., Gauthier, G. (eds.) ITS 1992. LNCS, vol. 608, pp. 651–660.
Springer, Heidelberg (1992)
[Dupuy J.-P, 99] Aux origines des sciences cognitives, édition La Découverte, Paris, p. 188
(1999)
[Giroux S., Paquette G., Pachet F., & Girard J., 96] EpiTalk - A Platform for Epiphyte Advisor
Systems Dedicated to Both Individual and Collaborative Learning Intelligent Tutoring Sys-
tems, 363–371 (1996)
[Gouarderes, G., Minko A., & Richard L., 99] Simulation et environnement multi-agents pour
l’apprentissage de la maintenance d’avions. revue STE 6(1), 143–187 (1999)
[Guéraud V., Pernin J-P., Cagnat J-M., & Cortes G. 99] Environnements d’apprentissage bases
sur la simulation- Outils auteur et experimentations. revue STE 6(1), 95–141 (1999)
[Johnson, W. L., Rickel, J. W., & Lester, J. C. 00] Animated pedagogical agents: Face-to-face
interaction in interactive learning environments. International Journal of Artificial Intelli-
gence in Education 11, 47–78 (2000)
[Koedinger, K.R. & Anderson, J.R. 97] Intelligent tutoring goes to school in the big city. Inter-
national Journal of Artificial Intelligence in Education 8, 30–43 (1997)
[Kort, B., Reilly, R., & Picard, R. W. 01] An affective model of interplay between emotions
and learning: Reengineering educational pedagogy-building a learning companion. In: Pro-
ceedings IEEE International Conference on Advanced Learning Technologies (2001)
[Uresty J. & du Boulay B., 04] Expertise, Motivation and Teaching in Learning Companion
Systems. International Journal of Artificial Intelligence in Education 14, 67–106 (2004)
The Relationship between Game Genres, Learning
Techniques and Learning Styles in Educational
Computer Games
Kowit Rapeepisarn1, Kok Wai Wong1, Chun Che Fung1, and Myint Swe Khine2
1
School of Information Technology, Murdoch University,
South St, Murdoch Western Australia 6150
{k.rapeepisarn,k.wong,l.fung}@murdoch.edu.au
2
Emirates College for Advanced Education, United Arab Emirates
mskhine@westnet.com.au
Abstract. Educational computer game has many similar characteristics like any
other genres of games. However its particular aspect is designed to teach, and in
which main objective involves in learning a topic. To develop an effective edu-
cational computer game, different game genres, learning activities and techniques,
and learning styles are important issues for consideration. This paper presents an
analysis by comparing and establishing relationships between the game genres
and learning techniques based on the types of learning and potential game style of
Prensky [1] and learning styles based on the study of Chong et al. [2].
1 Introduction
Educational computer games and various forms of edutainment have gained much
attention in the discipline of learning and teaching. Educators [3], [4], [5] believe that
most children learn best through play. Most studies also show that ‘learnt through play’
[6] has proven to be a successful learning experience. Therefore, it is desirable to use
educational computer games for teaching, which carry the objectives of play and learn
in the classrooms. There are reasons for using computer games as a learning tool to
enhance the learning experience of students. These reasons include the incorporation of
rules, goals, engagement, challenge, feedback, fun, interactive, outcome and immediate
reward [1], [7], [8], [9]. Even though most genres of the computer games in some ways
are educational, educational computer games are designed with explicit educational
purpose. When educational computer games are adopted in supporting learning in the
classroom, the pedagogical aspects such as learning style should be taken into account.
As different people learn and process information differently, it is important to under-
stand individual learning style which allows the prediction of the way learners react and
feel in different situations. Selecting the appropriate game genres for learning is an-
other important issue for consideration to develop effective educational computer game.
Recently, most studies focus on several variables when selecting game genres. This
includes age level, gender, racial diversity, number of players, and the role of teacher.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 497–508, 2008.
© Springer-Verlag Berlin Heidelberg 2008
498 K. Rapeepisarn et al.
Unfortunately, previous studies on the learning styles of the learners have not pro-
vided sufficient guidelines to design effective educational computer games to meet the
needs of individual learners. However, there are two important studies which are
relevant to the focus of this paper: Prensky [1] and Chong et al. [2]. Prensky’s study
presented a theory based on computer games and learning, whereas Chong’s study
focuses on the impact of learning styles using digital games. However, there are still
some gaps between their works. Therefore, this study aims to explore alternatives by
focusing on the learning techniques and the learning activities to match possible game
genres discussed in Prensky and Chong’s experimental research on computer game
types and the four learning styles. In addition to the literature review, this paper takes a
further step to develop a conceptual model based on those two studies in order to make
a contribution to the knowledge.
impact on this situations. Games deploy rich visual that draw players into fantasy
worlds and motivate the player via fun, via challenge and via instant feedback. With
instant feedback and immediate reward computer games provide, it makes a crucial
aspect for learning.
4 Learning Styles
Research into the use of games in education is growing rapidly [5], [7], [14], [15]. In
order to understand the potential roles of mainstream games in supporting learning, we
need first answer the questions “what is learning?” and “what forms of learning are
suitable for incorporating games in the classroom?” This is related to pedagogical
theory which includes learning theory which describing how people learn or what
styles of learning people like. Learning style is useful in identifying the methods by
which people prefer to receive information from their environment and undertake their
500 K. Rapeepisarn et al.
learning. Each person has his or her own way of converting, processing, storing, and
retrieving information. Some people prefer to learn through reading and reflecting on
how this might apply to their own experience, whilst others prefer to learn through
trying ideas out and learn through reviewing their experience before planning the next
step. Among the learning styles which classified as experiential, Honey and Mumford
learning style is one of the well known experiential learning [16]. This learning style
proves that people learn better when the teaching is adapted to the learning styles [2].
Honey and Mumford classify learners into activist, reflector, theorist, and pragmatist as
illustrated in Table 1.
Characteristics
Activists Reflectors Theorists Pragmatists
• Immerse in • Stand back and ob- • Think in a logical • Keen to put
new experience • serve. • Cautious, take a manner, rationally ideas, theories and
Enjoy here and back seat • Collect and and objectively. • techniques into
now • Open analyse data about Assimilate facts into practice. • Search
minded, enthu- experience and events, coherent theories. • new ideas and
siastic, flexible • slow to react conclusion Fit things into rational experimental • Act
Seek to centre • Use information from order. • Keen in basic quickly and confi-
activity around experience to maintain assumptions, princi- dently on ideas, get
themselves a big picture perspec- ples, theories, models straight to the
tive and thinking system. point. • Are impa-
tient with endless
discussion
new conceptual model by comparing and matching learning styles, learning activities
and game genres based on those two studies [1], [2] in section 5.3.
Fig. 1. Model of relationship of learning content, learning activities and possible game styles
5.2 The Study of Chong et al.: Learning Styles and Educational Game
Many educational researchers studied on learning styles, but the study on the rela-
tionship of learning styles to learning within game text is scarcely found. Researchers
are emphasizing that the education computer games should be developed considering
the learning styles of students [2]. However, there is a research conducted by Chong et
al. on the impact of learning styles on the effectiveness of digital games in education.
They conducted the survey based on the Honey and Mumford four types of learning
styles on 50 undergraduate students in INTI College Malaysia. They choose three
502 K. Rapeepisarn et al.
Table 2. Finding Summary of Prensky’s Learning Content, Leaning Activities and possible Game
Styles [1]
5.3 Bridging the Gap between the Prensky and Chong’s Studies
When reviewing the studies of Prensky and Chong at al., we realized that more need to
be done in order to provide a better framework for designing good educational games.
Prensky focuses on learning techniques, learning contents, and learning activities but
lack of learning style. Whereas, Chong et al. focus on learning style but uses only three
The Relationship between Game Genres, Learning Techniques and Learning Styles 503
different game genres as an example. Therefore, it is the attempt of this study to bridge
this gap by establishing the linkage of these two studies. Two conceptual models
proposed in this study are:
Firstly, as mentioned in section 5.1, Prensky proposes the relationship of learning
content, learning activities and the game styles. He also suggested 14 essential learning
techniques which he claimed that it should be considered and used when designing
learning materials. However, these 14 learning techniques have not been matched to
learning activities and game genres in his study. In order to use all those learning
techniques in learning with educational computer game, the relationship of each
learning techniques and game genres should be studied. Hence, it is the objective of
this paper to compare and match his 14 learning techniques to learning activities and
game genres. The new model and the result of this matching are illustrated in Figure 2
and Table 4.
Second, in section 5.2 Chong et al. [2] study the impact of learning styles on
computer game in education. They use three types of games in their study as the ex-
amples to prove that different learning styles do prefer different types of games. From
the finding of their study, the behaviors of each style of learners while playing game are
also described. However, they do not match this behavior of each style of learners with
learning technique. Additionally, only three different game styles are studies. While
the study of Prensky choose all the standard categories of computer games matching
with learning activities, but lacks of the comparison of learning style of the users. Thus,
this paper proposes the new conceptual model of the relationship between learning
styles, learning activities, and possible game genres based on these two studies [1], [2]
as illustrated in Figure 3.
The process that led to the new model is conducted by: 1) exploring the behavior
when playing games for each type of learner based on Chong’s study, 2) matching the
behavior of each type of learner to learning activities based on Prensky’s study, and 3)
finding the possible game genres which can relate to each learning activities. As an ex-
ample, the results from this study found that the possible game genres for the activist
learners could be multiplayer interaction, action games and role-playing game. Accord-
ingly, the results of matching learning styles; behavior when playing game; behavior
when using computer; learning activities; and possible game genres are shown in Table5.
504 K. Rapeepisarn et al.
Fig. 2. Model of relationship of learning techniques, learning activities and possible game styles
Table 4. The relationship between learning techniques, learning activities and possible game
styles
Learning by doing Interact, practice, drill, Strategy game, action game, role
imitation playing game
Learning from Feedback, problem Role-play game, puzzle game
mistake
Discovery learning Feedback, problem, Adventure game, puzzle game
& guided discovery creativity play
Task-based learning Understand principle, Simulation game, puzzle game
graduated tasks
Question-led Question/ questioning, Quiz or trivia game, game show
learning problem competition, construction game
Situated learning Immersion Immersive style game such as
role-playing game, flashcard game
Role playing Imitation, practice, Role-playing game, strategy game ,
coaching reflex game, adventure game
Constructivist Experimentation, Building game, constructing game
learning questioning
Multisensory Imitation, continuous Game in which introduce new
learning practice, immersion technologies such as locatable sound or
force feedback, reflex game
Learning object Logic, questioning Games which are becoming
object-oriented
Coaching Coaching, feedback, Strategy game, adventure game, reality
questioning testing game
Intelligent tutors Feedback, problem, Strategy game, adventure game, puzzle
continuous practice game, reflex game
The Relationship between Game Genres, Learning Techniques and Learning Styles 505
Fig. 3. Model of relationship of learning styles, learning activities and possible game genres
6 Discussion
As researchers have found that computer games have significant educational value,
computer games can become part of the school curriculum. There are different types of
computer games and games technologies, which have been used positively, both di-
rectly and indirectly to support and assist teaching and learning in the classroom. Green
and McNeese [17] conclude in their paper “Using Edutainment Software to Enhance
Online Learning” that the attributes of high quality educational computer games
comprise of 1) clear learning goal and objective, 2) provide review on concepts newly
learned and allow for questions and answers, 3) develop higher thinking skills, 4)
challenging but focus on learning rather than on winning or losing, 5) clear rules so
learners know how to play, 6) providing a means for collaboration, feedback, or
guidance, 7) be fun so learners are more relax, more alert, less fearful and open to
learning, 8) provide a means for debriefing to recap what was learned and allow for
question and answer. However, there are a number of issues, which need to be ad-
dressed before using computer games in the classroom.
Most studies concern about age of student, gender, racial diversity, and role of the
teacher. Unfortunately, not many researches focus on learning style when designing the
appropriate game genres for each style of learners. Different people have different
style of learning. No single learning preference is better than any other. In fact, indi-
vidual student may have more than one single learning style. This reflection can be
shown that when some learners prefer kinesthetic instruction, they can also have ability
to learn orally and visual [18]. One learner has the active type of learning; he/she may
have theory or pragmatic style of learning in other learning situation. Therefore, there
are many possible ways in choosing appropriate game genres for one particular student.
However, it can be determined by looking at learner’s dominate learning style. In other
word, the most preferred learning style of that learner. The model of relationship of
learning style, learning activities and possible game genre presented in this paper is
only the potential example proposal. To understand educational gaming, many factors
have to be examined. These include design, pedagogy, and literacy. It should also
focus on the classroom use, what is learned and what can be taught with educational
computer game [13]. Moreover, some variables such as the experience in playing game,
culture, language, and nurture should also have been examined. Game developers and
educational psychologists should work together with other professionals as a team to
formulate the educational content in order to build quality educational computer games.
506 K. Rapeepisarn et al.
Table 5. The relationship of learning style, behaviour when playing game, behaviour when using
computer, learning activity and possible game genres [1], [2], [16]
The further questions are: How do educators convince parents, teachers, and admini-
stration about the importance of gaming in education? [13]; Are there any significant
gains when determine the appropriate genre of game with learning styles? And are
there any relationship between teaching styles and using of educational computer
games?
The Relationship between Game Genres, Learning Techniques and Learning Styles 507
7 Conclusions
Educational computer games bring together the idea of game, play, fun, hands-on
experience and also with explicit educational purpose. Like other genres of computer
games, educational computer games have elements that benefit learning. This includes
rules, goals, active engagement, content/story, feedback, interactive, problem solving,
quick adaptation and immediate reward. However, educational content in education
computer games should be considered as the heart of game, and feedback that support
learners should also be built into difficult content. When designing educational com-
puter games for supporting learners in classroom, apart from educational content, they
can also embrace the pedagogical benefits such as learning style with game genres for
developing quality learning experience in class. This paper shows the comparison and
matching of learning techniques, learning activities, learning styles to possible game
genres. However, this study merely proposes the potential model used as one possible
conceptual guideline for further study in order to create effective educational computer
games. Further study may find out the benefit gain when determine the appropriate
genres of game with learning styles.
References
[1] Prensky, M.: Computer Games and Learning: Digital Game-Based Learning. In: Raessens,
J., Goldstein, J. (eds.) Handbook of Computer Game Studies, pp. 97–122. The MIT Press,
Cambridge (2005)
[2] Chong, Y., Wong, M., Thomson Fredrik, E.: The Impact of Learning Styles on the Effec-
tiveness of Digital Games in Education. In: Proceedings of the Symposium on Information
Technology in Education, KDU College, Patailing Java, Malaysia (2005)
[3] Morgan, A., Kennewell, S.: The Impact of Prior Technological Experiences on Children’s
Ability to Use Play as a Medium for Developing Capability with New ICT Tools. ICT
Research Bursaries, Becta (2005)
[4] Gee, J.: What Video Games Have to Teach Us About Learning and Literacy. Longman,
New York (2003)
[5] Kirriemuir, J., McFarlane, A.: Literature Review in Games and Learning
http://www.futurelab.org.uk/download/pdf8/research/
litreviews/Games_Review1.pdf (accessed in November 2005)
[6] Rapeepisarn, K., Wong, K., Fung, C., Depickere, A.: Similarities and Differences Between
“Learn Through Play” and “Edutainment”. In: Proceeding of the 3rd Australian Confer-
ence on Interactive Entertainment. Perth, Australia., December 4-6, pp. 28–32 (2006)
[7] Blunt, R.: A Causal-Comparative Exploration of the Relationship Between Game-Based
Learning and Academic Achievement: Teaching Management with Video Games, Ph.D.
dissertation, Applied Management and Decision Sciences; Walden University (2006)
[8] Garcia, G.: Digital Game Learning. In: Hoffman, B. (ed.) Encyclopedia of Educational
Technology (2005),
http://coe.sdsu.edu/eet/articles/digitalgamlearn/start.htm (accessed in November 2006)
[9] Juul, J.: The Game, the Player, the World: Looking for a Heart of Gameness. Level Up. In:
Digital Game Research Conference Proceedings (2003)
[10] Ferdig, R.: Preface: Learning and teaching with electronic games. Journal of Educational
Multimedia and Hypermedia 16(3), 217–223 (2007)
508 K. Rapeepisarn et al.
[11] Chuang, T., et al.: The Effect of Computer-Based Video Games on Young Children: A
Literature Review. In: Crawford, C., et al. (eds.) Proceedings of Society for Information
Technology and Teacher Educational International Conference, pp. 3954–3957. AACE,
Chesapeake,VA (2005)
[12] Dreyfous, R.: Cognitive and Affective Variables Involved in Recreational Computer-
Generated Games. Dissertation Abstracts International, 55(07), 1807A (UMI No.
9432617) (1994)
[13] Fisch, S.: Making Educational Computer Games. Educational, 56–61
[14] Ju, E., Wagner, C.: Personal Computer Adventure Games: Their Structure, Principles and
Applicability for Training. Data Base for Advances in Information Systems 28, 78–92
(1997)
[15] Amory, A., et al.: The Use of Computer Games as an Educational Tool: Identified of
Appropriate game types and game elements. British Journal of Educational Technol-
ogy 30, 311–322 (1999)
[16] Honey, P., Mumford, A.: The Manual of Learning Styles. Peter Honey, Maidenhead
(1992)
[17] Green, M., McNeese, M.: Using Edutainment Software to Enhance Online Learning. In-
ternational Journal on ELearning 6, 5–16 (2007)
[18] Dunn, R.: Individualizing Instruction for the Mainstream Gifted Children. In: Milgram, R.
(ed.) Teaching Gifted & Talented Learners in Regular Classrooms, Charles C. Thomas,
Springfield, Ill (1989)
EFM: A Model for Educational Game Design
1 Introduction
With the reform and development of education, contemporary educators increasingly
concerned about the overall development of learners, who will be promoted as a
separate entity for treatment. Under the guidance of such an educational idea, a vast
number of educators and parents commonly concern how to inspire learners' motiva-
tion and help them truly learn from playing. In the modern society, with the popularity
of network games and electronic games, the youth become infatuated with the com-
puter games, which promotes some scholars more actively research how to turn the
games effect into education. However, when many educational experts are concerned
about the educational value of games, the games was just viewed as teaching media. In
fact, the game can not only be regarded as a kind of teaching media, but also as a
learning environment to study, because the game itself contains the basic elements
which are necessary in learning activities. Through referring to relevant information at
home and abroad, it is clearly found the close connection of motivation, flow, effective
learning environment and educational game. Educational game can serve as an effec-
tive learning environment, so that learners produce motivation during the process of
flow experience and change from passive learning to active learning, so as to enhance
the quality of learning.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 509–517, 2008.
© Springer-Verlag Berlin Heidelberg 2008
510 M. Song and S. Zhang
2 Related Theory
2.1 Motivation
According to the Great Chinese Dictionary (CiHai), motivation refers to the driving
force in promoting people to learn. Whether the students study positively, what do they
glad to study, and how do they study, all have direct relations with the motivation. As
some scholars said, motivation can not only result in learning activities, but also en-
hance the efficiency and improve the effect of learning. The first step of instructional
design is to consider how to inspire motivation and maintain it. To help understand
motivation in instruction, we can look at the ARCS Model of Motivational Design as
developed by John M. Keller of Florida State University. The ARCS Model identifies
four essential strategy components for motivating instruction [1]:
[A]ttention strategies for arousing and sustaining curiosity and interest.
-Learners are more motivated when the instructional design generates curiosity
and interest about the content or learning context.
[R]elevance strategies that link to learners' needs, interests, and motives.
-Learners are more motivated when goals are clearly defined and align with
learners' interests.
[C]onfidence strategies that help students develop a positive expectation for suc-
cessful achievement.
-Learners are more motivated when challenge is balanced in such a way that the
learning process is neither too easy as to bore the leaner, or too difficult such
that success seems impossible.
[S]atisfaction strategies that provide extrinsic and intrinsic reinforcement for effort.
-Learners are more motivated when there are rewards for correctly executed
actions.
2.2 Flow
Flow Theory developed by Csikszentrnihalyi of Chicago University, as a method for
understanding and implementing motivation, is a theoretical bridge between the con-
cerns of instructional design and motivational design theory [2]. Flow theory has been
widely applied in the discussion about behavior and psychology in the man-machine
interaction environment, and it has been confirmed that the flow experience does exist
in the use of network.
Csikszentrnihalyi (1991) defined the phenomena of flow state as having nine di-
mensions [3]:
z Goals of an activity.
z Unambiguous feedback.
z Challenge-skill balance.
z Action-awareness merging.
z Concentration.
z Control.
z Loss of self-consciousness.
z The transformation of time.
z Autotelic experience.
EFM: A Model for Educational Game Design 511
Based on the process of flow experience, Novak and others (2000) classified these nine
dimensions into three categories [4]:
z Conditional factors, which could stimulate flow experience, include goals of
an activity, unambiguous feedback and challenge-skill balance.
z Experience factors, the feeling of individuals in a state of flow experience,
include action-awareness merging, concentration and control.
z Result factors, the results of experience of individuals in the flow state, in-
clude loss of self-consciousness, the transformation of time and autotelic
experience.
Education and game are originally an indivisible whole. Especially for children, game
is learning. Children's study start from observation, imitation and inquisition, and the
best activity which can embody Children's spirit of learning is game. When the children
play role playing games, play with sand or imitate adult activities they are interested,
their active attitude, cooperation spirit, explore awareness, imagination, and even a
certain degree of creative intelligence all will get trained and developed. "Learning
from playing" has been stressing since the era of Confucius, and the principles of all
learning activities modern educators advocated can be reflected in the game. The game
is one of the important ways for children and adult to study.
"Educational game" is still a newly emerging thing in our country, and there is no
explicit definition nowadays. Narrowly speaking, educational game refers to the inte-
gration of education and game, and the education effect naturally generated from the
process of playing games, in other words, it means "a type of computer game software
which generates education effect through interest" [8]. The educational game in narrow
512 M. Song and S. Zhang
Educational Game
7 Basic Requirements of
an Effective Learning
Environment
A Sense of Direct Appropriate Motivation Avoiding
Engagement Tools Distractions
Specific Goals High Intensity of A Continual Feeling
and Established Interaction and of Challenge
Procedures Feedback
9 Dimensions of
Flow Experience
Conditional
Factors Goals of an Unambiguous Challenge-Skill
Activity Feedback Balance
4 Essential
Strategy Attention Relevance Satisfaction Confidence
Components Strategies Strategies Strategies Strategies
for (Interest) (Goal) (Feedback) (Challenge)
Stimulating
Motivation Motivation
Fig. 1. The connection of motivation, flow, effective learning environment and educational game
Educational game has educational aims or instructional objectives, which separate from
or integrate with the original goals of game [8]. New curriculum standard emphasizes
the 3D objective, as perfect combination of knowledge and skill, process and method,
attitude and values, which is one of the important references to set up the goals of
educational game. So it is very essential to emphasize the process of copious and
natural experience of game, and the cultivation of correct passion, attitude and values in
addition to the knowledge and skill. The goals of educational game should be set up in
the principle of promoting full development of students.
To ascertain the type and characteristics of game according to the corresponding cur-
riculum and content first, and then confirm the frame of game scenes based on the
detailed teaching unit. We can divide the content into several units on the basis of the
goal of educational game, and then divide the game into several scenes on the basis of
the content units. That is to say, when we finish the frame of game scenes and the
relevant rules, we fulfill the whole game procedure and provide the learning procedure
for the learner accordingly. In the game environment, the learner can do everything
according to the scheduled procedure, and what he needs to do is to immersing himself
into the game body and soul.
Suitable props should be provided to help the learner to fulfill the tasks so that he will
not give up when he meet difficulties. Prop is a kind of important motivator in the
game, which can awake the curiosity of the learner so as to enhance the recreation of
the game. When the task comes with tools, the attention will be moved from learning
goal to play, as a result, the learners can reach the state of flow experience.
EFM: A Model for Educational Game Design 515
All the available resource will be used to deal with the relevant information (main
tasks) but not the game control because of the limitation of information treatment by
human beings. Therefore, in an ideal situation, game control will be transparent, and the
learner can be absorbed in the game.
It will destroy the first reception and interest to a game if there are long-winded in-
structions for the game operation and procedure. And if the learner needs to study how
to play the game because of the complicated operation, it will occupy the attention and
other cognitive resource of the learner, which will slow down the process of the game
and even block the achievement of flow experience.
Consequently, the game interface should be concise and will not confuse the learner,
the relationship and the layers of guide system should be clear and will not let the
learners get lost, the definition of common function key should be similar or same to the
common games, so that the learner can learn it easily. In conclusion, the transparency
of game control can guarantee that the learner will pay maximum attention to the game
and can reach the state of flow experience without any interference.
The backgrounds of stories could be powerful, unconstrained and imaginative, but their
plots should be logical and common. More real the stories are, more immersed the
players are, and more sensations of direct engagement are accepted. For example, the
United Nations World Food Programme (WFP) developed an educational game named
Food Force. Through the role playing of WFP staff, the players try to transport food to
a virtual island damaged by wars, seek refugees, airdrop succor materials, fight with
enemies and plan rebuilding programs for the farm. This kind of story background is
novel and attractive, and the plot is practical. The players will assume themselves as
heroes and take part in the tasks actively.
Different from our living environment, in the situation of game, it allows us to act
according to our free aspirations despite of considering the results. Being like the
learning environment, which should not restrain the capabilities of learners to construct
their own knowledge structures, the game situation should not restrain the cognitive
processes of players. It should allow players to try freely with their own aspirations, and
choose when and where to start or stop game. Therefore, the plot in a story background
should be designed as multi-choice, which will let the players try with curiosity, take
part in their own choosing plot and enhance the sense of direct engagement as well as
the enduring interest of the game.
Interaction and feedback connecting with the objective dimensions are important
elements to encourage the study activities. Communicating with the environment of
games, the players will know whether their action is positive, within the rules and close
516 M. Song and S. Zhang
Challenge is the core element of the games, and the learner will strive to research the
strategy to increase his ability to overcome the system when the challenge balances the
skill. Therefore, the designer of educational game should set up different grades for
different learners with different skills, which are selectable by the learner. At the same
time, the hardness of game will be adjusted automatically according to the ability and
performance of the learner. That is to say, there will be clear and reachable challenges
step by step in accordance with the improvement of the learner's skills. And the hard-
ness of challenge will be increased accordingly. The learner will continue the game in
the sense of self-affirmation and self-fulfillment whenever he overcomes the challenge.
That is the inner motivation of game.
The game should reflect the impact of grades and empirical values on the learner,
which means different grade and empirical value enjoy different power, which can
force the learners to continue the game in the purpose of higher power. The designer
can set up different grade according to different teaching objective to provide learning
objective step by step to the learners; and the designer also can set up different em-
pirical value according to sub-goal in the same grade. Take "Virtual Life" as instance,
the participants can accumulate intellect, fascination and energy through different
tasks, and when the empirical value reaches a certain extent, the participants can up-
grade to a higher grade and to reach the complete learning objective finally.
5 Conclusion
The research and development of educational game in our country is still in a primary
stage, effective models and ideas for educational game design are of great lack. This
paper proposed the EFM model, in order to provide a new thought for researchers and
developers of educational game. In the follow-up study, we will further study how to
build the game learning environment, and take the model in educational game design
practice to see how the model can help to improve it.
EFM: A Model for Educational Game Design 517
References
1. Keller, J.M.: Motivational Design of Instruction. In: Reigeluth, C.M. (ed.) Instructional De-
sign Theories and Models: An Overview of Their Current Status. Erlbalum, Hillsdale (1983)
2. Chan, T.S., Ahern, T.C.: Targeting Motivation–Adapting Flow Theory to Instructional De-
sign. Journal of Educational Computing Research 21(2), 151–163 (1999)
3. Kiili, K.: Evaluations of an Experiential Gaming Model. An Interdisciplinary Journal on
Humans in ICT Environments 2(2), 187–201 (2006)
4. Wan, L., Zhao, M., Zhao, C.: Viewing the Design of Digital Educational Game from Expe-
riential Games-Learning Model (in Chinese). China Audiovisual Education 10, 5–8 (2006)
5. Zhong, Z.: Discussing the Design of Learning Environment (in Chinese). E-education Re-
search 7, 35–41 (2005)
6. Zhu, X.: Discussing the Design of Learning Environment (in Chinese). China Audiovisual
Education 7, 16–18 (1996)
7. Houser, R., Deloach, S.: Learning from Games: Seven Principles of Effective Design.
Technical Communication 45(3), 319–329 (1998)
8. Zhao, H., Zhu, Z.: The Analysis of Definitions and Typology about Educational Game (in
Chinese). In: 10th GCCCE2006 Collected Papers, pp. 39–46. TsingHua University Pub-
lishing House, Beijing (2006)
Towards Generalised Accessibility of Computer Games
1 Introduction
Computer games have become an important part in child and youth culture, and most
children, in developed countries, have considerable experience of such games. Addi-
tionally these games are used by a growing part of the population, including especially
young adults (on average 25 years old, including 40% of women1) but the proportion of
players is also growing in other age groups of the population.
A lot of people with impairment are excluded from the computer games world be-
cause of accessibility. Indeed games accessibility have initially been regarded as an
area of minor importance as there were much more “serious” topics to focus on. Since
the middle of the nineties, lots of works have been focusing on making office computer
applications accessible, and it’s a fact that nowadays word processor and spreadsheets
applications are reasonably accessible as well as web browser and mail readers.
Today, as Zyda claims, “the time has come to take computer games seriously, really
seriously” [1]. Indeed the mainstream commercial market for computer games and
other multimedia products has shown an impressive growth in the last five years. Costs
for the development of a game may reach the level of major movie productions, in-
volving more than a hundred employees [2]. The expectation by games players of ever
1
TNS Sofres, Le marché français des jeux vidéo (The market of video games in France). afjv,
November 2006. http://www.afjv.com/press0611/061122 marche jeux video france.htm
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 518–527, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Towards Generalised Accessibility of Computer Games 519
more impressive games has seen increasing development budgets, and with a more
focused use of new technologies.
Academia and also R&D over the last years have started to focus on “serious games”.
Leading experts speak of “creating a science of games” [1] with the goal of implementing
games and game like interfaces of general importance for a growing number of applica-
tions and as a general trend in the design of Human-Computer Interfaces (HCI) [3].
In addition, general HCI is beginning to use concepts and methods derived from
games as they promise an increased level of usability. Games and game-like interfaces
2
are recognised as a means to implement educational, training, general HCI and web
applications with usability and effectiveness. Particular examples of interest are:
– eLearning and edutainment which more and more implement or use didactic
games [4]. As an example of this trend it should be noted that critical issues like
mathematics and science education are approached with game-based learning
infrastructures and edutainment to address the well known didactic problems in
this domain. “Games stimulate chemical changes in the brain that promote
learning.” [5].
– Avatar based interfaces. We assist to a growing number of applications in such
environments: for instance in France, real Job interviews have been organised in
Second Life2 .
– Emerging Non Classical Interfaces (e.g. virtual/augmented reality, embedded
systems, pervasive computing).
– A lot of Cultural Multimedia products, like Museum CD-Roms or DVD-Roms.
– Web 2.0
– Other software considered as inaccessible until today might come under accessi-
bility discussions based on the principles, guidelines and tools developed
for games and games like interfaces (e.g. simulation software, charts, virtual/
augmented reality).
Then, even if it would be considered as questionable to use the limited resources that
are available for research on accessibility to address problems of people with disabili-
ties using games or edutainment software, the general evolution of HCI towards
game-like interfaces compel us to also consider a “serious” look at games from the
accessibility perspective, in order to keep pace with the general level of accessibility
achieved over the last decades in standard HCI. When standard HCI changes also
accessibility has to change and this is closely related to games.
People with disabilities form one of those groups benefiting most from ICT3. Indeed,
Assistive Technology (AT) enables them in a lot of situations in their daily lives, at
school as well as at work or at home, in mobility, etc. The possibilities offered to the by
eInclusion makes a difference in the life of a lot of people. Therefore it seems important
that children get used to using technology as early as possible. Computer games are
often a good training for the use of AT, for children as well as for adults after accidents,
diseases. In addition playing games contributes considerably in establishing and ame-
liorating the skills in dealing with HCI.
2
Linden Lab, http://www.secondlife.com
3
Information and Communication Technologies.
520 D. Archambault et al.
2 Software Accessibility
Today it is state of the art that people with disabilities can interact with the standard
desktop/WIMP4based interface using Assistive Technology. Specific access software
applications, like screen readers and screen magnifiers, alternative input devices, al-
ternative information rendering – sound, text, signs, colour/size/contrast of objects,
allow them to access to many computer applications. This is mainly the case, as men-
tioned above, for text based software: word processors, spreadsheets, mail clients, web
browsers. The problem is that these access software applications are not able to access
any software application whatever the way it has been developed. Indeed they need to
collect accessible information from the applications to render it using alternative output
modalities, or to control them using alternative input modalities.
In other terms, to achieve accessibility of software applications, it is necessary to
have accessibility support embedded in the applications. During the last decade, ac-
cessibility frameworks have been developed and are available in the main environ-
ments. For instance, Microsoft has developed Microsoft Active Accessibility5, to make
their Windows Applications accessible, application developers have to implement the
IAccessible interface6. There exist similar frameworks on Mac7 and on Linux desktop
environments8. Theoretical works can be cited too [9]. Furthermore specific develop-
ment frameworks need to support accessibility, for instance Java9 and Mozilla10 .
It is not enough that applications respect accessibility standards. In most cases,
content must be accessible too. For instance, in the case of a web site, the accessibility
of web browser is necessary but the web contents must also be accessible. Graphical
elements for instance must have textual alternatives, and this depends on the content
itself. In that respect, the W3C launched the Web Accessibility Initiative to developed
Web Content Accessibility Guidelines [10]. These guidelines indicate how to use each
of the HTML tags to make a web site accessible. Accessibility of content has been
developed in other content formats such as the proprietary PDF and Flash formats11.
Of course there are still a lot of barriers in access to software and web content, but
basically there are technical solutions asking for according political and practical
measures to put this potential in place.
4
WIMP: Windows/Menus/Icons/Pointers
5
MSAA:http://msdn2.microsoft.com/en-us/library/ms697707.asp
6
http://msdn2.microsoft.com/en-us/library/accessibility.iaccessible.aspx
7
Apple accessibility: http://www.apple.com/accessbility
8
Gnome Accessibility: http://developer.gnome.org/projects/gap
KDE Accessibility: http://httpaccessbility.kde.org
9
Desktop Java Accessibility
10
Mozilla Accessibility Project: http://www.mozilla.org/access
11
Adobe Accessibility Resource Center: http://www.adobe.com/accessibility
522 D. Archambault et al.
Indeed, the results of a game can not be easily quantified, like in the standard case of
classical desktop applications. In a word processing software, it is easy to measure the
time needed by a user to write a document or to edit a document produced by a col-
league. In a game we can as well observe if a player succeeds, and measure the time to
finish a level or any case relevant for the game considered. But this is far not enough.
Unlike others software, games have to provide special good feelings to players. There
are probably some emotional factors to consider in the desktop applications, but they
are usually not taken into account, or at least unless they affect the productivity. In the
case of a game these factors are the most important.
Video images and audio messages contain emotional components and specific pat-
terns which can easily be perceived and due to empathy can be experienced by the
viewer and listener. For the same reason interactive video games are attractive and
popular among the youth and adolescence. Empathic arousal has a strong influence on
people viewing, listening and reading by forming their social response to the external
events through mental estimation of the problem and simulation of possible solutions
and actions [11, 12].
It has been demonstrated in numerous psychological studies that some emotions can
motivate a specific human action and behaviour. The development of the emotional
intelligence in youth depends on a social inclusion and personal experience which usu-
ally rely on observing others’ actions and behaviours presented in a real life (the cultural
milieu) and in the artificial situations disseminated by movies, television and video
games [13–16]. Being deprived of access to information with emotionally rich content,
blind and visually impaired children have experienced a significant emotional distress
which can lead to depression and deceleration in cognitive development [17, 18].
As we stated above: Accessible games must still be games! Visually impaired adults
in work situation accept relatively big constraint on usability to be able to use the same
software as their sighted workmates and to work on the same documents. This is not the
case with children, especially playing. In other terms it is not enough to find a technical
way allowing to access to all information needed in the interface, the result must be as
interesting and as usable as the original game, and additionally it must be possible to
succeed!
This helps us to understand that game interfaces are of a profound different nature
than standard HCI and use their own technology (game engines). Usability and acces-
sibility ask for freedom in time, speed, undo, mode of interaction,... It is a key criterion
outlined by the W3C/WAI guidelines and software accessibility guidelines that the
interface must not prescribe a certain interaction behaviour. But it is the core idea of
games for realising immersion into a game, joy and setting up the gaming feeling to
prescribe a restricted action and reaction behaviour and to force the user to be successful
in this “reality”. It seems that the more the player has to follow a strict behaviour, the
more it seems that the game “takes the player into it” and puts immersion in place.
Therefore game accessibility goes beyond standard HCI and content accessibility
measures. It must allow the prescription of behaviour by the system but it asks for
alternatives and freedom of adaptation in the level of prescription and usage of mo-
dalities of interaction. If a mainstream game has put accessibility in place it is the role
of adapted AT interfaces to realise immersion. Therefore it is inevitable to work on
these adapted AT interfaces in game accessibility.
Towards Generalised Accessibility of Computer Games 523
The first period that we identified is the period 2000-2005, that we will call the “basic
studies”. During this period we have seen the development of various games
specifically designed for specific groups of people with disabilities. These kind of
games are usually funded by foundations or non-profit organisations. Most of them
are very nice for the group they were developed for but have little interest for main-
stream, except maybe a few of the audio games. What is additionally of importance in
the context of this proposal is that they demonstrate how to render various interaction
situations with alternative modalities. This can be supplemented by the number of
research papers about new uses of various modalities in the game play (Braille de-
vices, haptic...).
The largest number of such games concerns audio games. Actually audio games
include three different concepts in which the main play modality is audio.
The first meaning involve mainstream video rhythm games like Guitar Hero II. The
second definition is related to artistic musical experiments. The third correspond to
games which is based on sound environment (sound scenes, sound characters, actions)
and can be played without vision, and therefore are accessible to visually impaired
players (like interactive audio books, stories and tales): In 10 years, over 400 accessible
audio games have been developed (which is very small as compared to video games).
The web site http://audiogame.net refers interesting interactive works. There exist a
few visual audio games which can be very impressive and playable as well with or
without sight. Terraformers [19] was developed with accessibility as part of the original
concept of the game. On the other hand, AudioQuake [20] was developed as a research
project to make a non-accessible game accessible.
A few tactile games can be found, that are games where the inputs and/or the outputs
are done by tactile boards or by Braille displays, in combination with usually audio
feedback. The use of Braille displays for gaming is only experimental by now. Some
research is currently carried out in order to find models to represent a 2D space on a
linear Braille display [21]. A few experimental games were designed in order to
evaluate these models, for instance a snake game and a maze game. During the TiM
project [22] (IST-2000-25298), a number of tactile games have been created or adapted
from existing mainstream contents. Tomteboda resource centre in Sweden have pub-
lished a report where they try to stimulate parents and educators to develop their own
games using tactile boards [23].
[24] studied the possibilities offered by haptic technologies for creating new inter-
actions usable by blind people. He worked especially with the Senseable Phantom.
Then a number of papers explore the possibilities of using haptics in experimental
games: [25–30].
524 D. Archambault et al.
The outcomes of this period is that we can base our work now on a lot of studies on
playing games in various situations of functional limitation and on the adaptation of
computer game situations.
playing the game it self which is fun and which provides learning benefits. In other
terms, games accessed with AT still must be games and due to this it challenges the
usage of AT. Then increasing accessibility of games will mean developing a new
generation of assistive software taking into account much more parameters than current
AT has access to, via the existing accessibility frameworks: characterisation of infor-
mation available (including ranking of importance regarding the current task to fulfil),
relative importance of events, speed, etc.
These new assistive software applications, which we will call Assistive Game In-
terfaces (AGI) will not likely be unique for a specific kind of impairment (like today a
screen reader allows to access any office application). Depending of the ability con-
straints, some could be dedicated to a specific game or a game engine, some would be
dedicated to a kind of games and finally some others would be generic (covering a large
range of games).
We could for instance imagine a ”captioning application” allowing lots of different
games to have captions when a character is speaking. On the other hand, for blind
gamers, we could have specific AGI for text based games, another AGI working with a
specific game engine, and a third one dedicated to a popular car race (since in this case
the interaction would have to be completely redesigned).
To achieve these goals, the AGI will need to collect information from the core of the
game itself. Indeed most of the information needed cannot be efficiently processed
automatically from the mainstream game (for instance the captioning information). We
have seen that the existing accessibility frameworks are not sufficient to provide these
AGI with the necessary information. This means that it is necessary to design a new
Game Accessibility Framework (GAF). This framework will have to take into account
the specific data needed by various alternative devices to work properly. To continue
with the example of the Captioning application, this application will need access to the
complete transcription of the texts spoken by the characters in the game. The Game
Accessibility Framework will have to specify how and in what format.
The first steps to carry out the specification of the Game Accessibility Framework
are (a) a typology of game interaction situations and (b) a characterisation of Accessi-
bility in terms of functional requirements. From the study of these expected results, the
specification of the GAF can be produced, including the data formats and exchange
protocols to transmit information between game and AGI.
Now it is time to make a significant move. This implies some participation from
assistive technology specialists as well as from mainstream games developers.
The proposed solution may seem not realistic but it has to be considered that:
– The state of the art shows that the technology is ready
– This solution will be the lighter for game developers (consider for instance the
work that would be needed to add a ”caption” option in a game, compared to the
implementation of the access to texts that are already existing somewhere in the
production process).
– the societal need for improving inclusion is growing in some leading countries
(Northern European countries, Austria, Canada, Japan, etc) and the political
pressure will necessarily follow, leading to laws and recommendations. We ex-
pect that this situation extends to the rest of Europe and North America, and to the
rest of the world.
– the evolution of standard HCI towards game like interfaces will soon make these
applications enter in the scope of existing laws
526 D. Archambault et al.
References
[1] Zyda, M.: Creating a science of games. ACM Communications 50(7) (July 2007)
[2] Natkin, S.: Video games and Interactive Media: A Glimpse at New Digital Entertainment.
AK Peters (2006)
[3] Kellogg, W., Ellis, J., Thomas, J.: Towards supple enterprises: Learning from N64’s Super
Mario 64, Wii Bowling, and a Corporate Second Life. In: “Supple Interfaces”: Designing
and evaluating for richer human connections and experiences (September 2007)
[4] Chatham, R.E.: Games for training. ACM Communications 50(7) (July 2007)
[5] Mayo, M.: Games for science and engineering education. ACM Communications 50(7)
(July (2007)
[6] Hildén, A., Svensson, H.: Can All Young Disabled Children Play at the Computer. In:
Miesenberger, K., Klaus, J., Zagler, W. (eds.) ICCHP 2002. LNCS, vol. 2398. Springer,
Heidelberg (2002)
[7] Archambault, D., Olivier, D., Svensson, H.: Computer games that work for visually im-
paired children. In: Stephanidis, C. (ed.) Proceedings of HCI International 2005 Confer-
ence (11th International Conference on Human-Computer Interaction), Las Vegas, Ne-
vada, July 2005, 8 pages (proceedings on CD-Rom) (2005)
[8] Buaud, A., Svensson, H., Archambault, D., Burger, D.: Multimedia games for visually
impaired children. In: Miesenberger, K., Klaus, J., Zagler, W. (eds.) ICCHP 2002. LNCS,
vol. 2398, pp. 173–180. Springer, Heidelberg (2002)
[9] van Hees, K., Engelen, J.: Non-visual access to guis: Leveraging abstract user interfaces.
In: Miesenberger, K., Klaus, J., Zagler, W., Karshmer, A.I. (eds.) ICCHP 2006. LNCS,
vol. 4061, pp. 1063–1070. Springer, Heidelberg (2006)
[10] W3C: Web Accessibility Initiative — Web Content Accessibility Guidelines 1.0. Tech-
nical report, World Wide Web Consortium (W3C) (May 1999),
http://www.w3.org/TR/WAI-WEBCONTENT
[11] Grézes, J., Decety, J.: Functional anatomy of execution, mental simulation, observation,
and verb generation of actions: a meta-analysis. Human Brain Mapping 12, 1–19 (2001)
[12] Prinz, W., Meltzoff, A.: An introduction to the imitative mind and brain. In: Meltzoff, A.,
Prinz, W. (eds.) The imitative mind: Development, evolution and brain bases, pp. 1–15.
University Press, Cambridge (2002)
[13] Segall, M.H., Campbell, D.T., Herskovits, M.J.: The influence of culture on visual per-
ception. Studies in Art Education 10(1), 68–71 (1968)
[14] Patterson, J.: Theoretical secrets for intelligent software. Theory into Practice 22(4),
267–271 (1983)
[15] Fromme, J.: Computer Games as a Part of Children’s Culture. The International Journal of
Computer Game Research 3(1) (May 2003)
[16] Sebanz, N., Knoblich, G., Prinz, W.: Representing others’ actions: just like one’s own?
Cognition 88, B11–B21 (2003)
[17] Barresi, J., Moore, C.: Intentional relations and social understanding. Behavioral & Brain
Sciences 19(1), 107–122 (1996)
[18] Chartrand, T.L., Bargh, J.A.: The chameleon effect: the perception-behavior link and so-
cial interaction. Journal of Personality and Social Psychology 76, 893–910 (1999)
[19] Westin, T.: Game accessibility case study: Terraformers -a real-time 3d graphic game. In:
Proceedings of the Fifth International Conference on Disability, Virtual Reality and As-
sociated Technologies, Oxford, UK, pp. 95–100 (2004)
Towards Generalised Accessibility of Computer Games 527
[20] Atkinson, M.T., Gucukoglu, S., Machin, C.H.C., Lawrence, A.E.: Making the mainstream
accessible: What’s in a game? In: Miesenberger, K., Klaus, J., Zagler, W., Karshmer, A.I.
(eds.) ICCHP 2006. LNCS, vol. 4061, pp. 380–387. Springer, Heidelberg (2006)
[21] Sepchat, A., Monmarché, N., Slimane, M., Archambault, D.: Semi automatic generator of
tactile video games for visually impaired children. In: Miesenberger, K., Klaus, J., Zagler,
W., Karshmer, A.I. (eds.) ICCHP 2006. LNCS, vol. 4061, pp. 372–379. Springer, Hei-
delberg (2006)
[22] Archambault, D.: The TiM Project: Overview of Results. In: Miesenberger, K., Klaus, J.,
Zagler, W., Burger, D. (eds.) ICCHP 2004. LNCS, vol. 3118, pp. 248–256. Springer,
Heidelberg (2004)
[23] Hammarlund, J.: Computer play for children who are severelly visually impaired: Using an
alternative keyboard with tactile overlays. Technical Report 20, Tomteboda resource
centre, Stockholm, Sweden (1999)
[24] Sjöström, C.: The sense of touch provides new computer interaction techniques for dis-
abled people. Technology and Disability 10(1), 45–52 (1999)
[25] Johansson, A.J., Linde, J.: Using Simple Force Feedback Mechanisms as Haptic Visuali-
zation Tools. In: Proc. of the 16th IEEE Instrumentation and Measurement Technology
Conference (1999)
[26] Wang, Q., Levesque, V., Pasquero, J., Hayward, V.: A Haptic Memory Game using the
STRESS2 Tactile Display. In: CHI2006, Montréal, Québec, Canada. ACM, New York
(April 2006)
[27] Raisamo, R., Patomäki, S., Hasu, M., Pasto, V.: Design and evaluation of a tactile memory
game for visually impaired children. Interacting with Computers 19(2), 196–205 (2007)
[28] Evreinov, G., Evreinova, T., Raisamo, R.: Mobile games for training tactile perception. In:
Rauterberg, M. (ed.) ICEC 2004. LNCS, vol. 3166, pp. 468–475. Springer, Heidelberg
(2004)
[29] Crossan, A., Brewster, S.: Two-handed navigation in a haptic virtual environment. In:
CHI2006, Montréal, Québec, Canada. ACM, New York (extended abstracts, 2006)
[30] Rodet, X., Lambert, J.P., Cahen, R., Gaudy, T., Guedy, F., Gosselin, F., Mobuchon, P.:
Study of haptic and visual interaction for sound and music control in the Phase project. In:
Proceedings of the 2005 conference on New interfaces for musical expression, Vancourer,
pp. 109–114 (May 2005)
[31] Grammenos, D., Savidis, A., Stephanidis, C.: Ua-chess: A universally accessible board
game. In: Salvendy, G. (ed.) Proceedings of the 3rd International Conference on Universal
Access in Human-Computer Interaction, Las Vegas, Nevada (July 2005)
[32] Grammenos, D., Savidis, A., Georgalis, Y., Stephanidis, C.: Access invaders: Developing
a universally accessible action game. In: Miesenberger, K., Klaus, J., Zagler, W., Karsh-
mer, A.I. (eds.) ICCHP 2006. LNCS, vol. 4061, pp. 388–395. Springer, Heidelberg (2006)
[33] Ossmann, R., Archambault, D., Miesenberger, K.: Computer game accessibility: From
specific games to accessible games. In: Mehdi, Q., Mtenzi, F., Duggan, B., McAtamney,
H. (eds.) Proceedings of CGAMES 2006 Conference (9th International Conference on
Computer Games), Dublin, Ireland, pp. 104–108 (November 2006)
[34] Tollefsen, M., Flyen, A.: Internet and accessible entertainment. In: Miesenberger, K.,
Klaus, J., Zagler, W., Karshmer, A.I. (eds.) ICCHP 2006. LNCS, vol. 4061, pp. 396–402.
Springer, Heidelberg (2006)
[35] International Game Developers Association: Accessibility in games: Motivations and
approaches (2004), http://www.igda.org/accessibility/
IGDA_Accessibility_WhitePaper.pdf
Designing Narratology-Based Educational Games with
Non-players
1 Introduction
Games are basic human activities and they are older than culture (Huizinga, 2006).
Computer games are also driven from social, cultural and economical aspects of so-
cieties. They are inseparable part of entertainment especially for youngsters (Fromme,
2003). Squire (2002) reports that games can be used to explore historical issues, in-
vestigate some complex learning situations, and they can manage and govern some
digital places even whole civilization. Squire and Jenkins (2003) state educational
values of games like socialization, interaction and understanding the concepts which
are aimed to give players are important for learning.
Games are investigated according to two major aspects; ludology and narratology
(Ang, 2006). These two aspects define complexity and interactivity of a game envi-
ronment. Especially narratives seem to be important issue for interactivity of player
and the game emotionally and narrative structures of games provide complexity of
games (Lindley, 2002). Characteristics of interaction in educational games make
students active, decision maker and problem solver. Narratives of games seem play an
important role in educational games for cognitively engaging students during learning
process. Therefore, in design processes of games, narrative structures and storytelling
parts should be constructed properly.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 528–534, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Designing Narratology-Based Educational Games with Non-players 529
Narratives provide decision points and interaction with game environment. Embed-
ding an educational context within a story and narrative structure might be beneficial
and easy for game designers during game design process. Therefore, it seems that
narratives are as important parts as ludology parts in educational games. Mateas and
Stern (2005) stated that small pieces of narratives provide a response from players
sequentially in a dynamic system, thus narratives can provide a local and global
agency to provide players valuable experiences. According to Mallon and Webb
(2005), narratives are not simple beings which can be analyzed in end-unit level,
rather they are complex beings which need detailed “specifications, properties and
concrete examples” to show the narratives in games.
According to Squire and Jenkins (2003) story of component of games is important
part of educational games especially for science education. They especially empha-
sized that stories of games should be explored in a way that stories help imagination
of students and what can be done for students to learn with stories in games. They
also added some research results that stories influence decisions of students related
their future career. Mallon and Webb (2005) urged that conflicts between game, inter-
action, hyper-structure and narratives should be solved and empirical studies should
be conducted to reveal several solutions between gameplay and narrative structure.
They also argued that lack of empirical studies that put what aspects of games provide
the significant experiences for players.
Computer games, especially role playing ones, make students force to follow and
think on problems (Squire & Jenkins, 2003). Narratives in games are important in role
playing games because narratives include who the player is, what he is supposed to
do, which decisions should be made etc. In this case, it should be considered that
narratives in games should be designed wisely to guide students’ learning process
throughout gameplay. Reiber (2001) stated that children like story of games, chal-
lenge, competition and the subject matters in game environment while playing educa-
tional games.
Juul (2001) believes that in games, players also produce stories, and games include
some narrative parts. Players can realize these narrative elements during the play, and
games and narratives have similar common structures. Therefore, in educational
games, developers should take into account integration of content with stories and
make students actively participate in these stories. Besides, teachers’ contributions are
invaluable for computer games (Squire & Jenkins, 2003) because teachers know
about target groups and how context and content can be integrated in a compatible
way. In this study, researchers’ aim was to understand how preservice foreign lan-
guage teachers embed educational context within a game environment and how they
organize storytelling and narrative structure in the game.
530 Y. Inal, T. Karakus, and K. Cagiltay
2 Methodology
In this study, the following research problems were examined;
1) What are the roles / missions / duties of avatars in designed educational
game environment?
2) What are the major patterns of preservice teachers in terms of storytelling in
educational games?
3) What type of narrative structures do preservice teachers prefer to embed
within the educational game environment?
4) What types of game genres do preservice teachers prefer to design educa-
tional games?
3 Results
In the study, students were asked to design educational games to teach English as a
second language basing on narratology. Results were categorized as “subject matter”,
“avatars in game environment”, “patterns of storytelling”, and “designed game genre
for educational games”.
Designing Narratology-Based Educational Games with Non-players 531
In the study, students were asked to select a subject matter concerning the scope of
English Teaching. Majority of them (n=24) preferred teaching “vocabulary” in a
game-based learning environment. Teaching names of the animals, numbers, coun-
tries, and shopping terms were most preferred topics under the vocabulary teaching.
“Directions” were second favorite subject matter (n=10) that students preferred. Find-
ing directions to save someone or reach target aimed in an adventure in game
environment were considered as a subject. Besides, teaching “grammar” was other
favored subject matter. Totally, 8 students stated that game-based learning environ-
ments might be appropriate for teaching grammar.
All of the students in the study preferred avatar-based educational game throughout
their designs. They gave a role or mission to avatars and story of the game was based
on these avatars. One of the students stated that “There is a pilot who encounters a
master caution during the flight. Therefore, he needs help immediately in order to
tend, so he tries to get help from the airspace”.
Most of the students tried to embed educational context in narrative structures and
storytelling in games, and avatars are the vital part of their game designs. They asked
players immerse avatars during game play, and while doing missions of avatars, play-
ers could learn the topic easily. For instance, one of the students stated “There is a
sick child coughing seriously, and saying “please help me!”. The mission of the
player is to help the child gathering clues in the game environment concerning the
illness. The game includes four stages and the player will try to find all illness by
giving right answers of vocabularies in the environment”. One of the interesting find-
ings of the study was that almost none of the students specified their avatars as male
or female. They only determined the missions or roles of the avatars player will use,
but they didn’t prefer explaining gender of avatars.
to solve this problem. In the game, player has to find the child”. To students, design-
ing educational games behind such storytelling patterns was more educational be-
cause player has to try to find correct answers or questions related to educational
context. Therefore, so as to motivate students and keep their attention high, before
asking question, such stories might be helpful.
All of the students preferred to use avatars while designing an educational game
and they preferred to give a mission or role to players. One of these missions was
helping an avatar (a character) in the game environment. Students tried to base stories
by saving human-beings in the world from disasters or catastrophes. One of the stu-
dents stated “Some of the animal species have been disappearing suddenly in the
ecology. However, this disappearing is not results of deaths of such animals. It was
suspected that some guys have been kidnapping these animals. The mission of the
player is following clues related to countries and their characteristics and give correct
answers to questions so as to find kidnappers”. Another interesting storytelling pat-
terns in this category was growing an object or bringing up a baby in the game envi-
ronment. Students aimed to teach vocabulary concerning an object or baby while
dealing with.
Finding directions in a labyrinth or puzzle environment was one of the most favor-
ite storytelling patterns of students. For the game environment, generally jungle, city,
under the sea or zoo was preferred. One of the students stated “In order to find the
treasure under the sea, the one thing that player has to do was to give correct answers
of animal names and meaning of them when they were asked. After each correct an-
swer, the player will take a clue and it will help him/her to reach the treasure.” Simi-
larly, another student stated “A group of students was taken to city zoo to visit the
animals. But, because the zoo was extremely large, some of them got lost. According
to the map they have, they will try to find exit of the zoo”.
Some of the students, especially females, preferred to design game environment
which are similar to Sim City or Barby-like games. They aimed to teach grammar or
vocabulary to players while designing game environment or dressing avatars with
clothes or other accessories. A female student stated “There is a 3D avatar in the
game environment. Players will select their own avatar at the initial point of the game.
Then, they will dress selected avatars in a shopping center by getting correct answer
of meaning of the clothes or accessories.”
4 Discussion
In the present study, preservice teachers’ educational game preferences basing on
storytelling and narrative structures in terms of teaching an educational context con-
cerning foreign language teaching were investigated. Results of the study revealed
that avatar-based educational games were mostly preferred. Participants aimed to
provide story by using an avatar and explain educational purposes with avatar-based
game design.
For vocabulary learning, there are many games and teachers prefer to create a
game for vocabulary. Role-playing games are most preferred one because most of the
preservice teachers give a mission to players. In this study, preservice teachers gave
importance to vocabulary teaching. They stated that a role-playing game basing an
avatar-based design might have beneficial and educationally valuable for learners.
Johnson, Vilhjalmsson and Marsella (2005) give importance to mission games in
language education because in these kinds of games, player has several target actions,
players are engaged to complete their missions, and they should interact with other
avatars by using verbal communication, and attempt to make some tasks. In foreign
language education, conversation is important to make practice with grammar and
vocabulary. Thus, it seems that selection of game type was related to teaching meth-
ods of foreign language in classrooms.
In the study stories of games were also related to real life. The participants
preferred some stories which might happen in classroom, helping someone to find
direction, stating the names of some objects and making conversations with some
people. This might be caused from teachers’ game playing habits and game prefer-
ences because participants of the present study were almost non-players. Dickey
(2005) argues that playing games provide new insights to develop different games. In
this context because teachers have little experience with games, it might be said that
they depended on their teaching strategies while designing game. Besides, teachers
tended to develop detailed stories rather giving just rules and actions in the game that
was because stories are important elements for games according to the participants.
With this implication it can be assumed that stories are attractive elements for foreign
language teachers to design / develop / or prefer educational games.
5 Implications
Computer game applications have been emerged in classroom settings for educational
purposes. However, it has been investigated for several years that effective and effi-
cient educational game design issues are not clear for both educators and game
designers. There need to be conducted studies focusing on game design basing on
educators’ preferences and needs. In the present study, a gap on educational game
design issue from educators’ points of views will be aimed to fulfill.
With the results of the study, it might be concluded that narratology-based game
design might be appropriate for educational games that are related to foreign language
teaching. Especially designing avatar-based educational games was very popular
among participants of the study. Developing a game by giving a mission to players
might be a way while designing a game. Besides, when the educational subjects that
534 Y. Inal, T. Karakus, and K. Cagiltay
participants preferred were considered, mostly vocabulary teaching was most pre-
ferred among all. We believe tat results of the present study will guide to both educa-
tors and educational game designers while designing games with foreign language
teaching.
References
1. Ang, C.S.: Rules, gameplay, and narratives in video games. Simulation & Gaming 37(3),
306–325 (2006)
2. Dickey, M.D.: Engaging By Design: How Engagement Strategies in Popular Computer
and Video Games Can Inform Instructional Design. Educational Technology Research and
Development 53(2), 67–83 (2005)
3. Durdu, P.O., Tufekci, A., Cagiltay, K.: A comperative study between METU and Gazi
University students: Game playing characteristics and Game preferences of University
Students. Euroasian Journal of Educational Research 19, 66–76 (2005)
4. Fromme, J.: Computer Games as a Part of Children’s Culture. Game Studies 3(1) (2003),
[online journal], (viewed 28 December 2007), http://www.gamestudies.org/0301/fromme/
5. Huizinga, J.: Homo Ludens. A Study of the play element in culture. Beacon Press (2006)
6. Johnson, W.L., Vilhjalmsson, H., Marsella, S.: Serious Games for Language Learning:
How Much Game, How Much AI? In: Proceedings of AIED 2005:The 12th International
Conference on Artificial Intelligence in Education, pp. 306–313. IOS Press, Amsterdam
(2005)
7. Juul, J.: Games Telling stories? A brief note on games and narrative. GameStudies 2(1)
(2001), [online journal] (viewed December 25, 2007), http://gamestudies.org/0101/juul-
gts/
8. Lindley, C.A.: The gameplay Gestalt, narrative and interactive storytelling. In: Proceed-
ings of computer games and digital cultures conference, June 6-8, 2002, Tampere, Finland
(2002)
9. Mallon, B., Webb, B.: Stand Up and Take Your Place: Identifying Narrative Elements in
Narrative Adventure and Role-Play Games. ACM Computers in Entertainment 3(1) (Janu-
ary-March 2005)
10. Mateas, M., Stern, A.: Build It to Understand It: Ludology Meets Narratology in Game
Design Space. In: DIGRA Changing Views & Worlds in Play, Vancouver, June 17 (2005)
11. Squire, K., Jenkins, H.: Harnessing the Power of Games in Education. Insight, The Insti-
tute for the Advancement of Emerging Technology in Education 3(1), 5–33 (2003)
12. Squire, K.: Cultural framing of computer/video games. GameStudies 2(1), [online journal]
(viewed December 28, 2007), http://www.gamestudies.org/0102/squire/
Interactive Game Development with
a Projector-Camera System
Andy Ju An Wang
1 Introduction
The computing discipline is currently facing an unprecedented array of pressures to
change. The enrollments in traditional Computer Science programs have been fluctu-
ating for the last a few years while the skill set of computing professionals keeps
growing in a rapid and unpredictable fashion. The computing profession itself is be-
coming more complex, with traditional disciplinary boundaries blurring or disappear-
ing in several emerging technological areas. According to a recent study conducted by
the Computing Research Association (CRA), the number of newly-declared computer
science majors in the fall of 2007 was half of what it was in the fall of 2000 – 7,915
versus 15,958. On the other hand, students have many misconceptions about computing.
Many believe that there are no opportunities and most do not see the connection be-
tween computing and innovations in fields such as entertainment. Some reforms in
computing education have been undertaken in response to these challenges.
More and more video game-related courses and degree programs are being offered
in colleges around the county in response to the digital media industry’s need for skilled
workers and the tastes of a new generation of students raised on Game Boy and Xbox
[6] – [13]. In the spring semester of 2006, the author offered a special topic course
entitled “IT 4903/6903 Entertainment Computing and Technology” at the School of
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 535–543, 2008.
© Springer-Verlag Berlin Heidelberg 2008
536 A.J.A. Wang
z If the player attempts to place a male professor in the wrong office, and the of-
fice is a female professor’s office, a scream is heard.
z If the player attempts to place a male professor in the wrong office, and the of-
fice is a male professor’s office, a male voice rebukes the player.
z Images of the professor will be displayed to reflect complementary characteris-
tics of the respective professor.
z Several images of students will also be in play. After the professor is in their
own office the student can enter to consult with the professor. Students cannot
enter a professor's office unless the professor is present. If the student attempts
to enter an office without a professor, a sound clip will play.
z At a random time a sound clip announces a faculty meeting is to be held and all
professors are to be herded into the central conference room.
z Professor's images are not recycled. When they drop to the bottom of the play
field they gather there until they are struck or popped up by the players. The
game is over when all professors are in their right office place. A congratulatory
sound clip is played at the conclusion of the game.
Pop the Professor was based on PlayMotion’s Critters game [1], which is a game that
contains areas into which the player moves an appropriate game piece. The player
moves a game piece with their hands and movement to match the appropriate place on
the projected screen or on a wall. These locations on the wall correspond to an area
called a grabber slot that matches the entity. A grabber slot can be moved to different
places on the playing field. Pop the Professor required multiple grabbers that could be
located on either side of the screen representing professor’s offices as shown in Figure
2 below. Electronically generated sketches of each professor’s face were presented to
the player to allow them to associate each professor with the appropriate office. Indi-
vidual cubes representing professors are released onto the playing field, the player then
maneuvers the cube into the appropriate office area.
Upon success, i.e., placing all professors into their appropriate offices, a new screen
is displayed as Figure 3 below, where the offices are moved to a different location.
Appropriate sounds are displayed when the cube is hit and upon successful completion
of the game, the professor’s voice stating her/his name, etc.
The feedback from students in developing this interactive game was very positive.
There were a few significant learning points worth for a discussion here.
z The team was exposed to a unique interactive technology, the camera-projector
system, in which the users of the system need no keyboard or mouse to play the
game. Instead, the users play the game by standing before a wall screen and
manipulate the game objects with their hands, body and movement.
z The team was exposed to a use of XML different from that seen in the usual Web
design context. The ability to set attributes for program elements using XML is
very powerful. In addition, it allows a level of abstraction such that the designer
can configure the system characteristics without having to manipulate the actual
game engine.
z Considerable knowledge was gained in capturing and manipulating audio and
video files. This involved working with multiple open source applications. It
allowed the team to evaluate and draw usability conclusions for many applica-
tions. The team discovered capabilities of Microsoft PowerPoint that they had
not previously been exposed to. Another interesting achievement was the ability
to create a background image for a page or presentation which looks similar to a
watermark.
would come across the screen, hearing the drum sound and learning from a pop-up text
panel containing pertinent information about the culture of the country accompanied by
a sound clip. A single user could play multiple drums while the drums were still moving
across the displaying surface, either a wall or a screen. Multiple users could interact and
play with the Drumline at the same time. After playing this game, the users would have
enjoyed, understood, and learned more about drum-related cultures. Drumline included
seven different kinds of drums from different continents. An XML file was used to
capture the characteristics of a drum and could be configured easily. A sample XML
file is shown below:
<?xml version=”1.0” ?>
<TITLE
baseDirec-
tory=”c:/Drumline/”>
<SOUNDTRACK volum=”0”
baseDirectory=”.”>
<PATH
file=”IntroDjembedrum32.wa
v”/>
</SOUNDTRACK>
<CONFIG>
<BALL radius=”1.2” mass=”300”
popUpDirectory=”../Drumlinefacts/AfricanCongaDrum”
popUpColor=”1.3,1.0,2” tiltAngle=”22.40”
popUpSound=”./Africandrum.wav” tex-
ture=”./AfricandrumCongaDrum.jpg”
quantity=”1” />
<!--More drums defined here-->
<TABLE
forceAmplifier=”0.8”
width=”32 height=”24”
catchTimeForPopUp=”2500”
popupTimeout=”5000”
shadowTimeout="5"
kineticFriction="0"
staticFriction="0"
backgroundFile="../../data/Drumline2/5DjembW.jpg"
gravityMagnitude="0.03"
collisionDetect="tight"
roll="on"
popUpSound="../../data/sounds/popup_open.wav"
popDownSound="../../data/sounds/popup_close.wav"
clackSound="../../data/sounds/Whomp-soft.wav"
screenWrapping="on"
shadow="../../data/textures/ballshadow.tga"
glowSprite="../../data/space/glow.tga" />
</CONFIG>
</TITLE>
The third project team developed an interactive game called “College Life”, tar-
geting new college students in a way that they could learn how to make a smooth
transition from high school life to college life, balancing their life and activities in a
college atmosphere. Activities were represented by images falling from the top of the
Interactive Game Development with a Projector-Camera System 541
screen. The players could grab objects that would have consequences on their everyday
life and future achievement. A score system was designed to challenge the player and
encourage them to make the right decisions. For instance, if the user grabbed SPSU 101
Orientation logo, her score would be increased by 50 points for participating in the
hitchhike course. The user had many objects to choose from that represent different
activities and social activities, some of which include parties, attending classes,
studying, tests, exercises, work, and dating. Selecting too many of either category of
activities would reduce the points, but selecting a well-balanced activities would make
the score increase until the graduation ball dropped then the game was over.
Our experiment utilized OpenCV and DirectX. OpenCV was used for image/vision
processing and DirectX taking the result and implementing it to the game. OpenCV is
an open source library for computer vision and image processing. It provides support
for applications such as human-computer interaction (HCI), object identification,
segmentation and recognitions face recognition, gesture recognition, motion tracking,
ego motion, motion understanding, and mobile robotics [2]. DirectX [3] is a Microsoft
product used especially in game programming because it contains numbers of Appli-
cation Programming Interface (API) to help users develop game. Components of
DirectX that are extremely useful are DirectDraw (produce raster graphics), Direct3D
(for 3D graphics), DirectSound (for playback and recording sound), DirectInput
(process data from keyboard, mouse, joystick, or game controllers), DirectMusic (for
playback soundtracks), DirectSetup (for DirectX components installation), and
DirectX Media (for animation).
542 A.J.A. Wang
The initial purpose of this project was to create an interactive game where players
can have fun and physical exercise at the same time when playing the game. The in-
tegration required that the game provide both proper content physical excise and game
play factors for enjoyment. Our game used a camera-projector system as depicted in
Figure 4, where the camera is used to capture the player motion and gesture image and
map it to the game content by OpenCV. The objects are drawn using DirectX com-
ponent (DirectDraw) and used as the “balls” in a typical “brick and balls” game. The
Mouse pointer and Bar are also created using DirectDraw. Background music is used
because the nature of this game is motion action and the combination of the two would
give the player more entertainment experience. There was a configuration file where all
the features of the game could be modified including the bar, objects, mouse pointer,
background image, background music, and click sound. This student team built this
interactive game using the camera-projector system we discussed in previous section.
From design to implementation and testing, a prototype was implemented. The game
idea was not easy to put into realization without a solid teamwork and significant effort.
Throughout the project, students learned to use some tools that they were not familiar
with such as OpenCV and DirectX programming. Other tools that were useful for the
project include Audacity, which is a powerful tool for sound editing and a freeware,
GIMP, which is also a freeware similar in its functions to Photoshop, and lastly Mi-
crosoft PowerPoint, which can be a practical tool for creating background and it is easy
to use.
art. The author is working on the combination of smart mobile devices with the cam-
era-projector system to provide more immersive and rich experience in interactive
games.
References
1 Introduction
In 3D computer graphics, an additional consideration is that rendering complex
shapes realistically requires significant resources, when very often the same shapes
can be effectively depicted in a line drawing style that uses less data, modeling effort,
and computation time. The most basic non-photorealistic rendering uses little or no
shading, and simply draws lines along silhouettes and sharp features. Recently, De-
carlo et al. introduced suggestive contours [1]. These are additional view-dependent
lines (like silhouettes) that convey a more complete impression of shape while adding
relatively few additional lines. Rendering is then equivalent to a resampling process
where surfels are blended with a Gaussian distribution in the image space[2].
The computational workload in graphics processing systems is generally split be-
tween a central processing unit (CPU) and a graphics processing unit (GPU). A com-
bination of software, firmware and/or hardware may be used to implement graphics
processing. For example, graphics processing, including rendering can be carried out
in a graphics card, graphics subsystem, graphics processor, graphics or rendering
pipeline, and/or a graphics application programming interface (API), such as
OpenGL.
The vertex shader is traditionally used to perform vertex transformations along
with per-vertex computations[3].Once the rasterizer has converted the transformed
primitives to pixels, the pixel shader can compute each fragment's color. This pipeline
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 544–550, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Animated Impostors Manipulation for Real-Time Display in Games Design 545
Characters
Updated Characters
Impostor Rendering
Game Console
3D Graphics
Central Processing Processing Video
Encoder
USB Host
Flash ROM Memory Controller
Perceptions Management
Fig. 1. The system structure of animated impostors manipulation for real-time display in games
design
Animated Impostors Manipulation for Real-Time Display in Games Design 547
3 Multi-resolution Modeling
In this section we show how a multi-resolution virtual human can successfully be
constructed and animated.
We first use Predefined Animations and Motion Capture: in many cases, the poses
of animations will be known in advance. This may be the case either due to manual
animation of frames, or the result of motion capture data. In the known examples,
some bones may have imperceptible movement, and it may be visually acceptable to
simplify their skin polygons as if they did not deform. From the examples given, we
can compute the probability distribution of the configurations directly, and use this
information to guide our simplification.
Our real-time virtual human model consists of an invisible skeleton and a skin. The
underlying skeleton is a hierarchy of joints that correspond to the real human main
joints. Each joint has a set of degrees of freedom, rotation and/or translation, which
are constrained to authorized values based on real human mobility capabilities [8].
Unlike other attempts we did not model a multi-resolution skeleton because our pur-
pose was not to demonstrate the effectiveness of animation level-of-detail. Hand
joints can be replaced with a single joint though.
After carrying the surface along with the particles, it is deformed under the action
of surface forces, similar to balloons. The forces are derived by minimizing the poten-
tial energy of the surface. The potential energy is composed of external potentials
which depend on the particles and internal potentials which depend on the surfels. We
derive an implicit and an attracting potential such that the energy is minimized when
the surfels are attracted to an implicit surface and to the particles, respectively. Mini-
mizing the internal potentials, consisting of the smoothing potential and the repulsion
potential, yields a locally smooth and uniformly sampled surface. From the potential
energy we derive forces acting on the surfels. While the derived forces from the im-
plicit, attracting and smoothing potential act in normal direction, the repulsion force is
applied in tangential direction.
Each primitive is attached to its proximal joint in the underlying human skeleton.
The set of primitives then defines an implicit surface that approximates the real hu-
man skin. Sampling this implicit surface results in a polygonal mesh, which can be
directly used for rendering. Sampling the implicit surface is done as follows: we start
by defining contours circling around each limb link of the underlying skeleton. We
then cast rays in a star-shaped manner for each contour, with ray origins sitting on the
skeleton link. For each ray, we compute the outermost intersection point with the
implicit surface surrounding the link. The intersection is a sample point on the cross-
section contour.
Depth information is estimated from multiple fixed cameras and allows easy
segmentation of the user from other people and background objects. An intensity-
invariant color classifier detects regions of flesh tone on the user and is used to iden-
tify likely body part regions.
As for the head, hands and feet, we still have to rely on a traditional decimation
technique to simplify the original mesh. Manual intervention is still needed at the end
of this process to smooth the transition between LODs(Levels of details). The body
548 Y. Yuan and L. Yan
extremities can cleverly be replaced with simple textured geometry for the lowest
resolution which dramatically cuts down the number of triangles.
Finally a face detection module is used to discriminate head regions from hands,
legs, and other body parts. Knowledge of the location of the user’s head in 3D is
passed to the application. Motion model has a sequence of motion parameters for each
action. In case of transmitted information with additional parameters such as 3-D
positions, the inverse kinematics theory is used for generating motion parameters
depending on the additional parameters. The system lets an avatar act according to a
sequence of motion parameters corresponding to an action decided by behavior
model. Fig.2 shown a few virtual humans in game design.
4 Results
The proposed technique was evaluated on a Pentium Xeon computer at 3.2Ghz with
512 MB of memory, and rendered to a 1280 x 1024 window. A different number of
characters was used to evaluate the performance of the proposed technique. Render-
ing performance for different number of characters is shown in figure3. Maximum
frame rates were achieved when rendering all characters as impostors, while mini-
mum frame rates involved a mixed rendering of impostors and instanced geometry.
Table 1.shows different number of characters.
Fig.4 shown the interface of games we designed.
60
Maximum frame rate
Minimum frame rate
50
40
Frame rate
30
20
10
0
0 200000 400000 600000 800000 1000000 1200000
Number of characters
improve the flexibility and efficiency of GPU crowd rendering. Thus the hardware
would be able to allocate appropriate storage for each invocation, as well as allocate
the number of indices generated by this invocation.
References
1. DeCarlo, D., Finkelstein, A., Rusinkiewicz, S., Santella, A.: Suggestive contours for con-
veying shape. ACM Transactions on Graphics 22, 848–855 (2003)
2. Alexa, M., Behr, J., Cohenor, D., Fleishman, S., Levin, D., Silva, C.T.: Point Set Surfaces.
In: Proc. of IEEE Visualization 2001, pp. 21–28 (2001)
3. Cabral, B., Cam, N., Foran, J.: Accelerated Volume Rendering and Tomographic Recon-
struction using Texture Mapping Hardware. In: Proceedings of the 1994 Symposium on
Volume Visualization, pp. 91–98 (1994)
4. Baerlocher, P., Boulic, R.: Task Priority Formulations for the Kinematic Control of Highly
Redundant Articulated Structures. In: IEEE IROS 1998, Victoria (Canada), pp. 323–329
(1998)
5. Boulic, R., Magnenat-Thalmann, N., Thalmann, D.: A Global Human Walking Model with
Real Time Kinematic Personification. The Visual Computer 6, 344–358 (1990)
6. Boulic, R., Becheiraz, P., Emering, L., Thalmann, D.: Integration of Motion Control Tech-
niques for Virtual Human and Avatar Real-Time Animation. In: Proceedings of the VRST
1997, pp. 111–118 (1997)
7. Bordeux, C., Boulic, R., Thalmann, D.: An Efficient and Flexible Perception Pipeline for
Autonomous Agents. In: Proceedings of Eurographics 1999, Milano, Italy, pp. 23–30
(1999)
8. Boulic., R., Capin, T.K., Huang, Z., Kalra., P., Lintermann., B., Magnenat-Thalmann, N.,
Moccozet, L., Molet, T., Pandzic, I., Saar, K., Schmitt, A., Shen, J., Thalmann, D.: The
Humanoid Environment for Interactive Animation of Multiple Deformable Human Charac-
ters. In: Proceedings of Eurographics 1995, Maastricht, August 1995, pp. 337–348 (1995)
Virtual Avatar Enhanced Nonverbal Communication
from Mobile Phones to PCs
Jiejie Zhu1, Zhigeng Pan1, Guilin Xu1, Hongwei Yang1, and David Adrian Cheok2
1
State Key Lab of CAD and CG, Zhejiang University, Hangzhou, P.R. China
2
Mixed Reality Lab, National University of Singapore, Singapore
1 Introduction
Phones are built for human-to-human voice connection for long years. To bring people
face-to-face in this situation, videophone technology are explored. These terminals
connect voice and video stream which is captured by phone camera simultaneously.
Thanks to recently growing bandwidth offered by wireless networks, such as UMTS,
WLAN, WiMAX, and the success of H.264, acceptable real time videophone are
available on mobile phones now.
Although mobile video phone has many advantages, no manipulation to the content
may make troubles. Sometimes, people are unwilling to expose their real feelings or
they may not show the real situation while talking to others. On the other side, most of
people want to embed their digital representations, such as cartoon like characters, 3D
talking heads and virtual avatars to mobile platform. At same time, people are likely to
use mobile phone as an important information input to rich their digital repository.
The main contribution of this paper is providing techniques to enrich user’s ex-
pressional content in their digital repository and studies technology transfer of content
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 551–561, 2008.
© Springer-Verlag Berlin Heidelberg 2008
552 J. Zhu et al.
modification from mobile phones to PCs. We select mobile phones as an input device
because it is most often used and most of them have built in cameras. We presents our
work on facial feature extracting, 3D face personifying, facial expression generating,
connecting technique both on mobile phone and Internet.
2 Related Work
A number of research of 3D face modeling have been proposed. Overall methods can
be grouped into two parts: geometry-based and image-based[1]. Geometry based
methods standardize the face framework which describes face model using grids, such
as polygons, surfaces and volume grids. It requires special 3D scan equipment to get
initialized data. The scanning process and data editing are tedious and time consuming.
Image based methods avoid scanning by recovering 3D information from multiple
views. Many vision information are extracted to accelerate this modeling process such
as structure information, stereo correspondences, face shading region and face silhou-
ettes. With prior knowledge of human face, image-based method achieves better ap-
pearance. However, they are subjective to image noise.
To improve the speed and quality of transmitting facial expression data on low-bite
network, several facial animating specification are reported[2, 3]. The most widely
used is Facial Action Coding System (FACS)[4]. A FACS coder ”dissects” an observed
expression, decomposing it into the specific Action Units (AUs) which produce the
facial movement. The scores for a facial expression consist of the list of AUs. MPEG-4
face animation standard[5] is another widely used coding standard. It supports the
transmission and composition of facial animation with natural video. Not like FACS,
facial animation parameter (FAP) set is defined based on the study of minimal facial
actions and is closely related to muscle actions. The FAP set enables model-based
representation of natural or synthetic talking-head sequences and allows intelligible
visual reproduction of facial expressions, emotions, and speech pronunciations at the
receiver[6].
To accelerate the rendering speed of 3D models on mobile phones, 3D graphic
hardware accelerators are designed and used. ARM[7] has implemented several types
of graphic hardware accelerator for mobile phones. Based on their reports, the ren-
dering capacity of ARM9, ARM10E and ARM11 series are all exceed the capabilities
of PCs of 1995. To enhance compressing capacity, MPEG hardware accelerator is also
embedded on mobile phone like SuperH Mobile Application Processor Series[8].
On-chip MPEG-4 hardware accelerator, it increases over twice MPEG-4 processing
performance. SH-Mobile series handle most MPEG-4 processing by means of mid-
dleware, which make it possible to implement high-performance, low power con-
sumption systems which incorporate moving-picture playback, Videophone, and
similar sophisticated functions.
However, realistic 3D face modeling and efficiently coding on low bit-rate wireless
network still remain an open issue. We focus on the first problem and presents our work.
Since MPEG-4 hardware accelerator has been available on mobile phones, we choose
MPEG-4 facial animation standard to code facial expression and transmit data both on
wireless network and Internet. In addition, we use VRML to render 3D face models on
Internet.
Virtual Avatar Enhanced Nonverbal Communication from Mobile Phones to PCs 553
Fig. 2. System Architecture across 3 platforms: Mobile Client on J2ME; Centric Server on Unix;
3D Modeler on Windows
Phone Client: This client is presently regarded as mobile phones with built in cam-
eras. To help user capture front view image and establish nonverbal communication, a
user interface is designed and encapsulates internal mobile phone APIs. Two main
sub-interfaces are implemented: facial feature point selection and personality cus-
tomization. Without them, selecting facial features and changing their 3D accessories,
such as hair color, skin color, configuring with glasses are unavailable. In addition,
communicating scheme to centric server is implemented using HTTP protocol, which
554 J. Zhu et al.
is a sub-protocol of WAP. This connection makes upload, download and change re-
sources available. All these mixed with voice connection, which means the client can
talk while doing nonverbal communication at the same time.
Centric Server: This server takes a role as middleware between mobile phone and
Internet. Each side is assigned one servlet to handle communications. On the client side,
Phone Servlet is compatible with common mobile phone servers. On the Internet side,
Modeler Servlet is used to exchange data between PCs and Centtric Server. While
communicating, polling and pushing techniques make upload and download resources
on this server available. The server also stores clients’ information in a database which
is accessible both by mobile phones and PCs.
3D Modeler: To reduce system burden on Centric Server, 3D face modeling is im-
plemented on this PC by 6 steps (see Figure 3).
Radial Distortion removes image distortions; Facial Feature Detection detects facial
features using Skin Color Possibility (SCP) and Gradient Change Map (GCM); Head
Model Adjustment module morphs general model to a personalized one; Texture
Mapping maps polished face image to the 3D face model and Facial Expression Gen-
eration creates several primary facial expressions by changing the position of detected
facial features.
4 Key Techniques
A friendly interface is very important for mobile phone application. Figure 4 illustrates
our implemented user interface.
Four layers are divided on the main screen from up to bottom. The up-most region
shows the program’s logo, memory size, battery power ; below this layer is a text
Virtual Avatar Enhanced Nonverbal Communication from Mobile Phones to PCs 555
Fig. 4. User Interface of Mobile Client: Left Image is our concept of user interface; the center
image is a given example of main menu; the right image is a given example of facial feature
selection
highlights of selected functions; on the third layer, the 3D face model is rendered by
JSR-184 functions[11] with 6 sub-menus surrounded it which rendered by MIDP2.0
APIs; on the last layer, it shows the description of function sequences.
We use two steps to detect facial features on user’s front view image. First step locates
the face region by Skin Color Possibility (SCP). Second step detects facial features
according to prior knowledge of human face’s division.
To locate user’s face region, we calculate the skin possibility of image pixels and
match them with sample SCPs. The probability is clustered in a small area of the
chromatic color space due to the color distribution of human skin and Gaussian dis-
tribution is used to calculate the likelihood of a skin pixel:
p(r ,b ) = exp[−0.5( x − m)T C −1 ( x − m)] (1)
Fig. 5. Intensity changing map: 5 regions are divided on this face template by different ratios. In
each region, intensity change is calculated and compared with templates.
(a) Original front view image. (b)Transforming RGB to YCC color space.
(c)Recognizing face region using SCP and deleting isolated points by image mor-
phology operation.
(d) Clipping the face region by ellipse fitting. (e)Two horizon gradient changes are
calculated for half face. (f) Determining which feature by comparing with ICM.
(g)Calculating vertical intensity changes. (h)Determining the facial feature points by
maximum intensity intersections.
Considering the changes of illumination condition, the skin possibility may not
correct. For example, the right side of the face in (c) is missed because the light is
occluded by left side. So, users have to capture their face image under a global light
environment. In addition, ellipse fitting requires high accuracy of isolated point dele-
tion. Otherwise, the ellipse will be too large with background.
calculate the face region and this time using the stored facial features. Again, an ellipse
approximation method is used to get the face region. Re-sampling algorithm resizes the
image and image operations blend it to a predefined skin background image. Figure 7
shows a sample of the results.
In our system, cylinder mapping algorithm is used to map face texture to a 3D face
model. Its process is shown in Figure
Fig. 7. Process of generating a visual face texture: (a)Original face image. (b)Ellipse face region.
(c) Re-sampled face region. (d)Re-sampled facial features. (e)Blended to a skin background.
(f)Final face texture after image operations.
This process maps a facial points on 3D face model to facial points on 2D face
texture. It first maps 3D points to a cylinder and then this cylinder is extended to an
image plane. Hardware matching of texture coordinates is carried to all facial points,
and non-facial points’ texture coordinates is achieved using 2D interpolation. Before
interpolation, the 3D points should build basic triangle strips. Since the number of 3D
points are constant, we can predefine these triangular strips. Figure9 shows a solution
using 23 points.
558 J. Zhu et al.
However, based on our work, we find no standard Mobile3D APIs for all mobile
phone available due to commercial protection. The most widely applied Mobile3D
specification in embedded system is OpenGL ES, others like OpenKode, OpenML,
Open VG and Microsoft’s Direct3D Mobile are also used. In addition, different Mo-
bile3D APIs are implemented using different language, like Mascot Capsule using C++
language and JSR-184 using Java language. All these prohibits a unified architecture
design.
Low memory and small screen are another bottle necks for 3D application on mobile
phone. The size of common 3G mobile phones’ memory is 64M with 200M HZ of
Fig. 10. 2 examples of facial expressions on PC. Left is in high resolution, while right is in low
resolution.
processing speed. The highest terminals can reach 1G memory and 600 M Hz. The size
of common 3G mobile phone’s screen is 240320. Powered by batteries, rendering high
quality of 3D graphic on these devices is challenging due to small resolution of the
display. From research results[17], the average eye-to-pixel angle is larger for mobile
phones than for PCs. This implies the quality in each pixel should be better for a mobile
device than for a PC.
With fast development of mobile phones’ hardware and wireless network, above
problems may be settled down in near future. We believe using mobile phone as an
input system to enrich one’s personal digital repository, and providing toolkit for
connecting mobile phone with Internet have a large potential industrial market. Our
next work will focus on video processing since mobile phone can record a clip of video
and send them to server in realtime now. Video-based reconstruction[18] and optimi-
zation[19] method will be studied first and results will be used to improve the quality of
face texture generation and 3D face model adjustment.
Acknowledgements
This work is co-supported by Key NSF Project on Digital Olympic Museum (Grant no:
60533080), 863 project (grant no: 2006AA01Z303), and Intel/University research
grant on 3D HCI and Realistic Avatars Based on Computer Vision. We’d like to give
thanks to Dr. You Kin Choong, Sucianto Prasetio, Seah Peck Beng, Clara. Their con-
tributions to this work are invaluable. Dr. Gaoqi He revised my description in detail.
References
[1] Xin, L., Wang, Q., Tao, J.H., Tang, X.O., Tan, T.N., Shum, H.: Automatic 3D Face
Modeling fromVideo. In: Proc. of IEEE ICCV, pp. 1193–1199 (2005)
[2] Wen, Z.C., Liu, M., Cohen, J., Li, K.Z., Huang, T.: Low Bit-rate Video Streaming for
Face-to-Face Teleconference. In: Proc. of IEEE International Conference on Multimedia
and Expo, pp. 309–318 (2004)
[3] Worrall, S.T., Sadka, A.H., Kondoz, A.M.: 3-D facial animation for very low bit-rate
mobile video. In: Proc. of 3rd International conference on 3G Mobile Communication
Technology, pp. 371–375 (2002)
[4] Ekman, P., Friesen, W.V.: Facial Action Coding System. Consulting. Psychologist Press,
Palo Alto, CA (1978)
[5] Eisert, P.: MPEG-4 facial animation in video analysis and synthesis. Journal of Imaging
Systems and Technology 2(3), 27–34 (2003)
[6] Tao, H., Chen, H.H., Wu, W., Huang, T.S.: Compression of MPEG-4 facial animation
parameters for transmissionof talking heads. IEEE Transactions on Circuits and Systems
for Video Technology 9(2), 264–276 (1999)
[7] http://www.arm.com
[8] http://cn.renesas.com
[9] Andrea, L., Ames, D.R., Nadeau, H.L.M.: VRML2.0 Sourcebook. John Wiley and Sons
Inc., New York (1997)
[10] http://www.blaxxun.com
[11] http://jcp.org/en/jsr/detail?id=184
Virtual Avatar Enhanced Nonverbal Communication from Mobile Phones to PCs 561
[12] Zhang, Z.Y.: A Flexible New Technique for Camera Calibration. IEEE Transactions on
Pattern Analysis and Machine Intelligence 22(11), 1330–1334 (2000)
[13] Harris, C., Stephens, M.J.: A combined corner and edge detector. In: Proc. of Alvey Vision
Conference, pp. 147–152 (1988)
[14] Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision. Cambridge
University Press, Cambridge (2003)
[15] Gong, Y., Sakauchi, M.: Detection of regions matching specified chromatic features.
Journal of Computer Vision and Image Understanding 61(2), 263–269 (1995)
[16] Pan, Z.G., Zhu, J.J., Hu, W.H.: Interactive Learning of CG in Networked Virtual Envi-
ronments. Computers and Graphics 2(29), 273–281 (2005)
[17] Jacob, S., Tomas, A.M.: iPACKMAN: High-Quality, Low-Complexity Texture Com-
pression for Mobile Phones. In: Proc. of Graphics Hardware, pp. 63–70 (2005)
[18] Wu, Q., Tang, X.O., Shum, H.: Patch Based Blind Image Super Resolution. In: Proc. of
IEEE ICCV, pp. 709–716 (2005)
[19] Shan, Y., Liu, Z., Zhang, Z.Y.: Model-Based Bundle Adjustment with Application to Face
Modeling. In: Proc. of IEEE ICCV 2001, pp. 644–651 (2001)
Analysis of Role Behavior in Collaborative Network
Learning
1 Introduction
Computer supported collaborative learning (CSCL) is changing the traditional ways
with the rapid development and effective application. Owing to implementation of
CSCL, learners, evaluators and tutors can fulfill their tasks on the basis of a collabora-
tive environment.
In recent years, a number of CSCL applications have appeared. For example, WEB
CT at the University of British Columbia has developed a multimedia platform for
learners [1]; Virtual-U, a cyberspace campus, can establish a collaborative learning
group based on different roles [2]; GRIDCOLE, a Grid Computing Environment, is
convenient for learners to access to grid resources, for educators to integrate applica-
tions, and for them to participate in collaborative learning applications [3].
The applications above provide multi-mode collaborative mechanism for learners
in network environment; however, there still exist several factors to be considered
when applied in practice: (1) Learning patterns in the network usually include tutor-
ship, fellowship, and individual and so on. Sometimes learner prefers to finish his task
individually in the collaborative environment. Users should be granted privilege to
select learning patterns instead of the assigned one during learning. (2) When a user
need find a suitable pattern to learn collaboratively, the awareness of role should be
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 562–572, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Analysis of Role Behavior in Collaborative Network Learning 563
provided appropriately. So he can get rid of his isolated feeling and interface with
other users in the environment. (3) If learning tasks are generated for a learner to
attain his goal, the knowledge background of learners, such as learning level, learning
motivation and learning style should be taken into account in the collaborative envi-
ronment. Learning tasks improper to a learner is likely to result in the failure.
Considering the factors above, in this paper we analyze role behavior in collabora-
tive network learning, indicate any learning organization must have its occurrence,
development, and disappearance, and disclose circulation of generating tasks; our
contributions can be summarized as follows:
(1) We introduce WF-net to describe role behavior in collaborative learning;
(2) We analyze the organizing of a learning pattern by awareness of role, and make
sure aim, condition, and procedure awareness of role;
(3) We provide a mechanism of generating tasks oriented roles, which enable a
CSCL system to run continually.
This paper is organized as follows. Section 2 defines notions such as knowledge
object, role, and member and so on. Section 3 introduces Petri net and describes of
organizing the learning pattern. Section 4 presents how to generate tasks oriented
roles in the learning organization. In Section 5 we give a CSCL example implemented
on the basis of our thought. Finally, we make conclusions in Section 6.
Petri net is a directed bipartite graph with two node types called places and transi-
tions. Directed arcs connected with the nodes. Places are showed by circles and transi-
tions by rectangles. Zero or more tokens are included in the Place, represented by dots
in circles. Places in the set correspond to conditions; transitions in the set correspond
to activities.
Definition 9. A Petri net structure is triple N (P, T, F) [8]:
(1)P= {Pl, . . .Pn} is a finite set of places with n>0;
T= {tl, . . . , tm}is a finite set of transitions with m >0 and P I T= Φ .
(2)F ∈ (P × T) U (T × P) is the flow relation, a mapping representing arcs between
places and transitions. The arcs represented by F prescribe pre- and post- relations for
places and transitions.
(3)p ∈ P,t ∈ T,·t={p|(p,t) ∈ F}is termed the pre-set of a transition t, and
t·={p|(t,p) ∈ F} is called the post-set of t . The pre-set and post-set for a place p are
defined similarly as the sets of transitions incident on s and following s, respectively.
Definition 10. A Petri net PN (P, T, F) is a WF-net (workflow-net) if and only if [9]:
(1) PN has two special places: i and o. Place i is a source place and place o is sink
place, satisfying ·i= Φ , o·= Φ .
(2)A transition t is added to PN which connects place o with i(i.e. ·t={o} and
t·={i}),the resulting Petri net is strongly connected.
t Signification
t1 Request for awareness
t2 Respond to awareness
t3 Decide whether to accept
t4 Acknowledge decision
t5 Execute tasks
t6 Request for dismission
T7 Respond to dismission
Fig. 1. Behaviors of roles with WF-net
A role should be entitled to select learning patterns instead of the assigned one during
learning. Usually organization of learning pattern is to build relations among roles by
566 X. Xu et al.
awareness, thus roles can complete tasks cooperatively at the next stage. Relations
among roles consist of ones between roles. Before a role possessed by an entity de-
cides to complete some tasks, it should tries to find another role in order that the for-
mer assign tasks to by the latter in an environment. If the former fails to do, it should
keep waiting. For example, a learner, as a real role, wants to select another role to
evaluating its learning result in the network, it requires finding an appropriate evalua-
tor and building mutual relations between them. If any, the learner can begin to learn
successfully in the collaborative environment. However, sometimes the learner cannot
find any of its proper evaluators, or all evaluators refuse to its request, the learner has
to wait until invited by a proper evaluator. Thereof, we declaim that organization of
collaborative learning is awareness of role. Awareness of role is corresponding to
stage 1 in WF-net (see Fig.1).
During awareness of role as above, it is important for a role to find another. As we all
know, an arbitrary role is belonged to an entity. The entity, either a member or an
agent, is distinct by its identifier. Let an entity have identifier E001, then Role(E001)
denotes a set of its roles ,and Knowledge(E001) denotes a set of its knowledge nodes.
So we can know the background information of any role by means of querying prop-
erties of the entity which possesses the role. When a role tries to find a proper group
to join in, it may check respectively up those roles who respond it according to their
background information from corresponding entities, then decide whether to join. So
awareness of role means to understand entity in a collaborative group.
We assume that r be a role. Before r tries to find a proper role, it gives constrained
conditions and broadcasts them in the collaborative environment (see transaction t1 in
fig.3).Only other roles satisfying conditions can respond to r and build relation with r.
The constrained conditions mainly embody three aspects: role type, knowledge rela-
tions and entity set. Let q be one of role objects who can respond r. So role type indi-
cates q must be an object of the role required by r; knowledge relation indicates that
entity possessing q must match to the set of knowledge nodes given by r; knowledge
relation’s value is in set {⊂ , ⊆ , =, ⊃ , ⊇} .entity set indicates that q must be longed to
one of entities limited by r.
For example, Let role r be a learner, the entity possessing r identified by E001, and
the entity possessing q identified by E002 .When r need awareness of evaluator, q has
responded to r. So q must be an evaluator. Meanwhile, if q is a qualified evaluator for
⊆
r, knowledge relation Knowledge(E001) Knowledge(E002) must be satisfied. Oth-
erwise q will fail to evaluate r because knowledge of entity E002 is less than that of
entity E001. If r prefers to evaluators coming from a special part of all entities, q must
be one element of the limited entity set.
Especially, Let role r be a learner, role q, q' be two evaluators, member M001 pos-
sess both r, and member M002 possess q'. When r need awareness of evaluator, q can
respond to r. if q has become the evaluator for r, then member M001 has to evaluate
himself, thus an individual learning has been built in a collaborative environment. On
the other hand, if q' becomes the evaluator instead of q, member M002 should bear the
task of evaluating member M001 after member M001, as a learner, finish a learning
task, thus a collaborative learning has been built in a collaborative environment. In a
word, there is not a distinct gap between individual learning and collaborative learn-
ing; they can transfer each other by means of role behavior in a collaborative envi-
ronment.
Step 1: Member M001 broadcasts its requesting for evaluator role in the group,
Role r becomes active, and its t1 is fired (see Fig.3): r gives its constrained conditions
with three parameters: T, R, and S, where T’s value is role type (i.e. evaluator), R’s
value is one element ( ⊆ ) from set {⊂ , ⊆ , =, ⊃ , ⊇} , and S dictates the limited entity
set satisfying S ⊆U.
Step 2: Member M002 receives the requesting message, checks every role object in
Role(M002) and finds q is an evaluator without much business. Role q becomes ac-
tive, and its t2 is fired (see Fig.3): q tests whether relation Knowledge
(M001) ⊆ Knowledge(M002) is satisfied, and whether M002 is in entity set S, if any, q
responses to role r.
Step 3: Member M001 receives the response messages from some roles; it selects
q from all responders by some algorithm, and build the relation with q. Then r’s t3 is
fired (see Fig.3): r decide to send acceptance message to q, and refusal message to the
rest. In its next step, one token will appear in p3.The token in p3 means that r is ready
to complete tasks or to generate complex organization by keeping awareness. On the
other hand, Member M001 doesn’t receive any response within limited time, r’s t3
also is fired and r restores its WF-net to the original status.
Step 4: Member M002 receives acceptance message, and build the relation with r.
Similarly,. q’s t4 is fired (see Fig.3): q acknowledges the decision from r .In its next
step one token will appear in p3. The token in p3 means q is ready to cooperate with r
or to generate complex organization by keeping awareness. For those entities who
receive refusal message, transition t4 of corresponding evaluator role also is fired: the
role restores its WF-net to the original status.
A correct schedule of awareness is the dot line showed in Fig.3: the dot line starts
from t1of r, passes by t2 of q, arrives to t3 of r, and end at t4 of q. After awareness
between two arbitrary roles, all roles establish collaborative relations. So they may
complete tasks regularly in the next stage; the detail will be analyzed in section 4.
If common goal of roles has been attained, collaboration among roles will terminate.
We should unchain relations of keeping awareness between two arbitrary roles, and
dismiss the old learning organization, so they can participate in awareness of role
repeatedly, a new learning organization may be generated.
Analysis of Role Behavior in Collaborative Network Learning 569
that p can evaluate the learning result of r; p builds relations with q such that q can
make learning plan according to evaluation and knowledge tree t.
In stage 2 of their WF-nets, by the awareness result of role we have
⎧q.execute (tk iq , tl iq ) = {( r , tk ir , tl ir )}
⎪
⎨r .execute (tk i , tl i ) = {( p , tk i , tl i )} .
r r p p
(2)
⎪
⎩ p.execute (tk i , tl i ) = {( q , tk i +1 , tl i +1 )}
p p q q
where i=1, 2… n.
In formula (2), symbols tk iq and tl iq denote that task tk iq with tool tl iq is assigned
to q in the ith time, rest of symbols have analogous meaning. At the beginning of
q q
collaboration (i.e. i=1), both tk1 and tl1 are default parameters given by the system.
Obviously, assignations of tasks among roles creates a cycle (See fig.4).After q fin-
ishes its tutoring task, it generates a new learning task for r. After learning generates,
a new evaluating task is sent to p. After examination of learning, p gives a new tutor-
ing plan to q. After analyze the tutoring plan according to background of the learner
and the knowledge tree, q lists another new learning task for r. So a collaborative
learning can keep continuing.
When function p.execute(tk ip , tl ip ) is called, a learning task tkir , a set of knowl-
edge units, is finished by learner r and sent from learner r to evaluator p as an element
of set tk ip .For any knowledge unit u∈ tkir , u.range is the learning result given by
evaluator p, is denoted as knowledge nodes mastered by learner r in knowledge unit
u, it is added to knowledge(E001) during evaluating. In a form, we have
Thus the updated knowledge (E001) is match up to current knowledge level of role r
in order that tutor q lists a new learning task for r according background of the learner
and goal of the learning.
Let r be a role, task tk be assigned to r with its tool tl, and the entity possessing r be
identified by E001. Suppose that function r.execute() is called, we have
r.execute (tk,tl)=Ø . (4)
Formula (3) means to stop generation of any new task, and terminate collaboration
between roles. In the example as above, if a tutor finds that his learner has mastered
all knowledge nodes (i.e. V(t)= knowledge (E001)), he should stop teaching the
learner, and dismiss the awareness; the learner does nothing because he can not
get any task from his tutor, but respond to dismissing from his tutor. All roles stop
Analysis of Role Behavior in Collaborative Network Learning 571
generating task, the collaborative learning terminates completely; all roles dismiss
awareness, the learning organization vanishes completely.
5 An Example
We have developed GLLS (Gas-Lift Learning System) to help students major in pe-
troleum engineering. The gas-lift is a basic mode of exploiting petroleum, so it is
important for students to design and analyze of gas-lift wells. They should master
knowledge nodes such as valves placement, gas allocation, simulation of pressure and
temperature in wells, valves diagnosis, gas-lift unloading, optimization of gas alloca-
tion in gas-lift block and so on. In the system, we introduce three roles: learner,
evaluator and tutor. Any user can play all roles. Firstly, those users build learning
organization by awareness of role, Secondly, every tutor creates learning task and
send to his student (See fig.5 (a)); the learner receives task and begin to finish it t(See
fig.5(b));the evaluator give scores to learners (See fig.5(c)).Since the given scores are
background knowledge for a learner, the tutor creates another new learning task ac-
cording them. Finally, users can dismiss the organization randomly. The learning
organizations are different owing to the fact that members possess roles.
The goal of the research presented in this paper is to disclose how roles organize a
learning pattern in the collaborative environment. We point out relations among no-
tions during collaborative learning. We introduce WF-net to describe role activity,
and find a learning pattern is established after awareness of role. We provide a
method of generating tasks oriented roles, which enable a CSCL system to run con-
tinually in a collaborative learning pattern.
In the future work, we are interested in the learning organization of adaptively,
changeability and complexity. One of the challenges in further researches is to ana-
lyze competition in collaborative learning. Both collaboration and competition should
appear in learning applications. We will try to discuss the cause and mechanism of
generating competition among roles, and focus on both relations and differences with
collaboration in collaborative network learning.
572 X. Xu et al.
References
1. Bruce, M., Chan, L.K.S.: Co-operative learning in integrated classroom. Curriculum and
Teaching 6(1), 48–52 (1991)
2. Celine, B.: Complementarity of information and quality of relationship in cooperative
learning. Social Psychology of Education 4(3/4), 335–357 (2001)
3. Bote-Lorenzo, M.L., Vaquero-Gonzalez, L.M., Vega-Gorgojo, G.: GRIDCOLE: a Grid
Collaborative Learning Environment. In: 2004 IEEE International Symposium on Cluster
Computing and the Grid, pp. 105–112. IEEE Press, Chicago (2004)
4. The Java Tutorials, http://java.sun.com/docs/books/tutorial/java/concepts/object.html
5. Guo, H., Sun, J.-m.: Research and design of intelligent teaching model and collaborative
learning mechanism. In: The 7th International Conference on Computer Supported Coop-
erative Work in Design, pp. 465–469. NRC Research Press, Ottawa (2002)
6. Mattas, A.K., Mavridis, I.K., Pangalos, G.I.: Towards Dynamically Administered:Role-
Based Access Control. In: Mařík, V., Štěpánková, O., Retschitzegger, W. (eds.) DEXA
2003. LNCS, vol. 2736, pp. 494–498. Springer, Heidelberg (2003)
7. Joshi, J.B.D., Bertino, E., Latif, U., et al.: A Generalized Temporal Role-based Access
Control Model. IEEE Trans. on Knowledge and Data Engineering 17(1), 4–23 (2005)
8. Reising, W.: Petri-nets: An introduction. Springer, New York (1982)
9. van der Aalst, W.M.P.: The application of Petri nets to workflow management. The Jour-
nal of Circuits, System and Computers 8(1), 21–66 (1998)
10. Eberspacher, H., Joab, M.: A role-based approach to group support in a collaborative
learning environment. In: Proc. of the Fifth IEEE International Conference on Advanced
Learning Technologies, pp. 64–65. IEEE Press, Taiwan (2005)
11. Zhang, R., Guo, L.-j., Liu, Z.: A novel a administrative model for collaborative learning
system. In: Proceedings of the Sixth International Conference on Machine Learning and
Cybernetics, pp. 4159–4162. IEEE Press, Hong Kong (2007)
Survey on Real-Time Crowds Simulation
1 Introduction
The wide use of computer graphics in education, entertainment, games, simulation,
and virtual heritage applications has led it to become such an important area of re-
search. In simulation, it is important to create an interactive, complex, and realistic
virtual world so that the user can have an immersive experience during navigation
through the world [1]. As the size and complexity of the environments in the virtual
world increase, it becomes more necessary to populate them with peoples, and this is
the reason why rendering crowds in real-time is crucial.
In general, crowd simulation constitute of three important areas. There are realism
of behavioral [2], high-quality visualization [3], and convergence of both areas. Real-
ism of behavioral is mainly targeted for simple 2D visualizations because most of the
attentions are concentrated on simulating the behavior of the group. . High-quality
visualization is regularly used for movie productions and computer games. In this
area, behavior is not really critical. The most important thing is how we can produce
very convincing visual. Convergences of both areas are mainly used for application
like training systems. In order to make the training system more effective, the element
of valid replication of the behavior and high-quality visualization will be added.
As we make comparison with non real-time system, developer for real-time crowd
simulation requires to consider various kind of challenges. One of the challenges is to
provide efficient management approach at every level of simulation in order to en-
sure, the agents composing a crowd should look different, move different, act differ-
ent, and so forth, similar to the real world. Another challenge is the increasing of
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 573–580, 2008.
© Springer-Verlag Berlin Heidelberg 2008
574 M. ‘Adi Bin Mohamed Azahar et al.
Fig. 1. This figure shows timeline on previous work for crowd simulation. More detail on
development of crowd simulations are discussed below.
The first procedural animation of flocks of virtual birds was shown in the movie by
Amkraut, Girard, and Karl called Eurhythmy, for which the first concept [5] was
presented at The Electronic Theater at SIGGRAPH in 1985 and the final version was
presented at Ars Electronica in 1989. The flock motion was achieved by a global
vector force field guiding a flow of flocks. A behavioral animation of human crowds
is based on foundations of group simulations of much more simple entities, notably
flocks of birds [6] and schools of fish [7].
Survey on Real-Time Crowds Simulation 575
In his pioneering work, Reynolds [6] described a distributed behavioral model for
simulating aggregate motion of a flock of birds. The technical paper was accompanied
by an animated short movie called “Stanley and Stella in: Breaking the Ice” shown at
the Electronic Theater at SIGGRAPH ’87. The revolutionary idea was that a complex
behavior of a group of actors can be obtained by simple local rules for members of the
group instead of some enforced global condition. The flock is simulated as a complex
particle system, with the simulated birds, called boids, being the particles. Each boid
is implemented as an independent agent that navigates according to its local percep-
tion of the environment, the laws of simulated physics, and the set of behaviors. The
boids try to avoid collisions with one another and with other objects in their environ-
ment, match velocities with nearby flock mates, and move toward a center of the
flock. The aggregate motion of the simulated flock is the result of the interaction of
these relatively simple behaviors of the individual simulated birds. Reynolds later
extended his work by including various steering behaviors as goal seeking, obstacle
avoidance, path following, or fleeing [8], and introduced a simple finite-state ma-
chines behavior controller and spatial queries optimizations for real-time interaction
with groups of characters [9].
Tu and Terzopoulos proposed a framework for animation of artificial fishes [7].
Besides complex individual behaviors based on perception of the environment, virtual
fishes have been exhibiting unscripted collective motions as schooling and predator
evading behaviors analogous to flocking of boids.
Brogan and Hodgins [10] and [11] simulated group behaviors for systems with sig-
nificant dynamics. Compared to boids, a more realistic motion is achieved by taking
into account physical properties of motion, such as momentum or balance. Their algo-
rithm for controlling the movements of creatures proceeds in two steps: first, a per-
ception model determines the creatures and obstacles visible to each individual, and
then a placement algorithm determines the desired position for each individual given
the locations and velocities of perceived creatures and obstacles. Simulated systems
included groups of one-legged robots, bicycle riders, and point-mass systems.
An approach similar to boids was used by Bouvier et al. [12] and [13] to simulate
human crowds. They used a combination of particle systems and transition networks
to model crowds for the visualization of urban spaces. At the lower level, attractive
and repulsive forces, analogous to physical electric ones, enable people to move
around the environment. Goals generate attractive forces, obstacles generate repulsive
force fields. Higher level behavior is modeled by transition networks with transitions
depending on time, visiting of certain points, changes of local population densities,
and global events.
Musse and Thalmann [14] and [15] presented a hierarchical model for real-time
simulation of virtual human crowds. Their model is based on groups, instead of indi-
viduals: groups’ are more intelligent structures, where individuals follow the groups
specification. Groups can be controlled with different levels of autonomy: guided
crowds follow orders (as go to a certain place or play a particular animation) given by
the user in run-time; programmed crowds follow a scripted behavior; and autonomous
crowds use events and reactions to create more complex behaviors. The environment
comprises a set of interest points, which signify goals and way points; and a set of
action points, which are goals that have some actions associated. Agents move be-
tween way points following Bezier curves.
576 M. ‘Adi Bin Mohamed Azahar et al.
Recently, Ulicny and Thalmann [16] and [17] presented a crowd behavior simula-
tion with a modular architecture for multiagent system allowing autonomous and
scripted behavior of agents supporting variety. In their system, the behavior is com-
puted in layers, where decisions are made by behavioral rules and execution is han-
dled by hierarchical finite-state machines.
Another work was exploring group modeling based on hierarchies. Niederberger
and Gross [18] proposed an architecture of hierarchical and heterogeneous agents for
real-time applications. Behaviors are defined through specialization of existing behav-
ior types and weighted multiple inheritances for creation of new types. Groups are
defined through recursive and modulo based patterns. The behavior engine allows for
the specification of a maximal amount of time per run in order to guarantee a minimal
and constant frame rate.
Most recently, a real-time crowd model based on continuum dynamics has been
proposed by [19]. In their model, a dynamic potential field integrates global naviga-
tion with moving obstacles, efficiently solving for the motion of large crowds without
the need for explicit collision avoidance.
3 Crowds Behavior
Crowd behavior modeling plays a major role in the domain of safety science and
architecture. Application such as crowd evacuation simulators need to apply crowd
behavior modeling to show believable visual and accurate result. Their objective is to
assist designers to understand the relation between the surroundings space and human
behavior [20].Many areas in computer graphics that can utilize crowds’ behavior,
such as:
1. Training and simulator system
2. Game industry
3. Simulation and animation
Crowd behaviors modeling play an essential part in police and military simulator
systems used for training. Simulator such as CACTUS developed to help in planning
and training for public order incidents such as large demonstrations. Game industry
nowadays still did not fully integrating virtual crowds in their game play environment
because of the need for high computational resource and costly production process.
Recently, there are some game genres that starting to change this situation, such as
real-time strategy genre that integrating groups of armies which will give direct effect
on the game play. In order to create an immersive simulation application using
crowds in virtual environments, researchers need to deal with various aspect of crowd
simulation such as behavioral animation, environment modeling, and crowd rendering
[4]. Nowadays, many movies and animation start to use crowd simulation in their
production. One of the most advanced crowd animation system for non-real-time
productions is Massive, this system was used to create battle scenes for The Lord of
the Rings movie trilogy.
There are two major crowd behavior models that have been used successfully,
which are PetroSim behavioral model [21] and physically based behavioral model
[22]. PetroSim is a system for simulation of evacuations of outdoor urban areas in real
Survey on Real-Time Crowds Simulation 577
time. Physically based behavioral model was proposed by Braun et al. [22] for simu-
lation of crowd evacuations from internal environments including several rooms and
obstacles.
4.1 Modeling
because they allow for neither procedural animations nor blending, and they also have
no composited, facial, or hand animation.
Impostors are the less detailed representation, and extensively exploited in the do-
main of crowd rendering [23] and [3]. An impostor represents a virtual human with
only two textured triangles, forming a quad, which is enough to give the wanted illu-
sion at long range from the camera. Impostor is only a 2D image of the posture that is
kept for each keyframe, instead of the whole geometry. Since impostors are only 2D
quads, normals and texture coordinates from several points of view need to be store,
so that, at runtime, when the camera moves, we can display the correct keyframe from
the correct camera view point. Since impostor is only 2D representation of the human
template, it’s usually placed farthest away from the camera.
4.2 Rendering
There are several technique used to speed up rendering process in crowd simulation.
Billboarded impostors are one of the methods used to speed up crowd rendering.
Impostors are partially transparent textured polygons that contain a snapshot of a full
3D character and are always facing the camera. Aubel et al. [24] introduced dynami-
cally generated impostors to render animated virtual humans. A different possibility
for a fast crowd display is to use point-based rendering techniques. Wand and Strasser
[25] presented a multiresolution rendering approach which unifies image based and
polygonal rendering. An approach that has been getting new life is that of geometry
baking. By taking snapshots of vertex positions and normal, complete mesh descrip-
tions are stored for each frame of animation as in the work of Ulicny et al. [26]. A
more recent approach to crowd rendering using geometry is through dynamic meshes
as presented in the work of de Heras et al. [27], where dynamic meshes use systems
of caches to reuse skeletal updates which are typically costly.
Fig. 2. This figure depicts the different stages: at each frame, data are flowing sequentially
through each one of them, beginning from the left, to the right [4]
5 Future Work
The Ancient Malacca Virtual Walkthrough [28] is a project that focuses on the model-
ing and visualization of Malacca city in 15th century. It is based on local and foreign
sources, such as the Sejarah Melayu and the descriptions by Portuguese writer, Tome
Pires. The focus area of visualization is Central Business and Administrative District
of the Malacca Empire. The project is visualized in real-time rendering mode using
SGI Onyx 3800 with 16 CPU, 32GB RAM and three Infinite Reality3 graphics pipes.
At the moment there are no crowd simulation have been develop for this project. For
future work, we will add crowd simulation into this walkthrough system with average
Survey on Real-Time Crowds Simulation 579
computational load that can be supported by normal personal computer. The chal-
lenge of this project is to bring crowd simulation into the walkthrough system on a
normal personal computer with the same or better quality compared to super com-
puter. As conclusion, we hope that this research is useful for the crowd simulation
system developer and also can give benefit to the computer graphics community.
Acknowledgement
We would like express our appreciation to Malaysian Ministry of Science, Technol-
ogy and Innovation under eScienceFund grant (01-01-06-SF0066) for financial
support of this research. We also like to thanks Creative Application Development
Centre (CADC), Multimedia Development Corporation, Cyberjaya, Malaysia for
allowing us to use the 3D model taken from Ancient Malacca project and permit us to
explore the potential improvement of the project.
References
1. Tecchia, F., Loscos, C., Chrysanthou, Y.: Visualizing Crowds in Real-Time. Computer
Graphics Forums (2002)
2. Thompson, P., Marchant, E.: Testing and application of the computer model ’simulex’.
Fire Safety Journal 24(2), 149–166 (1995)
3. Dobbyn, S., Hamill, J., O’Conor, K., O’Sullivan, C.: Geopostors: A real-time geome-
try/impostor crowd rendering system. In: SI3D 2005: Proceedings of the 2005 Symposium
on Interactive 3D Graphics and Games, pp. 95–102. ACM Press, New York (2005)
4. Thalmann, D., Musse, S.R.: Crowd Simulation. Springer, London (2007)
5. Amkraut, S., Girard, M., Karl, G.: Motion studies for a work in progress entitled ”Eurny-
thmy”. SIGGRAPH Video Review 21 (second item, time code 3:58 to 7:35) (1985)
6. Reynolds, C.W.: Flocks, herds, and schools: A distributed behavioral model. In: Computer
Graphics (ACM SIGGRAPH 1987 Conference Proceedings), Anaheim, CA, USA, vol. 21,
pp. 25–34. ACM, New York (1987)
7. Tu, X., Terzopoulos, D.: Artificial fishes: Physics, locomotion, perception, behavior. In:
Computer Graphics (ACM SIGGRAPH 1994 Conference Proceedings), Orlando, FL,
USA, vol. 28, pp. 43–50. ACM, New York (1994)
8. Reynolds, C.W.: Steering behaviors for autonomous characters. In: Proceedings of Game
Developers Conference 1999, pp. 763–782 (1999)
9. Reynolds, C.W.: Interaction with groups of autonomous characters. In: Proc. Game De-
velopers Conference 2000, pp. 449–460 (2000)
10. Hodgins, J., Brogan, D.: Robot herds: Group behaviors for systems with significant dy-
namics. In: Proc. Artificial Life IV, pp. 319–324 (1994)
11. Brogan, D., Hodgins, J.: Group behaviors for systems with significant dynamics. Autono-
mous Robots 4, 137–153 (1997)
12. Bouvier, E., Guilloteau, P.: Crowd simulation in immersive space management. In: Proc.
Eurographics Workshop on Virtual Environments and Scientific Visualization 1996, pp.
104–110. Springer, Heidelberg (1996)
13. Bouvier, E., Cohen, E., Najman, L.: From crowd simulation to airbag deployment: Particle
systems, a new paradigm of simulation. Journal of Electrical Imaging 6(1), 94–107 (1997)
580 M. ‘Adi Bin Mohamed Azahar et al.
14. Musse, S.R.: Human Crowd Modelling with Various Levels of Behaviour Control. PhD
thesis, EPFL, Lausanne (2000)
15. Musse, S.R., Thalmann, D.: A hierarchical model for real time simulation of virtual human
crowds. IEEE Transactions on Visualization and Computer Graphics 7(2), 152–164 (2001)
16. Ulicny, B., Thalmann, D.: Crowd simulation for interactive virtual environments and VR
training systems. In: Proc. Eurographics Workshop on Animation and Simulation, pp.
163–170. Springer, Heidelberg (2001)
17. Ulicny, B., Thalmann, D.: Towards interactive real-time crowd behavior simulation. Com-
puter Graphics Forum 21(4), 767–775 (2002)
18. Niederberger, C., Gross, M.: Hierarchical and heterogeneous reactive agents for real-time
applications. Computer Graphics Forum 22(3) (Proc. Eurographics 2003) (2003)
19. Treuille, A., Cooper, S., Popovic, Z.: Continuum crowds. ACM Transactions on Graph-
ics 25(3), 1160–1168 (2006)
20. Okazaki, S., Matsushita, S.: A study of simulation model for pedestrian movement with
evacuation and queuing. In: Proc. International Conference on Engineering for Crowd
Safety 1993 (1993)
21. Barros, L.M., da Silva, A.T., Musse, S.R.: Petrosim: An architecture to manage virtual
crowds in panic situations. In: Proceedings of Computer Animation and Social Agents, pp.
111–120. ACM Press, New York (2004)
22. Braun, A., Bodman, B.J., Musse, S.R.: Simulating virtual crowds in emergency situations.
In: Proceedings of ACM Symposium on Virtual Reality Software and Technology —
VRST 2005, Monterey, CA, USA, pp. 244–252. ACM, New York (2005)
23. Tecchia, F., Loscos, C., Chrysanthou, Y.: Visualizing crowds in real-time. Computer
Graphics Forum 21(4), 753–765 (2002)
24. Aubel, A., Boulic, R., Thalmann, D.: Real-time display of virtual humans: Levels of detail
and impostors. IEEE Transactions on Circuits and Systems for Video Technology 10(2),
207–217 (2000)
25. Wand, M., Strasser, W.: Multi-resolution rendering of complex animated scenes. Com-
puter Graphics Forum 21(3) (Proc. Eurographics 2002) (2002)
26. Ulicny, B., de Heras Ciechomski, P., Thalmann, D.: Crowdbrush: Interactive authoring of
real-time crowd scenes. In: Proc. ACM SIGGRAPH/Eurographics Symposium on Com-
puter Animation (SCA 2004), pp. 243–252 (2004)
27. De Heras, P., Schertenleib, S., Maïm, J., Thalmann, D.: Realtime shader rendering for
crowds in virtual heritage. In: Proc. 6th International Symposium on Virtual Reality, Ar-
chaeology and Cultural Heritage (VAST 2005) (2005)
28. Sunar, M.S., Mohd Zin, A., Tengku Sembok, T.M.: Effective Range Detection Approach
for Ancient Malacca Virtual Walkthrough. The International Journal of Virtual Real-
ity 5(4), 31–38 (2006)
TS-Animation: A Track-Based Sketching Animation
System
Abstract. Computer has been widely used in classroom. Yet most of them
can only provide static slides in the project mode, which are not attractive to
children, especially to the younger ones. On contrast animations are not only at-
tractive but also expressive. Using some animations in classroom makes the
children more active and enhances their learning efficiency. However, most of
current animation tools are either too complicated to common users, or specific
domain based. Aimed at removing the complexity barrier and allowing nonpro-
fessional users to create a wide variety of animations quickly, we propose a
sketching animation system. A kind of track-based method is adopted to make
creating animations as easy as drawing. Motion coordination is simplified by
providing the tool of motion time warping. The interface of our system allows
most of the important motions to be set by pen gestures.
1 Introduction
Since (at least) 1970s, computers entered classrooms as teaching assistants. Computer-
Aided Teaching has been of great benefits to teachers and students. The electronic
lecture systems, such as Microsoft’s PowerPoint, Apple’s Keynote and so on, are used
widely, which is often considered as the biggest technology revolution in classrooms
[1]. However, both of them don’t support freehand sketching, which is a familiar, effi-
cient, and natural way of expressing certain kinds of ideas, particularly in the class-
room. Teachers get used to sketching on the blackboard with chalk. Yet this archetypal
behavior is largely unsupported by current software used in classroom. More and more
researchers have recognized that such slides make the students keep in passive receiv-
ing, not active thinking, because the students could not get any response from their
teachers and are not involved in the lectures, since teachers can not interact with their
students in time [2]. Differently, pen-based interaction has some advantages of writing
easily, drawing freely, taking up just one hand, and so on. Hence, pen-based interaction
is particularly suitable for lecturing in class, since there is less text input, more sketches
and more operations in class, and since the teacher do not have much time to sit on a
chair in class in order to attract his students.
Although sketching rough diagrams is a useful tool in a classroom, sometimes static
diagrams are insufficient .Even very a simple animation can be more expressive, it is a
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 581–592, 2008.
© Springer-Verlag Berlin Heidelberg 2008
582 G. Wu, D. Wang, and G. Dai
convenient way to express moving visual images, it can represent dynamic concepts,
and it can make information more attractive and engaging[3], especially to the chil-
dren. Animation can also make students move active in the learning process. Research
on Pedagogy indicates that active learning could remarkably enhance students’ learn-
ing efficiency. And abundant interactions can give students participant feelings and,
hence, benefit students’ active learning and promote teaching effects.
The traditional way to create animation is to draw a series of images in a flip-book.
This is not only time consuming but also tedious. Although computer has been used to
help animators remove this tedium, one problem remains: animation tools and skills
are still in the hands of a small number of animation makers. Current animation tools
have extremely complex interfaces with many, specialized methods for generating
motion from static specifications [4]. This may be easier than designing every frame
by hand, but it is still too complicated for most of the teachers.
Aimed at these issues, this paper proposed a track-based sketching animation sys-
tem, which implements several methods to improve access to animation through a
simple interface for animation sketches. By focusing on motion-by-example and ges-
ture input, we hope to make animating to be an easy work, so that common users can
create animations without any complicated barriers.
This paper begins by reviewing the wealth of related work in animation tools. Then
we describe the track-based approach of animation creating, and motion coordination.
In Section 4 we give an outline of our system, its interface, a simple using scenario
and an informal evaluation. Conclusion and future works are given in Section 5.
2 Related Work
Researchers have long tried to create informal Animation sketching tools. The first
informal animation sketching tool was Baecker’s Genesys system [5]. In recent
years, the use on free–form strokes for creating 2D free–form objects and animation
has become popular. Richard C. Davis and James A. Landay conducted field studies
that investigated potential users’ needs in an “informal animation sketching” tool
.And they implemented a system that supports the most important needs, but they
neglected the coordination of motions, which is also an important aspect to create an
expressive animation. Tomer Moscovich and John F. Hughes discussed several
methods to aid sketching of animation, and had implemented a system as a test-bed
for these methods [6]. They proposed the approach of motion-by-example and even
synchronization, but their solution for object coordination is still incomplete. Their
work is the most similar to ours. There are also several complicated systems such as
TicTacToon [7], Fabian Di Fiore and Frank Van Reeth’s Multi–Level Sketching
Tool [8]. Although these systems can be used to create complicated and vivid
animations by professional users, they are too complicated for common users. For
education, there have been noteworthy systems that generate domain-specific 2D
animations from sketches of mechanical systems [9] and math equations [10].We
focus instead on general-purpose animations.
Free–form strokes are also being used for rapidly designing3D free–form objects
and 3D animations. Igarashi, Matsuoka, & Tanaka present a gesture–based sketching
interface (TEDDY) to quickly and easily design 3D free–form models out of 2D
TS-Animation: A Track-Based Sketching Animation System 583
sketched strokes [11]. Later, they introduce spatial key-framing, a technique for per-
formance-driven character animation [12]. This technique is especially useful for
imaginary characters other than human figures because it does not rely on motion-
capture data. On the contrary Matthew Thorne, David Burke, & Michiel van de Panne
presented a novel system for sketching the motion of a character [13]. Although
sketch has been used in 3D widely, creating 3D objects is still more difficult than 2D,
and some of them can only support very limited motions.
We attempt to create a simple sketching animation tool allowing inexperienced us-
ers to create a wide variety animation quickly. And we believe 2D animation to be
simpler and more accessible to the common users.
Fig. 1. The left is a rectangle containing one stroke only. The right is a clock composed by 5
strokes.
Fig. 3. The use’s selecting operations and the tree nodes created. The leaf-nodes are shown in
Figure 2.
Fig. 4. The user drags a rotation path and a translation path to a rectangle
One problem the user runs into when using track-based animation is that of coordina-
tion. In frame-based techniques, the animator can cause things to occur simultane-
ously, or in some other temporal relationship such as cause-and-effect simply by
drawing them in the appropriate frame [6]. In track-based animation, the animator
must rely on his sense of timing and response speed. For example, in Figure 4 the user
attempts to make the rectangle begin to rotate and translate at the same time, and also
586 G. Wu, D. Wang, and G. Dai
wants them to stop at the same time. First the user sets the rotate motion, then he/she
attempts to add the translate motion. After he/she sees the rectangle beginning to
rotate, he/she drags it to move, until he/she sees it stop rotating he/she stops moving
it. The progress depends on the user’s sense, and inevitably the translate motion starts
and stops behind the rotate motion. In order to help the user to succeed in this task, we
provide a time edit tool, which allows the user to reset the start and end time of a
motion. Assume the origin time of the points on a motion path is ( t1 t 2 … t m ), t1 is
the start time and t m is the end time. After the user resets time, the start time and the
* *
end time become to be t1 and t m respectively. We use the Formula (3) to reset all
points’ time on that motion path.
(t i − t1 )
t i* = t1* + (t m* − t1* ) (3)
(t m − t1 ) .
Sometimes, resetting the start and end time is not enough. As shown in Figure 6,
the user wants the rectangle to move along the sine curve, while the arrow to rotate
around the origin point. After reset their start and end time, they don’t collaborate
correctly in the mid time although they start and stop at the same time.
Fig. 6. Left: the rectangle’s translation motion and the arrow’s rotate motion. Right: the rectan-
gle moves to the position of α = π , but the arrow does not rotate to the degree of π .
We provide another tool to solve this problem, key-points. The user can set several
key points for a motion, seeing Figure 7. There are three key points on the translation
motion, and three key points on the rotate motion. Then the user adjusts the time of
the key points. He/she just need to adjust the collaborative key points to the same
time, then the motions can collaborate very well. There is no limit for number of key
TS-Animation: A Track-Based Sketching Animation System 587
Fig. 7. The key points set respectively for the translation motion and the rotate motion. Key-
point 1 and Key-point 4 have the same time. Key-point 2 and Key-point 5 have the same time.
Key-point 3 and Key-point 6 have the same time.
points the users can set for a motion. More key points a motion have, more well the
motion collaborates with others. However for most of simple motions, 3 to 5 key
points are enough.
4 System Description
We implement a system supporting track-based sketching animation, in which anima-
tion can be created using any of a variety of devices that provide the experience of
freehand drawing while capturing pen movement. We represent a stroke with the set
of points produced by the drawing implement between the time it contacts the surface
(mouse-down) and the time it breaks contact (mouse-up). A stroke can be either a part
of an object or a path of an animation. The animation methods supported by our sys-
tem includes translation of objects, rotation of objects, scale of objects, appear-
ance/disappearance, repeating motion and the composition of them. These include
most of the important methods used in an informal animation discussed by Davis RC
and Landay JA [4].
In order to make the system easy to use, we designed several gestures, selecting ob-
ject, setting key-point, deleting motions etc. When a stroke is inputted, the function of
Gesture and Sketch process is to identify the stroke as a gesture, a drawing or a mo-
tion (Figure 8).
The first part of Gesture and Sketch Process is a preprocessor, which gets the user’s
operation information. Pen is the main input device of this system. Hence, no matter
what mode the system is in, drawing or animation setting, the pen’ s movements and
588 G. Wu, D. Wang, and G. Dai
some other attributes are logged. The useful information for the following task includes
all the points on the pen’s trace, which are organized as a stroke, the speed, the pen’s
angle of inclination and so on. The strokes are real-time collected by the preprocessor
and sent to the classifier in time. Besides, it also does some work to make the strokes
visually more appealing without changing their basic shape. We refer the method de-
scribed by Tevfik Metin Sezgin [14].
The following part is a classifier, which classify the strokes into 3 categories:
gestures, drawings and motions, according to the context information and a group of
pre-defined rules. The last part is composed of three processors which are in charge of
motions, drawings and gestures respectively.
We have designed a simple animation user interface that will allow users to draw
objects and create motions that our system supports. Setting repeat times and editing
time of motions are also easy.
TS-Animation: A Track-Based Sketching Animation System 589
There are three areas, a blank canvas, a motion selecting area and a motion time ed-
iting area (Figure 9). Users draw objects and set motions for them on the blank canvas.
To define motions, the user presses “Start”, and all drawings and modifications are
recorded. In order to avoid confusion, the paths of motions are not visible, unless it is
selected to edit. When an object is selected, the selection widget in Figure 10 appears
and all the motions of the object are shown in the motion selecting area. In the motion
selecting area, user can select a motion just by double-click on it, delete a motion by a
simple delete-gesture, and reset its repeat num just by resetting the num on its top-left
corner. When a motion is selected, its start time, end time and key points’ time appear
in the time editing area. The user drags them to reset their time. The time editing area
can show two motions, so the user can make them coordinate easily.
Fig. 10. The selection widget with multiple control zones. Users may specify a variety of mo-
tions (such as translating or rotating) or other operations (such as moving the center of rota-
tion). This widget is similar to that of KSketch.
A teacher had prepared a lecture for the relationship knowledge among the sun, the
earth and the moon. The moon encircles around the sun while earth encircles around
590 G. Wu, D. Wang, and G. Dai
Fig. 11. Left: The teacher draws the sun, the earth and the moon. Right: The track for the moon
and the track for the earth.
the sun. In order to give a vivid image of this relationship, the teacher decided to use
an animation. First he draw the three celestial bodies on the blank canvas (figure 11
left), then he set an encircling path for the moon around the earth and an encircling
path for the earth around the sun (Figure 11 right). At last he reset their motion time
to make them synchronize well. It’s done.
In the classroom, the teacher showed the animation just by pressing the “start” but-
ton. The animation is shown in Figure 12. He also can pause or restart the animation
at anytime. Through this vivid animation, the students get to understand the relation-
ship more deeply.
After learning and practicing for a short time, they can all get hold of it. They find it
interesting, easy to use and can preserve a lot of time for simple animations. The
courseware makers said they are willing to use it as an aid. Besides, they give us a
lot of valuable suggestions to improve it. (1)One courseware maker said it is too
simple to support more complicated and attractive animations. (2)Composition of
different motions relies on their setting order, which may confuse the user. (3)
Morphing motion is important to create vivid animations. (4) Interactive animation is
a complement. In the future work we will take these issues into account.
Acknowledgement
The research is supported by National Grand Fundamental Research 973 Program
(Grant No. 2002CB312103), the National Natural Science Foundation of China (Grant
No. 60373056), National Facilities and Information Infrastructure Foundation for Sci-
ence and Technology(Grant No.2005DKA46300-05-32) and the CAS Pilot Project of
the National Knowledge Innovation Program (Grant No. KGCX2―YW―606).
References
1. Bonwell, Charles C., James A.: Eison, Active Learning: Creating Excitement in the Class-
room. ASHE-ERIC Higher Education Report. Washington, D.C. (1991)
2. Bransford, John, Brown, Ann, Rodney, C. (eds.): How People Learn: Brain,
Mind,Experience, and School. Committee on Developments in the Science of Learn-
ing,Commission on Behavioral and Social Sciences and Education, National Research
Council. National Academy Press, Washington (1999)
592 G. Wu, D. Wang, and G. Dai
3. Park, O., Hopkins, R.: Instructional Conditions for Using Dynamic Visual Displays: A
Review. Instructional Science 21, 427–448 (1993)
4. Davis, R.C., Landay, J.A.: Informal animation sketching: Requirements and design. In:
Proc. of the AAAI 2004 Fall Symp. on Making Pen-Based Interaction Intelligent and
Natural, Washington (2004)
5. Baecker, R.: Picture-Driven Animation. In: Proceedings of the AFIPS Spring Joint Com-
puter Conference, vol. 34, pp. 273–288 (1969)
6. Moscovich, T., Hughes, J.F.: Animation sketching: An approach to accessible animation.
Technical report, Brown University ( (2003)
7. Fekete, J.-D., Bizouarn, É., Cournarie, E., Galas, T., Taillefer, F.: TicTacToon: a paperless
system for professional 2D animation. In: Proceedings of SIGGRAPH, pp. 79–90 (1995)
8. Di Fiore, F., Van Reeth, F.: A Multi–Level Sketching Tool for “Pencil–and–Paper” Ani-
mation. In: Sketch Understanding: Papers from the 2002 American Association for Artifi-
cial Intelligence (AAAI Spring Symposium), Palo Alto (USA), March 25-27, 2002, pp.
32–36 (2002)
9. Davis, R.: Sketch Understanding in Design:Overview of Work at the MIT AI Lab. In:
AAAI Spring Symposium on Sketch Understanding, pp. 24–31 (2002)
10. LaViola, J.J., Zeleznik, R.C.: MathPad2: A system for the Creation and Exploration of
Mathematical Sketches. Proceedings of SIGGRAPH 23(3), 432–440 (2004)
11. Igarashi, T., Matsuoka, S., Tanaka, H.: Teddy: a sketching interface for 3D freeform de-
sign. Computer Graphics, 33(Annual Conference Series), 409–416 (1999)
12. Igarashi 1,3, T., Moscovich 2, T., Hughes, J.F.: 2Spatial Keyframing for Performance-
driven Animation Eurographics/ACMSIGGRAPH Symposium on Computer Animation
(2005)
13. Thorne, M., Burke, D., Van De Panne, M.: Motion Doodles: An Interface for Sketching
Character Motion. In: Proceedings of SIGGRAPH (2004)
14. Sezgin, T.M., Stahovich, T., Davis, R.: Sketch Based Interfaces: Early Processing for
Sketch Understanding. In: Proc. 2001 Workshop Perceptive User interface (PUI 2001), pp.
1–8. ACM Press, New York (2001)
Dynamic Axial Curve –Pair Based Deformation
1 Introduction
Deformation of 3D objects plays an important role in computer graphics, simulation
and computer-aided design. Using a deformation tool, a simple geometric model can
be deformed to take meaningful and creative shapes. Freeform deformation (FFD) [1 -
4] deforms an object by embedding the object in a space which can be warped, and
thereby deforming the embedded shape. Similar to FFD, other deformation tools [5, 6]
also require manipulating a large number of parameters to achieve a desired defor-
mation. The axial deformation technique was introduced by Lazarus et al. [7] which
deforms a 3D object by adjusting the shape of an axial curve. However, the lack of
control on the local coordinate frame of the axial curve may result in unexpected twist
of the object.
Hui [8] proposed a free-form design method using axial curve-pairs. A curve-pair
composing of a primary and an orientation curve provides explicit control on the local
coordinate frame of the axial curve. By associating 3D objects to the curve-pair, these
objects can be stretched, bended and twisted through manipulating the curve-pair. The
axial curve-pair deformation technique is an effective tool for manipulating complex
shapes in industrial and aesthetic design. However, to deform objects in a physically
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 593–601, 2008.
© Springer-Verlag Berlin Heidelberg 2008
594 M.L. Chan and K.C. Hui
viable manner, the control points of the axial curve-pair have to be adjusted such that
the axial curve-pair is deformed according to physical laws. Incorporating physical
properties in the axial curve-pair thus provides a convenient tool for deforming and
animating soft objects. The D-NURBS approach [10] effectively models an elastic
NURBS curve, but may not be applicable for modeling elastic curve-pair.
In this paper, a dynamic axial curve-pair is adopted. The deformation technique
makes use of an energy model, the mass-spring system, to emulate the physical prop-
erties of an elastic curve-pair as described below.
The curve-pair technique [8] provides explicit control on the local coordinate frame of
the axial curve. By associating an object to the curve-pair, the object can be stretched,
bended and twisted intuitively through adjusting the control points of the curve-pair.
Denote the primary curve as c(t), and the orientation curve as cD(t), where t’start ≤ t’ ≤
t’end , the curve-pair can be constructed as a set of k th-degree B-spline curves: c(t) =
∑ni=0 Ni,p(t) ui and cD(t) = ∑ni=0 Ni,p(t) vi, where ui is the i-th control point of the primary
curve, vi is the i-th control point of the orientation curve, vi = ui + d where d is the offset
distance, Ni,p(t) is the B-spline basis function of degree p. When a vertex of a model s is
attached to the curve c(t), a projection point sp = c(t) is obtained. A local coordinate
frame is constructed at sp as discussed below.
Dynamic Axial Curve –Pair Based Deformation 595
Fig. 1. The local coordinate frame of an axial curve-pair (left) The axial curve-pair model (right)
Using the same parametric value t at sp, a point sd is obtained by projecting the point
cD(t) onto the plane through the point sp with unit normal c’(t) / |c’(t)|. Hence, (sp – sd) ·
c’(t) = 0 and the local coordinate frame is given by
lz(t) = c’(t) / |c’(t)|
lx(t) = n(t) x lz(t) (1)
ly(t) = lz(t) x lx(t)
where n(t) = cD(t) – c(t) / |cD(t) – c(t)|
Assume the primary curve c(t) passes through point masses pi and the orientation curve
cD(t) passes through point masses qi. These mass points also satisfy the condition that
qi = pi + d, where d is the offset distance.
Given an axial curve-pair (c, cD), the local coordinate frame at pi is given by the unit
vectors li, mi, ti, where ti is the polygon tangent [8] at pi, and li = (qi – pi) x ti |qi – pi |,
mi = ti x li.
There are five degrees of freedom for the point pi including translation in the direc-
tions ti, li, mi and rotation about the axis li and mi . Similarly, there are six degrees of
freedom for the point qi including translation in the directions ti, li, mi, and rotations
about the axis ti, li and mi.
The motion of pi is governed by the mass-spring system simulating the bending and
stretching of the curve-pair. However, the motion of qi is not fully controlled by the
system. When pi is modified, qi has to be modified correspondingly. Only the rotation
axis ti of qi is governed by the dynamic system to represent the elastic twisting prop-
erties of the curve-pair.
Geometric parameters such as the rest length between two control points, the initial
angle between two local coordinate frames, and the tangent at a point mass determines
the geometry of the mass spring model. Physical parameters including forces, masses,
stiffness and stress governs the physical behavior of the system, and which can be
obtained from the geometric parameters.
596 M.L. Chan and K.C. Hui
A mass density m is associated with the point masses pi and qi. The linear stress at a
point mass is induced by the variation in linear displacement from the rest-length of a
spring along the direction of the spring. The torsional stress at a point mass is induced
by the variation in angular displacement from the rest-angle between the lattice edge
and lattice frame (Figure 2). To express these stress behavior, spring elements are
connected between the point masses. Different spring elements are adopted to represent
the different physical behavior of a curve-pair model.
Two types of springs are adopted. They are the metric springs and torsional springs.
The structure of the mass-spring model is specially designed for the 1-dimensional
curve-pair model. To maintain the linear structure and to capture the tensile stress
behavior, metric spring elements are connected between neighboring point masses (e.g.
pi, pi+1) of the mass spring model.
Elastic bending is modeled by two types of spring elements. The angular spring
elements defined at the angle of the lattice edge and the neighboring local coordinate
frame axes provide a controllable bending deformation. To capture the torsional stress
behavior, torsional spring elements are connected between the local coordinate frames
(li, mi, ti) at pi and (li+1, mi+1, ti+1) at pi+1.
in Figure 3. Equation (2) can be written in matrix form to obtain Equation (3), in which
R represents the 2N x 3 dimensional position vector, that is, [ p0 p1 … pN-1 q0 q1 …
qN-1]T. N is half the total number of point masses.
MR” + DR’ + KR = Fexternal (3)
In Equation 3, M and D are 2N x 2N diagonal mass and damping matrices respec-
tively, K is the 2N x 2N symmetric sparse matrix, and Fexternal is the 2N x 3 force
vector. Assuming there is no damping, Equation (2) can be simplified to:
MR” + KR = Fexternal (4)
The total force on the model is a vector sum of the internal and external forces. The
internal forces can be the tensile stresses or torsional stresses. User-interactions, reac-
tion forces and gravity are treated as independent external force in the dynamics for-
mulations. In the next section, the formulations of different forces are described.
where ti is the polygon tangent at pi, and li = (qi – pi) x ti | qi – pi |, mi = ti x li. The
torques are then converted to virtual forces on the neighboring point mass.
Fθj = Kθ (Δθ ) nθ,i * Pij (8)
λ β
The same holds for F i and F i. A restoring force is applied to pj in order to balance
the torque. That is Fθi = -Fθj. For the twisting deformation on Pij, the torque at the point
mass pi is proportional to the angular displacement Δω. The torque can be expressed as:
τωi = Kω (Δω) ti (9)
where Kω is the angular stiffness, ti is the tangent vector at pi, and Δω = ωi –ωj. The
torque is then converted to virtual forces on the point mass qi and qj of the orientation
curve and they are expressed as:
Fωi = Kω (Δω) ti * li (10)
Fωj = -Kω (Δω) tj * lj
Fig. 3.
Dynamic Axial Curve –Pair Based Deformation 599
3 Implementation
The proposed technique is implemented on a P4 3.0GHz PC with a GeForce 6600
256MB display card using VC++ and OpenGL. Figure 4 shows an example of simu-
lating the cloth movement of a game character. Dynamic axial curve-pair is used to
generate a prescribed cloth animation which can be further edited to produce a desired
effect. Figure 5 shows an example of simulating the movement of a ribbon in a ribbon
dance. A frame rate of 52 frames per second is achieved.
4 Conclusion
We have proposed and implemented a framework for physics-based axial deformation.
Our approach extends the axial curve-pair deformation technique to incorporate elastic
deformation. The technique provides a solution to a problem in axial deformation
where the lack of control on the local coordinate frame may lead to unexpected twist of
600 M.L. Chan and K.C. Hui
References
1. Sederberg, T.W., Parry, S.R.: Free-form deformation of solid geometric models. In: ACM
Computer Graphics (SIGGRAPH 1986) (1986)
2. Chadwick, J.E., Haumann, D.R., Parent, R.E.: Layered construction of deformable animated
characters. Proceedings of ACM SIGGRAPH, Computer Graphics 23(3), 243–252 (1989)
3. Faloutsos, P., van de Panne, M., Terzopoulos, D.: Dynamic Animation Synthesis with
Free-Form Deformations. IEEE Transactions on Visualization and Computer Graphics
(1997)
4. Feng, J.Q., Ma, L.Z., Peng, Q.S.: A New Free-form Deformation through the Control of
Parametric Surfaces. Computers&Graphics 20(4), 531–539 (1996)
5. Borrel, P.: Simple constrained deformations for geometric modeling and interactive design.
ACM Transactions on Computer Graphics 13(2), 137–155 (1994)
6. Terzoppoulos, D., Qin, H.: Dynamic NURBS with geometric constraints for interactive
sculpting. ACM Transactions on Graphics 13(2), 103–136 (1994)
7. Lazarus, F., Conquillart, S., Jancene, P.: Axial deformations: an intuitive deformation
technique. Computer-Aided Design 26(8), 603–617 (1994)
8. Hui, K.C.: Free-form design using axial curve-pairs. Computer-Aided Design 34, 583–595
(2002)
9. Christensen, J., Marks, J., Ngo, J.T.: Automatic motion synthesis for 3D mass-spring
models, Tech.Rep., MERL TR95-01 (1995)
10. Qin, H., Terzopoulos, D.: D-NURBS: A physics-based framework for geometric design.
IEEE Transactions on Visualization and Computer Graphics 2(1) (March 1996)
11. Volino, P., Magnenat-Thalmann, N.: Comparing Efficiency of Integration Methods for
Cloth Animation. In: Proceedings of Computer Graphics International 2001, Hong-Kong,
China, pp. 265–274 (2001)
12. Choi, K.-J., Ko, H.-S.: Stable but responsive cloth. In: Proceedings SIGGRAPH 2002, San
Antonio, TX, USA, pp. 604–611 (2002)
13. Baraff, D., Witkin, A., Kass, M.: Untangling Cloth. In: Proceedings SIGGRAPH 2003, San
Diego, CA, USA, pp. 862–870 (2003)
14. Eberhardt, B., Weber, A., Strasser, W.: A Fast Flexible Particle-System Model for Cloth
Draping. IEEE Computer Graphics and Applications 16(5), 52–59 (1996)
Dynamic Axial Curve –Pair Based Deformation 601
15. Haumann, D.R., Wejchert, J., Arya, K., Bacon, B.: An application of motion design and
control in physically-based animation. In: Proceeding of Graphics Interface 1991, pp.
279–286 (1991)
16. Peng, Q.H., Jin, X.G., Feng, J.Q.: Arc-Length-Based Axial Deformation and Length Pre-
serving Deformation. In: Computer Animation 1997, pp. 86–92. IEEE Computer Society,
Geneva (1997)
17. Faux, I.D., Pratt, M.J.: Computational geometry for design and manufacturing. Wiley,
Chichester (1979)
18. Klok, F.: Two moving coordinate frames for sweeping along a 3D trajectory. Com-
puter-Aided Geometric Design 3, 217–229 (1986)
19. Lossing, D.L., Eshleman, A.L.: Planning a common data base for engineering and manu-
facturing. In: SHARE XLIII, Chicago (August 1974)
20. Fehlberg, E.: Low-order classical Runge-Kutta formulas with step size control and their
application to some heat transfer problems, NASA Technical Report 315 (1969)
21. Cash, J.R., Karp, A.H.: A variable order Runge–Kutta method for initial value problems
with rapidly varying right-hand sides. ACM Trans Math Software 16, 201–222 (1990)
3D Freehand Canvas*
1 Introduction
Modern 3D animation movies are usually created through the following complicated
steps, including modeling, texture mapping and varieties of methods of rendering with
professional 3D animation software systems, which are exemplified by Autodesk 3D
MAX and Maya. The whole process must be done by a group of well-trained craftsmen
who are familiar with the complicated software and have a lot of experience on
computer animation work.
Some features of the traditional 2D animations, however, sink into oblivion because
of the emergence of the 3D animation, taking example as the freehand sketching style.
Additionally, many excellent 2D cartoon designers still stand far apart from the 3D
computer animation for the strange operations on modeling with a complex 3D
software system. They prefer freehand sketching style even giving up the convenient
3D features. So they choose the tools like Painter or Adobe Flash to create styled 2D
cartoons. The flaws can't be avoided either, which are that Painter supports varieties of
stroke with no animation functions, and Flash provides abundant animation functions
while it poorly supports the bitmap image, that means the lack of strokes and styles.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 602–612, 2008.
© Springer-Verlag Berlin Heidelberg 2008
3D Freehand Canvas 603
Moreover the workload on dealing with those 2D animation tools is obviously heavier
than that on handling 3D animation tools.
3D Freehand Canvas combines the advantages of 2D and 3D system. It is based on
lines with sketching style which is considered as basic element of constituting an
object, and includes translation of viewpoint for generating shielding effect. It provides
a novel fashion for artists to create artistic work of 3D freely. And it is also software
that is used to make 3D storyboard for supporting freehand sketching in the 3D space
and animation creation. It is a convenient tool for artists who prefer creating artistic 3D
works by using freehand drawing to doing them by modeling. Artists could get a "What
you see is what you get" result. Additionally, 3D Freehand Canvas supports some
animation functions which allows the objects and camera do some given movement.
2 Related Work
Freehand sketching plays an important role in both the past and the present phases of
the design process. Because of being extremely intuitive mode to by using pencils and
paper and they are always accessible, Most of the artists are inclined to drawing by
hand with them. Sketchy presentation [10] could help the designers to communicate
with each other better; the cognitive processes of the designer [7] can be expressed
clearly though sketching. The goal of the research is to transform the way of freehand
drawing to 3D data.
Some related work has been done. One approach to this is the usage of some 3D
input devices, such as tracking devices, data gloves or armatures which are used in
computer animation widely, there are many systems which has used 3D input devices
for sketching, for example, 3-Draw [8], 3DM [2], HoloSketch [4] and the system
presented by Diehl et al. [5]. Other systems which is for instance Tractus[9] and 3D6B
editor[6] use the normal input devices such as mouse and canvas. Tractus supports a
drawing canvas to do vertical movement, which makes the device able to maintain
direct spatial mapping between real and virtual spaces. The method how Tractus allows
drawing of non-planar curvature is remarkable. When drawing with a pen on the canvas
surface and moving the canvas vertically, complex 3D paths can be constructed by
user, which are difficult to input with 2D methods. 3D6B editor uses 2D input for
generating projective strokes on a user-definable 3D grid. The rendering is done with a
line-based renderer, and the data is never converted to 3D surface models but rather
stored as strokes. The systems above achieve some functions of 3D freehand sketching.
However, the shortage is obvious. Both of the systems deal with the simple lines
without any styles and strokes. This is disadvantageous for artists to do the artistic
works. Additionally, the drawing function is too limited to go into the realistic
animation process. At last, they don’t have any animation functions.
Another path in the research is to use the strokes or paths directly without trying to
convert them to 3D models. Cohen et al. [3] introduces a system where a 3D curve is
defined by first drawing its screen plane projection and then its shadow on the floor
plane. And Tolba et al. [11] present a system in which sketches of 3D space with fixed
604 M. Wang et al.
camera position can be made. However, these results are only panoramic sketches in
the whole scene, not full 3D with six degrees of freedom.
3D Freehand Canvas is the software that is used for freehand sketching and
animation producing. It is a convenient tool for artists who prefer to freehand drawing
to create the artistic 3D works. Additionally, 3D Freehand Canvas supports some
animation functions which allows the objects and camera do some given movement.
The traditional storyboard features are extended.
3 System Structure
stroke, firstly, calculating the arc length of the stroke by Gaussian quadrature, then
establish adaptable subdivision table of the stroke, the sampling points are the points on
the stroke which have same interval. The distance between two sampling points is the
width of non-planar mesh.
In short, the structure of our system is that there is a scene including serials of objects
and a camera, and an object includes several canvases which also include many strokes,
the camera and every object can be set key frames animation (figure 1).
4 Stroke Model
One of the significant differences between other system and ours is that our system
defines a kind of stroke with sketching style, which contributes to constructing artistic
work. In this section, I describe how to generate the stroke model that I have used.
In our system, the generation of our stroke is similar to the method of Stroke-Based
Rendering [1]. A basic brush stroke is defined by stroke curve, stroke mesh, stroke
color, stroke texture. The curve is also an endpoint-interpolating cubic B-Spline
defined by a set of control points. Curve points can be computed by recursive
606 M. Wang et al.
subdivision. But the difference is that our stroke must be drawn on canvas; it means that
each stroke needs to be converted to 3D space, so the first step is to calculate the points
that intersect on current canvas, and these points are the control points of a stroke. And
then, I bound to convert a stroke into a stroke mesh after all the control points have been
inputted. The basic technique for generating stroke mesh is to tessellate the stroke into a
triangle strip.
Considering that a dense list of control points Pi and a brush thickness R should be
moderate, we can get the geometrical topology of the stroke by the following steps
(Figure 2): curve tangents should be computed at each control point firstly and then is
the normal directions of curve. An adequate approximation to the tangent for an interior
stroke point Pi is given by Vi = Vi +1 − Vi −1 . The first and the last tangents are V0 = P1 − P0
and Vn −1 = Pn −1 − Pn − 2 . The next step is to calculate the points on the boundary of the
stroke, according to the distance R , these points will offset by R along the normal
direction. So the geometrical topology of the stroke can be generated as the Figure 2
showing above.
When the topological structure has been finished, it means that a stroke has become
a triangular mesh, so there are two ways to show a stroke, one is to fill the mesh with a
single color, and the other way is to use texture mapping.
For expressing abundant stroke with sketching style, our system supports to texture
mapping, TGA texture with alpha value can be considered as stroke texture.
The number of points on stroke mesh is dramatically large generated by the method
above, it can make stroke look smooth. But when the objects are stored in a file, we
need to record the stroke mesh and texture, the size of file will be large by this way; or
if the object is far away from the camera, there is no need to give a very precise
expression of the shape of the stroke, approximate expression is enough, thus, the
storage of stroke should be optimized, namely, the number of points should be
decreased in order to economize space of storage.
Our method is when a stroke finished, finding the feature points on the stroke firstly,
our arithmetic adopt a method which makes the spatial curve parameterized by arc
3D Freehand Canvas 607
length, calculates the adaptive subdivision table by given a desired accuracy, the spatial
coordinates of these points calculated by the adaptive subdivision table are the feature
points, every two feature points has a bounding box, namely feature bounding box. Then,
getting the bounding box of the whole stroke, the stroke in its bounding box will be
converted to a new texture, then get the texture coordinates of every bounding box on the
texture, the alpha value of the texture is equal to 0, if on which there is no stroke,
otherwise, the alpha value is equal to 1. When a stroke is saved, only the feature bounding
boxes of the strokes and a texture are stored. (Figure 3) In this way, not only the storage
space is optimized, but also efficiency is improved.
Fig. 3. When given different desired accuracy, the key bounding boxes will change, the bigger
the desired accuracy is, the more the number of the feature bounding boxes are
In OpenGL environment, because of the error which is caused by float computing, the
cross strokes on the same canvas presents a “sieve” in the cross area (Figure 4(a)). The
reason is that users ideally consider that strokes on the same canvas has the same
z-depth value while they are differ at random, which is changing in real time with the
different view.
Our solution to this problem is to mark stroke with an integer orderly when it is
drawn on the canvas, then add a tiny value to the z-depth value of the stroke with the
bigger mark value when it is crossed with another one. (Figure 4(b))
Fig. 4. (left) a “sieve” in the cross area because of the error which is caused by float computing.
(right) “Sieve” disappears by adding a tiny value to the z-depth value of the stroke.
608 M. Wang et al.
To judge whether the two strokes are crossed or not, it is important to find the feature
point in the curves. After computing the feature points by using the adaptive
subdivision table, the curve could be described approximately with the polygonal lines
that consist of the feature points orderly. Then test the cross among the polygonal lines
instead of the mesh stroke.
The amounts of the textures depend on the curve length. The length of every stroke
curve is computed after the operation of adding control points works, which is the sum
of the distance of every two sequential control points. The quotient of the curve length
divided by the width of a texture is defined as the amount of the textures, which is used
in texture mapping.
According to the texture mapping method of OpenGL, texture coordinates are
described as the value that starts at 0 and ends at 1 for one texture. The number above 1
will be considered to use another same texture. So the texture mapping of a stroke can
be generated by given the amount of the textures, the texture will be mapped to stroke
mesh one by one, the quality and size of texture can be ensured by this way, and the
texture will not be stretched.
At present, a stroke composes of stroke mesh and texture mapping, in the future, we
will adopt procedural texture, it will generate the beginning and the end of a stroke
automatically, in addition, procedural texture will be affected by the pressure and speed
in the process of drawing, and the direction of the brush, the stroke with sketching style
generated by this way, will be better than it by stroke mesh.
5 Animation
draw a path and bind objects to it, the path generated is by getting the key position of
the object in information of each key frames, according to the list of the key position of
object, when all key frames has finished setting, the path will be generated
automatically as a form of Catmull-Rom Spline curve. By using this method, it is
convenient to draw a path in 3D space, without any binding operations.
After generating a path curve, the speed curve can be generated automatically too.
Each object has its own motion, and each motion has one path and one speed curve. In
our system, speed curve adopts line curve mode and Multi-Hermite curve mode. Speed
curve, actually, is a distance-time curve. Both distance and time can be obtained by the
position of the object in 3D space and specifying key frames. So the most important
thing of generating a speed curve is to calculate tangents of these control points. Our
method is to calculate tangents based on control points given. The tangents of two
endpoints are calculated firstly, the directions of them are along the direction of joining
of the two endpoints and in opposite direction. And then, the tangents of middle points
can be obtained by the previous point and posterior point of this one, so find these two
points which the mid-one should be insert, and the direction of joining of them is the
tangent of the mid-points.
6 Experimental Results
Our system can create model on a standard PC at interactive rates, user can draw
objects and set object and camera animation. User can also change the sketched style of
stroke by selecting different stroke files.
In our system, we adopt a method of drawing the outline of object to represent an
object in 3D space, and the process of drawing is a kind of mode of “what you see is
what you get”. The system can simulate the sketching style of artists, describe object
by using strokes with different kinds of sketching style, and the whole scene can be
rendered in real-time. The system can generate not only the doodle effect like
drawing on 2D canvas, but also the shielding effect between objects. Thus, our
system has significant advantages compared with 2D and 3D software: compared
with 2D system, viewpoint in our system will be changed and the shielding
relationship between objects will be also changed following the viewpoint, figure 5
shows the shielding and perspective effect of different objects from different
viewpoints; compared with 3D system, our system can describe objects with
sketching style, which is nearly impossible in 3D software, an object composes of just
some strokes without modeling, the system is much freer than 3D software, and it is
in line with the habits of artists, has dramatic performance of art, figure 5(b) is
different from the others, the sketching style of strokes in these three pictures looks
like crayon, and the others simulate the style of watercolor painting, these pictures
represent different drawing styles. The following ones (Figure 5) show single objects
with changing viewpoint, (Figure 5(a)) and the whole scene with fog around it which
composes of several objects.(Figure 5(b),(c)).
610 M. Wang et al.
(a)
(b)
(c)
Fig. 5. Result pictures. (a) Showing the single objects from different view point. (b) These three
pictures are a little more complicated than the ones in (a), which composes of multi-object, and
the drawing style is crayon while others are watercolour. (c) Showing a whole scene which
describes nightfall of magic world with light yellow fog around it, it is more complex than others,
and includes the movement of people and viewpoint.
3D Freehand Canvas 611
In this paper we present an approach for establishing strokes with sketching style,
drawing objects and generating animation. The system is able to create rather
complicated stroke mesh with almost only sketch input, which can be used to generate
sketching style by specifying stroke texture and appending texture to mesh. Apart from
defining stroke mesh, our system also allows using stroke texture with Alpha value, and
does storage optimization of stroke. The system runs at interactive rates on a standard
PC. The whole system is adapted well to non-expert users.
There would be also lots of works to do to enrich the functions of this software.
Firstly, we provide some sketching styles of stroke to users, but the variety of stroke
style is still not rich, in the further study, we intend to map multi-texture to stroke mesh,
and allow users define stroke style by themselves. Secondly, strokes can be edited by
user, such as movement, rotation, copy, paste, but can not transform, in the further
work, the sharp of stroke will be transformed by pushing and pulling. Third, because of
the limit of the drawing styles, just some transparent texture created by other software
or some common format could be used in the storyboard; extending brushwork library
will be an effective supplement for the 3D Freehand Sketching system. Fourth, the
system is able to express a scene containing multi-object rendering with depth test, but
objects at distant position is too clear to have an effect on near ones, thus, an effect of
distant objects disappearing gradually is necessary for users, which makes near ones
more clearly.
References
1. Hertzmann, A.: Stroke-Based Rendering. In: SIGGRAPH 2002, pp. 3.1-3.31 (2002)
2. Butterworth, J., Davidson, A., Hench, S., Olano, M.T.: 3DM: a three dimensional modeler
using a head-mounted display. In: SI3D 1992: Proceedings of the 1992 symposium on
Interactive 3D graphics, pp. 135–138. ACM Press, New York (1992)
3. Cohen, J.M., Markosian, L., Zeleznik, R.C., Hughes, J.F., Barzel, R.: An interface for
sketching 3D curves. In: SI3D 1999: Proceedings of the 1999 symposium on Interactive 3D
graphics, pp. 17–21. ACM Press, New York (1999)
4. Deering, M.F.: HoloSketch: a virtual reality sketching/animation tool. ACM Trans.
Comput.-Hum. Interact. 2(3), 220–238 (1995)
5. Diehl, H., Müller, F., Lindemann, U.: From raw 3D-sketches to exact CAD product models
—concept for an assistant-system. In: Hughes, J.F., Jorge, J.A. (eds.) Eurographics
Workshop on Sketch-Based Interfaces and Modeling. Eurographics (August 2004)
6. Kallio, K.: 3D6B Editor: Projective 3D Sketching with Line-Based Rendering. In: Igarashi,
T., Jorge, J.A. (eds.) Eurographics Workshop on Sketch-Based Interfaces and Modeling
(2005)
7. Lim, C.-K.: An insight into the freedom of using a pen: Pen-based system and
pen-and-paper. In: Proc. 6th Asian Design International Conference (October 2003)
8. Sachs, E., Roberts, A., Stoops, D.: 3-Draw: A tool for designing 3D shapes. IEEE Comput.
Graph. Appl. 11(6), 18–26 (1991)
612 M. Wang et al.
9. Sharlin, E., Sousa, M.C.: Drawing in space using the 3D tractus. In: 2nd IEEE Workshop on
New Directions in 3D User Interfaces (IEEE VR 2005) (March 2005)
10. Schumann, J., Strothotte, T., Laser, S., Raab, A.: Assessing the effect of non-photorealistic
rendered images in CAD. In: CHI 1996: Proceedings of the SIGCHI conference on Human
factors in computing systems, pp. 35–41. ACM Press, New York (1996)
11. Tolba, O., Dorsey, J., Mcmillan, L.: Sketching with projective 2D strokes. In: UIST 1999:
Proceedings of the 12th annual ACM symposium on User interface software and
technology, pp. 149–157. ACM Press, New York (1999)
Sparse Key Points Controlled Animation for
Individual Face Model
1 Introduction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 613–618, 2008.
© Springer-Verlag Berlin Heidelberg 2008
614 J. Yao, Y. Wang, and B. Ding
Also the regulated model can be directly flattened. After remapping of key points’
motion vector, we interpolate vertexes’ motion vector with barycentric coordinate. All
the process needs less than 40 key points.
The rest of the paper is organized as follows. Section 2 describes how to regulate
scanned face models, especially about normal projection and the accelerating with
octree. Section 3 describes how to remapping animation from motion capture data to
the regular face model. Section 4 includes postprocess about mouth division. Then
Section 5 shows some examples and draws a conclusion.
In PCA based method of individual face generation, the regulation of scanned face
model is an inevitable step. The regulation means to remesh all the face models to the
same topology and make sure any one vertex has the same anatomy meaning for every
model. Also the uniformity of vertexes should be taken into account, since sparse mesh
will reduce the quality of texture. We use RBF and normal projection to regulate
scanned face models(Fig.1). Also, the vertexes’ color of regular model is retrieved
during normal projection process.
RBF (Radial Basis Function) is a supervised neural network and a powerful tool for
geometry smoothing and deformation [3]. RBF could perform deformation between
any two meshes with correspondence key points. Its performance is determined by
position and amount of the key points. Fig.1(c) shows the result of RBF deformation.
The RBF equation is
n
f ( pi ) = ∑ ω j ⋅ h j ( pi ) (1)
j =1
2
h j ( pi ) = Pi − Pj + s j2 (2)
C (ω ) = eT e + λ ⋅ ω T ω (3)
where e is the error vector between the training input and output coordinates.
After RBF deformation, the shape of source grid mesh is roughly similar to the
target model but the detailed difference is very large. A projection process must be
applied to make the source mesh ”close-fitting” to the target model(Fig.1(d)). J. Noh[2]
use cylindrical projection to improve the RBF deformation, but it could have large
distortion if the target triangle face is close to horizontal. We use the normal vector of
vertex as the direction of projection. This makes the projection result more uniform.
The comparation can be seen in Fig.2.
The projection process needs to traverse all the vertexes of source mesh and all the
triangle faces of target model, so this process is time-consuming. We introduce octree
to accelerate it(Fig.3). For each vertex, the coordinate and normal vector define a line in
3D space (green line). We compute to determine the octree nodes that intersect with the
line, then only the triangle faces (red region) in intersected nodes will be taken into
account. This will commonly spare more than 75 percent computation time. All the
intersection points will be computed and the nearest to the vertex is chosen as
projection point.
The scanned models are acquired with 3D scanner and contain color information.
Consequently, the regular face model also needs color to render. Vertex of grid mesh is
now located on one triangle face of the scanned model. Thus we can get the barycentric
coordinate and use it to compute vertex color(Eq.4).
~ ~ ~ ~
C g = α ⋅ Cs 0 + β ⋅ Cs1 + γ ⋅Cs 2 (4)
~
where C g = ( Rg , G~ g , B~g ) is
~
the color of grid mesh vertex, (α , β , λ ) is barycentric
coordinate, and C g 0 , Cg1 , Cg 2 is color of the three triangle vertexes respectively.
With all the process above, scanned face models have been regulated to a same
regular topology. Then, we can construct a matrix with vertexes’ coordinate as the
616 J. Yao, Y. Wang, and B. Ding
3 Animation Interpolation
The source animation data is obtained from Motion Capture with only 39 markers. For
the shapes of performer’s face and grid model are different, so direct transfer of the
motion data between models is inappropriate and may induce errors such as
overlapping. The direction and magnitude of motion vectors need to be modified before
remapping. We employ local bounding box[2] to do this job. Bounding box is a
directional cuboid contains a vertex and all its neighbor vertexes. As shown in Fig.5,
V and V ' are key point of source and target models respectively, m is motion vector of
V . The change of direction and magnitude from source to target bounding box
determines the transform matrix R , thus m' = R ⋅ m is the motion vector of V ' .
Our goal is realistic facial animation with sparse key points, which require an unique
interpolation method. Parameterization is the optimal method to achieve unique
correspondence and interpolation. Since all the face models have been regulated, which
means the model can be directly flattened similarly as plane parameterization. When a
vertex is flattened onto a plan, it will be located in one triangle face of a predefined key
point mesh. Similar to the color retrieval, motion vector of the vertex is interpolated by
the barycentric coordinate and motion vector of three key points. The interpolation
equation is the same as Eq.4.
For bounding box is local property of the key mesh and varies with the topology, so
it’s inevitable for errors in motion vector remapping. We smooth motion vectors using
Sparse Key Points Controlled Animation for Individual Face Model 617
ur ur
Eq.5, in m j which is smoothed motion vector of vertex i , m j is the motion vector of its
1
neighbor vertex j , and ωij = is reciprocal of the distance between vertex i
pi − p j
and vertex j . Fig.6 shows an example, the inflexion in underlip is corrected.
ur ' 1 ur '
mi = ⋅ ∑ ωij ⋅ m j
∑ω
(5)
ij j
j
4 Mouth Process
To enable the mouth to open, the mesh must be cut off along the mouth contact line and
vertexes should have motion vectors in different directions. Otherwise, vertexes in
different lips could have motion in same direction and thus the mouth cannot open.
Since all the vertexes of lip contact line lie in one row of the grid mesh, we only need to
duplicate these vertexes and assign them to triangle faces in different lips to achieve lip
division.
Vertexes in lip region are affected by key points in both sides, which requires
distinguishing upper and lower lip. We use Dijkstra’s algorithm[4] to compute the
shortest path from vertex in lip region and midpoint of upper lip. If the path intersects
with lip contact line, the vertex should be marked as lower lip. Otherwise, no
intersection means the upper lip vertex. Thus we can restrict the influence of lip key
points only in vertexes of its own side. Fig.7 shows a comparison of the result with and
without lip division. Yellow point is a key point and vertexes influenced by it are
represented by red points, the transparency represents influence scale.
618 J. Yao, Y. Wang, and B. Ding
References
[1] Blanz, V., Vetter, T.: A morphable model for the synthesis of 3D faces. In: Proceedings of
the 26th Annual Conference on Computer Graphics and interactive Techniques,
International Conference on Computer Graphics and Interactive Techniques, pp. 187–194.
ACM Press/Addison-Wesley Publishing Co., New York (1999)
[2] Noh, J., Neumann, U.: Expression cloning. In Proceedings of the 28th Annual Conference
on Computer Graphics and interactive Techniques. In: SIGGRAPH 2001, pp. 277–288.
ACM, New York (2001)
[3] Pighin, F., Hecker, J., Lischinski, D., Szeliski, R., Salesin, D.H.: Synthesizing realistic
facial expressions from photographs. In: Proceedings of the 25th Annual Conference on
Computer Graphics and interactive Techniques, SIGGRAPH 1998, pp. 75–84. ACM, New
York (1998)
[4] Johnson, D.B., Johnson, A.: A Note on Dijkstra’s Shortest Path Algorithm. J. ACM 20(3),
385–388 (1973)
[5] Tutte, W.T.: Convex representations of graphs. Proc. London Math. Soc. (10), 304–320
(1960)
[6] Sheffer, A., Praun, E., Rose, K.: Mesh parameterization methods and their applications.
Found. Trends. Comput. Graph. Vis. 2(2), 105–171 (2006)
[7] MathWorld, Wolfram Research, http://mathworld.wolfram.com
[8] Sumner, R.W., Popovic, J.: Deformation transfer for triangle meshes. In: SIGGRAPH 2004,
pp. 399–405. ACM, New York (2004)
[9] Langer, T., Belyaev, A., Seidel, H.-P.: Spherical Barycentric Coordinates. In: Proceedings
of the Fourth Eurographics Symposium on Geometry Processing (2006)
Networked Virtual Marionette Theater
1
Kwansei Gakuin University, School of Science and Technology
2-1 Gakuen, Sanda, 669-1337 Japan
{aaz61232,miyazaki,nakatsu}@kwansei.ac.jp
2
National University of Singapore, Interactive & Digital Media Institute
Blk E3A #02-04, 7 Engineering Drive 1, Singapore 117574
idmdir@nus.edu.sg
Abstract. This paper describes a system that allows users to control virtual
marionette characters based on computer graphics (CG marionette characters)
with their hand and finger movements and thus perform a marionette theatrical
play. The system consists of several subsystems, and each subsystem consists of
a web camera and a PC. It can recognize a hand gesture of its user and
transform it into a gesture of a CG marionette character. These subsystems are
connected through the Internet, so they can exchange the information of the CG
marionette character’s movements at each subsystem and display the
movements of all characters throughout the entire system. Accordingly,
multiple users can join the networked virtual marionette theater and enjoy the
marionette play together.
1 Introduction
The culture of controlling puppets with the hands to perform theatrical play has been
common throughout the world from ancient times. In Japan, there is a type of puppet
theater called Bunraku, which arose about three hundred years ago [1][2]. In Europe,
too, various kinds of puppet play have been performed and enjoyed. The puppet play
using a puppet called a “marionette” has been the most popular variety [3]. Marionette
play and puppets have become very popular in recent years, largely due to the movie
called “Strings [4]” (Fig. 1). This paper describes a networked virtual marionette theater
that is basically a distributed system consisting of several subsystems connected through
the Internet. Each subsystem can recognize the hand and finger gestures of the person in
front of its web camera and then transform them into the motions of a marionette
character based on computer graphics (CG marionettes). Each subsystem exchanges the
information of actions performed by its marionette character with such information from
the other subsystems. The display of each subsystem shows a virtual scene where
multiple marionette characters, each controlled by a different user, interact. Thus
multiple users, even if they are in separate locations, can gather in a virtual marionette
theater and perform a theatrical marionette play.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 619–627, 2008.
© Springer-Verlag Berlin Heidelberg 2008
620 D. Ninomiya, K. Miyazaki, and R. Nakatsu
2 Related Works
Technologies based on three-dimensional computer graphics have made tremendous
progress in recent years. We can see photographically real CG objects and CG
characters in movies and games. Furthermore, the technologies based on CG
animation have also progressed rapidly. Animations of fluid [5] and the destruction of
objects [6] have been studied. Moreover, the movements of a crowd based on an
artificial-intelligence approach [7] and movements of humans based on inverse
kinematics [8] have been proposed. Motion capture systems have been widely used
for the control of CG characters [9]. Although the realization of human-like motions
of CG characters has been eagerly pursued, the realization of marionette-like motions
has seldom been studied. Since the movements of marionette characters are unique
and have been loved by people throughout history, it is worth studying a system by
which non-experts of marionettes can easily manipulate their movements and
generate marionette-like behaviors using CG characters.
3 System Concept
The following elements typically compose a marionette theater.
(1) Puppets called “marionettes”
(2) Speech of each puppet
(3) Scene settings
(4) Music
In a large performance, various kinds of marionette puppets appear and the scene
settings are changed frequently, depending on the story’s plot, and even a live
orchestra is sometimes used to generate music. Therefore, even if people wanted to
enjoy manipulating marionette puppets and creating theatrical play, it could be very
difficult. On the other hand, if we introduced virtual marionette characters based on
computer graphics instead of using real marionettes with physical bodies, it would
Networked Virtual Marionette Theater 621
become significantly easier for users to generate and change most of the above
elements of marionettes, speech, backgrounds, and music. In addition, by developing
a networked virtual marionette theater, multiple users, manipulating their own
marionette characters, can gather in a virtual theater and let their virtual puppets
interact with other puppets, thus creating the performance of a virtual theatrical play.
4 System Structure
4.1 Overview
The entire system is made from a group of subsystems connected through a network.
The structure of the whole system is shown in Fig. 2, and the structure of each
subsystem is illustrated in Fig. 3. Each subsystem consists of a PC and a web camera
The posture of a user’s hand is captured by the web camera, and then hand-gesture
recognition is carried out. Then the recognition result of a hand posture is reflected in
the gestures of a CG marionette character.
In this section, a real-time hand-gesture recognition method is described for use in the
recognition of a user’s hand gesture for each subsystem [5]. There have been several
research efforts on the real-time recognition of hand gestures [6][7]. Most of them use
rather complicated systems such as multiple cameras. On the other hand, we tried to
develop a simpler system using a single web camera. The recognition process consists
of the following sub-processes.
Finger-bending
angle
Histogram
reflected directly in the length of each string. In this way, the angle of each joint of
the marionette, corresponding to p1, p2, p3, p4, p5, p6, p7, and p8, is determined.
Therefore, by bending each of the five fingers appropriately, a user can control the
motion and gestures of a virtual CG marionette.
We are planning a system that allows us to easily change scenes as well as characters,
so we have developed various kinds of backgrounds and characters based on
computer graphics. We are trying to develop an “Interactive Folktale System [8]” that
offer users the ability to generate Japanese folktales as animation and to enjoy the
interactions with creators of other characters in the system. Therefore, we have
prepared various kinds of backgrounds and characters for our virtual marionette
system. Figure 7 shows one of the marionette character in three different
backgrounds.
The virtual marionette system we have developed as a first prototype toward the
networked virtual marionette system consists of a hand-gesture recognition & control
unit and an animation generation unit. This prototype system would work as a sub-
system in the whole distributed system. These subsystems are connected using a
network environment construction library called DirectPlay that is one of the DirectX
libraries. In this case, instead of the client-server model, subsystems are connected
based on the peer-to-peer model. In most of marionette plays the number of
marionettes that appear in the play is less than ten. This means that this peer-to-peer
network model would be enough to construct a networked marionette theater.
As a first step, we have constructed a networked system consisting of two sub-
systems. The construction of the whole system is illustrated in Fig. 8. In each
subsystem, the recognition results of another subsystem is shared. Furthermore, all of
the CG characters and backgrounds are shared among these subsystems. Using these
Networked Virtual Marionette Theater 625
recognition results as well as the CG characters and backgrounds, each subsystem can
simultaneously create the same scene where multiple CG characters, each of which is
controlled by its own subsystem, appear and behave in the same way.
(3) Did you feel that your hand gestures were closely reflected in the movements of a
virtual marionette?
Definitely Yes (0), Comparatively Yes (15), Neutral (3), Comparatively No (1),
Definitely No (1)
From the first question, it is clear that 80% of the subjects said that there is some
unique aspect in the movement of the virtual marionette. This means that the authors
626 D. Ninomiya, K. Miyazaki, and R. Nakatsu
6 Conclusions
In this paper, we proposed a system in which users can easily manipulate virtual
marionette characters with their hand gestures. For the recognition of hand gestures,
simple real-time hand-gesture recognition was realized by using histogram
information of an extracted hand area. The recognition result is reflected in the
movement of the marionette character by connecting each finger movement to a
particular part of the virtual marionette by a virtual string. Furthermore, the concept of
networked marionette theater was proposed in which several subsystems are
connected by a network. Here, multiple users can perform theatrical marionette play
by manipulating their own marionette characters. Finally, we carried out an evaluation
test to assess the feasibility of a subsystem. By using twenty subjects and letting them
manipulate both a physical marionette as well as a virtual one, we obtained evaluation
results indicating that by using this virtual marionette system, even a non-expert of
marionette manipulation can have the feeling of manipulating marionettes and thus
can participate in a theatrical marionette performance.
For our further work, we need to improve the recognition accuracy of the hand-
gesture recognition. Moreover, we need to develop adequate contents to refine the
entire networked virtual marionette theater, and we also need to carry out an
evaluation of the whole system by letting people use the system.
References
1. Keene, D.: No and Bunraku. Columbia University Press (1990)
2. http://www.lares.dti.ne.jp/bunraku/index.html
3. Currell, D.: Making and Manipulating Marionettes. The Crowood Press Ltd. (2004)
4. http://www.futuremovies.co.uk/review.asp?ID=319
5. Stam, J., Fiume, E.: Depicting Fire and Other Gaseous Phenomena Using Diffusion
Process. In: Proceedings of SIGGRAPH 1995 (1995)
6. O’ Brien, J.F., Hodgins, J.K.: Graphical modeling and animation of brittle fracture. In:
Proceedings of SIGGRAPH 1999 (1999)
7. Courty, N.: Fast Crowd. In: ACM Siggraph/Eurographics SCA 2004 (2004)
8. Boulic, N., Thalmann, M., Thalmann, D.: A Global Human Walking Model With Real-
Time Kinematic Personification. The Visual computer, 344–358 (1990)
9. Lee, J., Lee, K.H.: Precomputing avatar behavior from human motion data. Graphical
Models 68(2), 158–174 (2004)
Networked Virtual Marionette Theater 627
10. Ninomiya, D., Miyazaki, K., Nakatsu, R.: Study on the CG Marionette Control Based on
the Hand Gesture Recognition. Annual Meeting of Game Society of Japan (in Japanese)
(2006)
11. Ng, C.W.: Real-time gesture recognition system and application. Image and Vision
Computing, 20 (2002)
12. Utsumi, A., Ohya, J., Nakatsu, R.: Multiple-camera-based Multiple-hand-gesture-tracking
(in Japanese). Transaction of Information Processing Society of Japan 40(8), 3143–3154
(1999)
13. Miyazaki, K., Nagai, Y., Wama, T., Nakatsu, R.: Concept and Construction of an
Interactive Folktale System. In: Ma, L., Rauterberg, M., Nakatsu, R. (eds.) ICEC 2007.
LNCS, vol. 4740, pp. 162–170. Springer, Heidelberg (2007)
Tour into Virtual Environment in the Style of Pencil
Drawing
1 Introduction
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 628–635, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Tour into Virtual Environment in the Style of Pencil Drawing 629
For these reasons, more and more researchers endeavor to depict virtual
environments by NPR means. Allison W. Klein etc. have firstly described a very
classic system for non-photorealistic rendering (NPR) of virtual environments [1].
Liviu Coconu etc. have also presents a NPR rendering pipeline that supports pen-
and ink illustration for complex landscape scenes [2]. The feature of his work is that
all NPR algorithms are integrated with photorealistic rendering, allowing seamless
transition and combination between a variety of photorealistic and non-photorealistic
drawing styles.
Thomas Luft etc. have presented algorithms that allow for real-time rendering of
3D-scenes with a watercolor painting appearance [3]. Their approach provides an
appropriate simplification of the visual complexity, and imitates characteristic natural
effects of watercolor. To enhance the efficient of the algorithm, they use image space
processing methods rather than an exact simulation.
Hyunjun Lee etc. [4] have presented some real-time technique for rendering 3D
meshes in the pencil drawing style. The shortage of the system is that the painting
speed is limited by the amounts of polygon surfaces on the 3D model. Therefore, the
computational complexity is very high when rendering the large scale and complex
3D scene.
In the research, we and Hyung W. Kang [5] have individually find that there are
two disadvantages of these techniques mentioned above: (1) It has to build a coarse
3D model of the realistic 3D scene, so it is not fit to the real-time rendering of the
complex scene. (2) Although many researchers have designed a various artistic style
on NPR algorithm to render 3D virtual scenes, they ignored pencil drawing is an
important form of pictorial representation and an effective way to convey lighting,
directions and texture properties.
In this paper, by using a hybrid NPR/TIP approach which we designed previously
[6], we present a new system for tour into a pencil drawing style virtual environment.
By rendering the real photos or images into pencil drawing style, we can reconstruct
the 3D virtual environment taking on artistic imago, at the same time we can easily
transfer more concise and effective information to the user than any other artistic
expressions practices.
2 System Overview
In this section we will simply introduce the basic architecture of our system. At a high
level, our system proceeds in two steps. First, during preprocessing, the system has to
do two important works (1) Reconstructs the foreground model and background
model of the cubic panorama. (2) By using our image-based pencil drawing algorithm
(This step is described in section 3), we can fast process the input 2D panorama
images into pencil drawing styles. Second, during real-time rendering, the system has
to do two important works (1) Re-renders the foreground model and background
model in the style of pencil drawing (2) Interactively tour into the pencil drawing
style cubic panorama based on the 3D model we created.
630 Y. Zhao, D.-e. Xie, and D. Xu
By observing and analyzing the real pencil texture, we simply suppose that: (1)
graphite marks present stochastic distribution according to the coarseness of papers;
(2) graphite marks stretch along the stroke tracks; (3) graphite marks on perpendicular
direction of a stroke present obviously black-white staggered distribution.
Considering the above supposes, we create a mathematic model for pencil filter.
Assume that the stroke length is len, the stroke direction is θ and the stroke width is
2D. If we know the stroke length and the stroke orientation, we can easily calculate
the template size by ( ⎡len * sin θ ⎤ × ⎡len * cos θ ⎤) .
The next problem is how to decide the value of each element in the pencil filter. As
shown in Figure 2, firstly, calculate the distance d from each point P to the central
axis l of a stroke. Then calculate the distance r from the point P to the center O of the
stroke. The value of each element in the pencil filter lies on the relation between d and
D, and also the relation between r and len/2.
Here, we take the upper right quarter of the template as an example (Figure 2(a)).
Obviously, only three kinds of points are presented in the template. Points in the
green area (e.g. P in Figure 2(b)) satisfy the conditions that r is less than len/2 and d is
less than D, so we choose D-d as their values. Points in the gray area (e.g. Q in Figure
3(c)) satisfy the condition that d is large than D, which means the stroke can’t covered
this area, so their values are set to be zero. Points in the blue area (e.g. R in Figure
3(d)), satisfy the conditions that r is large than len/2 and d is less than D. Blue area is
near to the stroke end, and the graphite marks is thin there, so the value in the
Tour into Virtual Environment in the Style of Pencil Drawing 631
template is less than D-d and none zero. We calculate the distance dx from point R to
the end of line l, and then we choose D-dx as the value of this area. In this way, it
decreases the value of the stroke end effectively, and the decrement is in proportion to
the distance r. When D-dx is less than zero, the value should be set to zero.
To make sure the pencil texture has stochastic distribution, we generate the black
noise image from the reference image. Our method for generating the black noise
image is similar to Mao’s approach [7]. We use the tone of the input image to guide
the distribution of the black noise. Let Iinput be the intensity of a pixel in the input
image, P is a floating-point number generated with a pseudo-random function, and
then the intensity Iinput of the corresponding pixel in the noise image is decided in
the following way:
⎧
⎪ 2 5 5, if P ≤T P ∈ [0 .0,1 .0 ] ⎛I ⎞
I n o ise = ⎨ , T = k ⋅ ⎜ in p u t ⎟ , k ∈ ( 0 .0,1 .0 ] (1)
⎪ 0, ⎝ 255 ⎠
⎩ o th erw ise
K is a coefficient for controlling the density of the black noise. In this way we can
ensure the pencil drawings have the stochastic distribution character, and also ensure
the density of the black noise correspond to the intensity of the input image.
It is also an important step to extrud the outlines in pencil drawing. Gradient operators
are commonly used in digital image processing to extract edges of an image.
Considering Kirsch operator has the bigger weighted factors, we prefer to choose it to
extract the contour lines in this paper.
Here, μ is a coefficient for controlling the weight value in the filter. One can adjust
the value of μ interactively. Figure 3 shows the different contour line maps with the
different value of μ . Generally, if an image has more details, the value of should be
smaller. A smaller can prevent the contour line conglutination. On the contrary, if an
image has little details, a larger should be set in order to make sure the consistency of
the contour lines.
In the process of pencil drawing, artists will choose strokes with different direction
when painting a region. Therefore, our system needs to be done by the following two
steps: (1) Segment the input image into different regions. (2) Set different stroke
direction to the different regions. Traditional image segmentation techniques can not
meet our requirements, so we have designed an interactive method for image
segmentation. In our system, users can interactively segment the input image into
different regions according to their creation needs, at the same time each of those
regions will be then set to the corresponding stroke direction. Our approach is not
only more flexible than Mao’s approach [7], but also more in line with the actual
process of pencil drawing. Figure 4(a) shows the input image. Figure 4(b) shows the
result of the different segment regions, Figure 4(c) shows the stroke direction of the
different regions.
According to the imagination and understanding to the input image, user can easily
distinguish foreground from background. Thus the whole 3D scene model
reconstructed mainly consists of background model and foreground model.
To background model, the input image is divided into two disjoint regions by the
vanishing line. The region below the vanishing line in the image corresponds to
ground plane, and that above the vanishing line corresponds to back plane. Since
vanishing line contains all vanishing points formed by all parallel lines on a plane,
back plane can be thought of as a plane of infinite distance and the points on it can be
called the ideal points (Figure 5).To foreground model, a foreground object specified
in the image is modeled as a 3D polygon called foreground model [9]. Suppose that
the polygons stand perpendicular to the ground plane. The coordinates of its vertices
are then computed by finding the intersection points between the ground plane and the
ray starting from the viewpoint and passing through the corresponding vertices in the
image plane. As proposed by H. Kang, a foreground object can have a hierarchical
structure in a more complex environment. That is, another foreground object can be
attached on a foreground object standing on the ground plane to form a hierarchical
structure.
For cubic panorama, there is no important content on the top and the bottom sides in
general. Thus the system just simplifies to model the four sides using vanishing line
based TIP techniques. Supposed that the center of cube (viewpoint) is positioned at
the origin. Figure 6 illustrates the local model of one side of cube. We make local
model for the four sides of cube respectively. Since the input image is panorama, the
vanishing line on each of the four sides will have the same height above the bottom
side. So the four local models will put together to form a global scene model. Then,
the system just needs to project the top and the bottom sides of cubic panorama on the
top and the bottom of the global model respectively. Finally, we get a close
hexahedron scene model.
Together with the scene model, the images called background image and
foreground mask are generated by segmenting the foreground objects from the input
image. The background image is used as a texture, which is to be mapped onto the
background model. Foreground mask is used to distinguish the exact portion of the
634 Y. Zhao, D.-e. Xie, and D. Xu
foreground object from the background. After performing the above all steps, we can
render the scene model by changing the parameters of camera.
5 Experiment Result
An image based pencil drawing system has been developed with the Matlab7.04.
When the input image is specified, it can generate the pencil drawing image
automatically. Users are allowed to specify parameters interactively. These
parameters control the stroke orientation, length, density of the noise and the
coefficients of the Kirsch operator. Compared with the LIC method, our method saves
lots of time. We average use 10 seconds to process an image (here the image size is
1024*768) into pencil drawing style. The main reason is that we needn’t to calculate
the visualization vector field of the input image, and needn’t to do the hundreds of
iterations. We only need to do the convolution once for each pixel. In addition, we use
Microsoft Visual C++ 6.0 and the OpenGL to develop our real-time interactive 3D
virtual scene rendering system, the rendering speed of our rendering system can easily
get to 25 fps. Through the combined use of the two systems we designed, users can
tour into the virtual environment in the style of pencil drawing in real-time (Figure 7).
6 Conclusion
In this paper, we have described a system which can provide a simple pencil drawing
scene model, the users can walk through in it freely, obtain enjoyable and real-time
artistic experience. Our system is not only fit to the fast rendering of the small-scale
scene, but also to that of the complex large-scale scene. Furthermore, we have also
designed an image based system to directly process 2D images into pencil drawing
style. According to their design requirements, users can make use of the system to
render the input scene images into pencil drawing style. Through the combined use of
the two systems we designed, this paper presents a relatively complete solution.
References
1. Klein, A.W., Li, W., Kazhdan, M.M., Corra, W.T., Finkelstein, A., Funkhouser, T.A.: Non-
photorealistic virtual environments. In: ACM SIGGRAPH 2000, pp. 527–534 (2000)
2. Coconu, L., Deussen, O., Hege, H.-C.: Real-time pen-and-ink illustration of landscapes. In:
NPAR 2006, pp. 27–35 (2006)
3. Luft, T., Deussen, O.: Real-Time Watercolor for Animation. J. Comput. Sci. Technol. 21(2),
159–165 (2006)
4. Lee, H., Kwon, S., Lee, S.: Real-time pencil rendering. In: NPAR 2006, pp. 37–45 (2006)
5. Kang, H.: Nonphotorealistic Virtual Environment Navigation from Images. International
Journal of Image and Graphics 5(2), 1–13 (2005)
6. Zhang, Y.-P., Zhao, Y., Shi, J., Xu, D.: Digitization of Culture Heritage Based on Tour into
the Picture. In: Edutainment 2006, pp. 1312–1320 (2006)
7. Mao, X., Nagasaka, Y., Imamiya, A.: Automatic generation of pencil drawing from 2D
images using line integral convolution. In: CAD/Graphics 2001, pp. 240–248 (2001)
8. Cabral, B., Leedom, C.: Imaging Vector Field Using Line Integral Convolution. In:
SIGGRAPH 1993 conference proceeding, pp. 263–270 (1993)
9. Kang, H., Pyo, S.Y., Anjyo, K., Shin, S.Y.: Tour into the Picture using a Vanishing Line
and its Extension to Panoramic Images. Computer Graphics Forum 20(3), 132–141 (2001)
Research and Implementation of Hybrid Tracking
Techniques in Augmented Museum Tour System
1 Introduction
Augmented Reality (AR) is a new research field developed on the basis of Virtual
Reality (VR), which superimposes computer generated virtual information onto the
surrounding real environments to augment user’s perception in real-time and
interactively[1]. Augmented Reality systems have a promising application prospect[2]
in the field of Equipment Maintenance, Entertainment, Medical treatment and
E-learning.
By applying modern AR techniques into conventional museum tour[3-5], named as
Augmented Museum Tour system (AMT), it becomes more effective and interactive
for visitors, helps them absorb profound information and experience a special
multi-channel interactive way, including panorama visual effect and audio approaches.
Although Augmented Museum Tour system may benefit visitors a lot, just like all AR
prototype systems, it also needs to solve tracking problems first to achieve a better
performance.
Tracking is one of important research fields in Augmented Reality (AR), and also in
Augmented Museum Tour (AMT) system; it aims at solving accurate registration
problems between virtual information and real environment and building a stable and
effective AR application system. Former researches on tracking techniques in
Augmented Museum Tour system mainly focus on vision-based technology. Noboru
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 636–643, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Research and Implementation of Hybrid Tracking Techniques in AMT System 637
used for predicting the dynamic position of the tracker; Third, visual assisted correction
is implemented to ameliorate inertial drift and compute the accurate position and
orientation information of the tracker.
Initial positioning will be first performed by recognition with vision tracker after the
initialization of the system. Just like every visual recognition algorithm, markers in
image sequences are first sampled for pre-processing, four to five best sampled markers
are detected and extracted for recognizing, and the position information of these
markers are looked-up in the pre-stored map, which will be used for computing the
initial position and orientation information with pose recovery algorithm.
Visual data are also used to ameliorate inertial drift by using Kalman filter and acquire
the accurate position information in the approximate area provided by inertial tracking
step.
There are four principal coordinate systems in Augmented Museum Tour system, as
illustrated in Figure 1: the world coordinate system W ( xw , yw , zw ) , the camera
coordinate system C ( xc , yc , zc ) , the inertial coordinate system I ( xi , yi , zi ) , and the 2D
image coordinate system U ( xu , yu ) .
(1)
.
where the Ri is a 3*3 rotation matrix, calculated with orientation data from gyroscopes
and the T is a 3*1 translation vector, calculated with position data from accelerometers,
i
characterizing the current orientation and position of camera in world coordinate
system.
where the matrix K :
(2)
contains the intrinsic parameters of the camera, f is the focal length of the camera,
Δx and Δy are the physical size of horizontal and vertical axis in 2D image plane, and
(u 0, v0) ) is the intersection point of camera-center axis with 2D image plane. The
intrinsic parameters are calculated offline.
VisTracker, a 6DOF hybrid tracker, has two standard units[9]. One is hybrid sensor
unit, used for recognizing markers and acquiring inertial information during moving
process; the other is Data Fusion processor, used for computing the position and
orientation information of the tracker in world coordinate system.
Our system takes ARToolkit[10] released by HIT Lab from University of
Washington as our software developing platform, OpenGL 3D Graphics API as virtual
information rendering tool; Hardware platform is made up of three parts: VisTracker,
CCD camera and a computer workstation. VisTracker and CCD camera are fixed
together onto a bracket with their centers calibrated in one point and a 90ºoffset in the
pitch angle.
System design introduces the way we had applied to implement our system.
Fig. 2. World coordinate system. The metric is in centimeter, and z-axis forwards inside the
paper.
(3)
Translation vector:
(4)
5 Conclusion
We tried to apply a hybrid tracking approach which integrates inertial (6-DOF) and
vision-based technologies to promote the tracking performance for Augmented
Museum Tour system, followed by analyzing the tracking problems and requirements
for it. An AR experimental system based on VisTracker, a Vision-Inertial Self tracker
is also designed. Experimental results and analysis demonstrate the system’s
effectiveness and its further application prospect in Augmented Museum Tour
system.
References
[1] Azuma, R.: A Survey of Augmented Reality. J. Presence: Teleoperators and Virtual
Environments 6(4), 355–385 (1997)
[2] Azuma, R., Baillot, Y., Behringer, R., Feiner, S., Julier, S., MacIntyre, B.: Recent
Advances in Augmented Reality. J. IEEE Computer Graphics and Applications, 34–47
(2001)
[3] Vlahakis, V., Ioannidis, N., Karigiannis, J.: ARCHEOGUIDE: Challenges and Solutions
of a Personalized Augmented Reality Guide for Archaeological sites. J. IEEE Computer
Graphics and Applications 22(5), 52–60 (2002)
[4] Koshizuka, N.: Museum Navigation System using Augmented Reality Technologies,
http://www.um.u-tokyo.ac.jp/publish_db/2000dm2k/english/01/
01-16.html
[5] Liarokapis, F., White, M.: Augmented Reality Techniques for Museum Environments. J.
The Mediterranean Journal of Computers and Networks 1(2), 90–96 (2005)
[6] Kang, B.: Tracking Technology for Augmented Reality. Computer Measurement and
Control 14(11), 1431–1434 (2006)
[7] You, S., Neumann, U., Azuma, R.: Hybrid inertial and vision tracking for augmented
reality registration. In: Proc. of IEEE Virtual Reality 1999, pp. 260–267 (1999)
Research and Implementation of Hybrid Tracking Techniques in AMT System 643
[8] Foxlin, E., Naimark, L.: VIS-Tracker: A Wearable Vision-Inertial Self-Tracker. In: Proc.
of IEEE Virtual Reality 2003 (VR 2003), pp. 199–206 (2003)
[9] Naimark, L., Foxlin, E.: Circular Data Matrix Fiducial System and Robust Image
Processing for a Wearable Vision-Inertial Self-Tracker. In: IEEE International
Symposium on Mixed and Augmented Reality (ISMAR 2002), pp. 27–36 (2002)
[10] Kato, H., Billinghurst, M., Blanding, R.: ARToolkit PC version 2.11.,
http://www.hitl.washington.edu/artoolkit/
Terrain Synthesis Based on Microscopic
Terrain Feature
Abstract. In this paper, we use real terrain elevation data and the mi-
croscopic terrain feature unit to synthesize the macroscopic terrain flex-
ibly and effectively. An interactive system provides users a convenient
and intuitive interface to profile microscopic terrain features using ter-
rain primitives. A number of terrain primitives, geometric objects such
as prism-like, tetrahedron-like, and cone-like objects regarded as the con-
cept representation object of microscopic terrain features, can be trans-
formed to construct the terrain profile. Then, these terrain primitives are
replaced by matching terrain units searched from real terrain elevation
database to seamlessly synthesize the macroscopic terrain, the landform.
As experimental results show, the resultant efficient synthesized terrains
are realistic, fitting user’s intuition.
Keywords: Terrain Modeling, Terrain Synthesis, Terrain Primitive.
1 Introduction
Terrain synthesis is essential to the construction of realistic outdoor virtual envi-
ronments. Many applications such as pilot training, scenery browser, game, etc.
need terrain models. Upon terrain synthesis, fractal terrain approaches and phys-
ical erosion approaches were two main techniques in the past few years. Fractal
terrain approaches based on the Brownian motion lack realism partly because
the statistical character of the surface is the same everywhere, i.e. the surface
has no global erosion features inherently. Physical erosion approaches simulate
fluvial, thermal, and diffuse erosion processes to create global stream/valley net-
works, talus slopes, and rounding of terrain features, so they are computation
intensive due to the complex physical model.
In this paper, we use several microscopic terrain features to construct a macro-
scopic terrain. Microscopic terrain features [1] are hill, mountain, plain, tableland
and plateau. They are classified by elevation, relative altitude and gradient. We
use several terrain units to construct a microscopic terrain feature. The macro-
scopic terrain represents a scene namely, a landform, being constructed by one or
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 644–655, 2008.
c Springer-Verlag Berlin Heidelberg 2008
Terrain Synthesis Based on Microscopic Terrain Feature 645
1. In the paper, we use the geometry and topology similarity to simplify the
matching procedure and exploit an alpha value to adjust the candidate set
of the terrain units that have the most geometry and/or topology similarity
such that the new method is more flexible and effective.
2. In Chiang [19], the matching method often obtains a strictly small number
of candidate terrain units and results in the stitching process performing not
well.
3. In the paper, we construct a graph to find the shortest path between all can-
didate terrain unit sets such that we can have a global matching result rather
than just locally deciding the best matched unit in Chiang [19]. Therefore,
we have synthetic terrains with better visual effect as you can see results
shown in the experimental section.
4. In short, the paper is more flexible and effective.
The rest of the paper is organized as follows. Section 2 reviews previous work
on terrain generation. Section 3 describes our approach. Section 4 demonstrates
the experimental results of our terrain synthesis method, and finally we give
some conclusions and future works in section 5.
2 Related Works
In the past few years fractal terrain approaches, [2],[3],[4],[5],[6],[8],[9],[10],
[13],[15], and physical erosion approaches, [11],[12], were two main approaches
in terrain synthesis.
Lewis [4] generalized the stochastic subdivision, constructed by Fournier,
Fussell, and Carpenter [2], based on the random process and estimation theories
to synthesize a noise with prescribed autocorrelation and spectrum functions.
646 S.-C. Tu, C.-Y. Huang, and W.-K. Tai
Lewis produced artifact-free noises with a variety of spectra, and the gray levels
in synthesized textures are interpreted as heights to obtain several distinct types
of synthetic terrains. The Poisson faulting process proposed by Mandelbrot [5]
is a sum of randomly placed step functions with random heights. It produced
a Brownian surface. While it is suitable for use on spheres for creation of plan-
ets, yet with O(n3 ) time complexity. The variation of stochastic subdivision de-
scribed by Miller [9] used a higher quality interpolation for the point estimation
that alleviated the problem of creases. Musgrave [10] generated fractal terrains
by terrain generation and erosion simulation. In terrain generation phase, Mus-
grave used summing band-limited noises which refer to as noise synthesis to
generate fractal terrain. In erosion simulation phase, Musgrave subdivided the
erosive processes into two categories: hydraulic erosion and thermal weather-
ing. The approach improves on previous fractal terrain synthesis and rendering
techniques and the result looks very realistic. Nagashima [11] used fractional
Brownian motion (fBm) obtained mainly from Poisson faulting by Mandelbrot
[5] and Peitgen and Saupe [13], and Fourier filtering by Peitgen and Saupe [13]
for creating fractal terrain models so that the approach could produce pictures
of realistic terrains. However, if the fBm and related fractal methods such as
Mastin [8] are used independently, it will produce unnatural terrains.
Physical erosion approaches are based on physical erosion theory. Nagashima
[11] proposed a simulation approach based on alluvial deposition laws, hydraulic
and thermal watering erosion. This approach shows amazing results. Roudier [12]
proposed a terrain simulation model for erosion and deposition using geological
parameters and laws from simplified geomorphological theories. However, the
approach only considered the processes that depend on running and infiltrated
water, and the timing costs a lot.
There are other methods which do not use fractal approach or physical simu-
lation to generate terrains. Marshall [7] proposes a procedural model to generate
scenes automatically. The procedural model uses fundamental data parameters
to create object, and the data parameters are always trivial for user. So this
method is useful even though its results are decided by the complete of proce-
dural model.
3 Approach
In this section, we specify the proposed approach for synthesizing macroscopic
terrains. First, user profiles a microscopic terrain in our interactive environment
using terrain primitives. Second, for specified terrain primitives, the matching
procedures, external and internal procedure, are applied to search for the best
matching terrain units from the terrain unit database. Third, the best matching
terrain units constitute for the corresponding terrain primitive. Finally, to have
the best smoothness and visual effect of the synthetic terrain, the minimum cut
approach is exploited to find an optimal seam such that adjacent terrain units
are well stitched.
Terrain Synthesis Based on Microscopic Terrain Feature 647
Fig. 1. (a) A mountain terrain can be divided into two boundary zones (red) and
one intermediate zone (blue). (b) Two tetrahedron-like geometry primitives and one
prism-like geometry primitive are suitable to conceptualize the mountain terrain. (c)
The cone-like geometry primitive can be used to conceptualize this type of mountain
well.
(a) (b)
Fig. 3. Two sample terrain units. (a) The prism-like terrain primitive is appropriate
to conceptualize the terrain unit at left side. (b) The tetrahedron-like terrain primitive
well profile to its terrain unit.
The terrain unit database contains as wide range of all types of microscopic
terrain features, like hill, mountain, plain, tableland and plateau, as possible
such that the matched terrain unit to user specified terrain primitive is as close
as possible. All terrain units are manually segmented from terrain elevation map
of real world according to rules as follows:
1. The height variation of each scanline in the terrain unit displays higher
elevation near the center and lower elevation around the boundary.
2. Each terrain unit contains only one main mountain ridge. This requirement
makes user specify the terrain primitive easier, and so does the matching
process.
Sample terrain units are shown in Figure 2. All segmented terrain units are
oriented to align each other and normalized in height and width length. Besides,
each terrain unit is conceptualized to correspond a terrain primitive using its
top and bottom cross sections and main range. Two sample terrain units with
corresponding terrain unit primitive are shown in Figure 3.
After user specifies terrain primitives in our system, the best matching terrain
unit is going to be searched to substitute for the terrain primitive. There are two
phases. First, the external matching phase finds out candidate terrain units that
have the most geometry and topology similarity to the specified terrain primitive.
The internal matching phase then determines the best matching terrain unit from
each candidate set.
(a) (b)
Fig. 4. (a) Two connected cross sections of adjacent terrain units are aligned with the
peak position. (b) The microscopic terrain is constructed by four terrain primitives and
a graph G = (V, E) is constructed to find the shortest path (pink) from v0 to v5.
specified terrain primitive with that of terrain unit primitive in the database
using
4or6
Gi,j = |T Pi (vk ) − T U Pj (vk )|
k=1
26
Ti,j = |T Pi (Pk ) − T Uj (Pk )|
k=1
Φi = {T Uj |(αgi,j + (1 − α)ti,j ) ≤ , 0 ≤ α ≤ 1}
where gi,j and ti,j are the normalized Gi,j and Ti,j respectively. The geom-
etry distance metric, gi,j , is used to evaluate the cross-section and mountain
ridge similarity and the topology distance metric, ti,j , measures terrain contour
(shape) similarity.
Since the microscopic terrain is composed of real terrain units, say I, the joint
of two adjacent terrain units matters the visual smoothness while rendering
the synthesized terrain. A smoothness cost measures the smoothness at joint as
follows:
3
ui,i+1 = |vi,k − vi+1,k | + M iddleSmotthCost(vi,1 , vi,2 , vi+1,1 , vi+1,2 , d, w)
k=1
+ M iddleSmotthCost(vi,1 , vi,3 , vi+1,1 , vi+1,3 , d, w)
In the first term, we concern the degree of elevation differences at three major
extreme vertices of two adjacent terrain units, say two end vertices and the vertex
with highest elevation at the adjacent cross section of a terrain unit as shown in
Figure 4(a). We further consider elevation variations between two pairs of major
vertices in the second term, MiddleSmoothCost function. Parameter d controls
the fineness of smoothness cost measurement between two endpoints, vx,1 and
vx,2 or vx,1 and vx,3 where x is i or (i+1), namely the depth of recursion. The last
parameter, w, weights the importance for each measurement. In our experiment,
the weighting w is assigned to 0.5, and the depth d is assigned to 10, i.e. 512
pairs of vertex will be measured. The pseudo code of the function is shown as
following.
pairs of candidate terrain units in adjacent candidate set Φi and Φi+1 , there is
an edge which is associated with the measured smoothness cost. In Figure 4(b),
for example, one constructs a microscopic terrain with four terrain primitives,
T P1 , T P2 , T P3 , and T P4 . Let Φi be the candidate set of terrain units of T Pi , i =
1, 2, 3, 4 respectively. Two extra vertices, a source vertex (v0) and a target vertex
(v5), are added in the G, where the smoothness cost is zero for all edges from
v0 to all candidate terrain units in Φ1 and from all candidate terrain units in
Φ4 to v5. The shortest path from v0 to v5 is found by Dijkstra shortest path
algorithm [16]. The vertices v12, v23, v32 and v42 on the path are the best
matching terrain unit of T P1 , T P2 , T P3 and T P4 respectively. In a short, we
employ the internal matching method to promise one optimal terrain unit set
from which the microscopic terrain feature demonstrates the best visual quality
and hopefully best fit to users intuition
Notice that users are allowed to assign a terrain unit from Φi as the best
matching unit of T Pi if they prefer some specific terrain features. If so, the
shortest path algorithm just applies to a graph which has some cut vertices (user
assigned terrain unit). The dark side of this assignment is that we may have a
shortest path where the terrain appearance might reveal some discontinuities at
joint boundaries.
Fig. 5. (a) Terrain units join together without stitching. (b) Terrain units join together
using the minimal cut approach for stitching. (c) and (d) are heightfield maps of (a)
and (b) respectively.
Simply substituting for specified terrain primitives may not make the synthetic
terrain look smoothness. Even for best matching terrain units, stitching for two
adjacent ones is still required. Let two adjacent terrain units be placed partly
overlapped. Our goal of stitching is to find the minimum cut at the overlap
region such that adjacent terrain units join together as seamless as possible.
Inspired by Kwatra et al. [17], we can simply employ the minimum cut [18] to
find a good seam, and then adjacent terrain units are stitched together along
this seam. Furthermore, to obtain the optimal seam, we first vary the overlay
region by iteratively offsetting one terrain unit within a range, and compute the
sum-of-height-differences cost [17] for the offset overlay region at each iteration,
and finally find the minimum cost at the offset overlap region with the lowest
sum-of-height-differences cost.
Figure 5(a) shows two terrain units which join together without stitching.
As you can see in the heightfield map of Figure 5(a), as shown in Figure 5(c),
the discontinuity is apparent around the joint. Figure 5(b) displays two terrain
units which are seamlessly joined together using the minimal cut approach. The
corresponding heightfield map of Figure 5(b), as shown in Figure 5(d), reveals
no discontinuity around the joint.
4 Experimental Results
We have implemented ant tested the proposed approach on a PC with a processor
of Pentium 4 2.4G and 1GB main memory. The graphic card is based on the
chipset of GeForce FX 5600. All the codes were written in C# and used DirectX
API. There are 698 terrain units segmented from real terrain elevation map of
Taiwan in our terrain unit database.
Figure 6 demonstrates heightfield maps of some matching terrain units with
respect to the user specified primitives. As you can see in the Figure 6, the
matched terrain units are very similar to the given terrain primitives in cross-
section, mountain ridge, and terrain contour. Consequently, if a composition of
terrain primitives conveys the overall macroscopic terrain in user’s mind con-
sistently, the synthesized terrain is intuitively predictable. Also, the resultant
synthesized terrain is realistic and reasonable because all terrain units are from
real world.
Figure 7(a) demonstrates a synthesized mountain which is composed of two
tetrahedron-like terrain primitives, and the corresponding terrain primitives are
shown in Figure 7(b) and 7(c). As you can see in Figure 7, the two terrain prim-
itives are used to simulate a cone-like single mountain. Figure 8(a) demonstrates
the synthesized mountain range which has one branch and the corresponding
terrain primitives are shown in Figure 8(b) and 8(c). It shows that each mi-
croscopic terrain feature can be effectively generated by a number of terrain
units. Moreover, several microscopic terrain features can be used to synthesize
a macroscopic terrain as shown in Figure 9. It shows that the synthesized scene
is realistic and flexible.
Terrain Synthesis Based on Microscopic Terrain Feature 653
Fig. 6. Terrain primitives and their matched terrain units. Top row: the hightfield map
of user specified terrain primitives. Bottom row: the heightfield map of matched terrain
units.
Fig. 7. (a) A mountain is synthesized by two terrain primitives. (b) The top view of
the corresponding terrain primitives of (a). (c) A perspective view of the corresponding
terrain primitives of (a).
Fig. 8. (a) A simple mountain range which has one branch is synthesized by several
terrain primitives, (b) The top view of the corresponding terrain primitives of (a). (c)
A perspective view of the corresponding terrain primitives of (a).
654 S.-C. Tu, C.-Y. Huang, and W.-K. Tai
Fig. 9. The synthesized macroscopic terrain using several microscopic terrain features
References
1. U.S. Geological Survey, http://www.usgs.org/
2. Fournier, A.D., Fussell, D., Carpenter, L.: Computer Rendering of Stochastic Mod-
els. Communications of the ACM 25, 338–371 (1982)
3. Gardner, G.Y.: Functional Modeling. In: SIGGRAPH course notes, Atlanta (1998)
4. Lewis, J.P.: Generalized Stochastic Subdivision. ACM Transactions on Graph-
ics 6(3), 167–190 (1987)
5. Mandelbrot, B.B.: Stochastic Models for the Earth’s Relief, the Shape and Fractal
Dimension of Coastlines, and the Number Area Rule for Islands. Proc. Nat. Acad.
Sci. USA 72, 2825–2828 (1975)
Terrain Synthesis Based on Microscopic Terrain Feature 655
1 Introduction
Digital watermarking technique is an important branch of information security. It can
be used in copyright protection, authentication, access control, and covert
communication etc [1]. Robust watermarking techniques are mainly used in copyright
protection etc. In theory, the robust watermark should be resistant to all kinds of
attacks. The attacks contain not only common image processing attacks, but also
some uncommon attacks, such as geometrical attack. In general, the geometrical
attack contains cropping, rotation, enlarging, shrinking, inclining, distortion etc. The
common image processing attack includes JPEG compression, Gaussian noise, and
filtering etc. In recent years, researchers have proposed many algorithms to resist the
geometrical attack. In references [2-5], the authors proposed the Fourier
transformation based methods to resist the typical geometrical attacks (rotation,
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 656–663, 2008.
© Springer-Verlag Berlin Heidelberg 2008
A Double Domain Based Robust Digital Image Watermarking Scheme 657
scaling, translation; RST), while the visual quality of the watermarked image is low
and they can not resist other geometrical attacks. In references [6], the authors
utilizing the techniques which are widely used in image registration to conquer the
geometrical attacks, while this kind of methods are very computationally complex and
sometimes can not recover the geometrically attacked image to the fine state, and then
the watermark cannot be extracted exactly. The performance of the existed
watermarking methods which are used to resist the geometrical attacks are very
limited by far.
In this paper, a double domains based digital image watermarking scheme is
proposed to resist both the geometrical attacks and the common image processing
attacks. Firstly, the watermark is embedded in the 3-level DWT coefficients by the
parity modulation method, and then the same watermark is embedded in the pixel
domain by the similar method. At the detector side, the watermark is extracted twice,
one is in DWT domain, the other is in pixel domain. Select the watermark which has
better visual quality as the final result. From the experimental results, we can clearly
see that, with the same embedding and extracting method, the spatial domain based
method has better robustness to the geometrical attacks, but very vulnerable to the
common image processing attacks. While the transformation domain based method
are on the contrary. By carefully regulating the parameters and selecting the
embedding sequence, the watermarks in the two domains are compatible to each
other.
The paper is organized as follows: section 2 is the concrete algorithms, section 3 is
the experimental results, and section 4 is the conclusions.
Assume the original gray image is A, the binary watermark image is W, W = ( wij ) ,
wij ∈ {0,1} . (i, j ) is a pair of coordinates. The embedding steps are as follows:
Step 1. For A, the 3-level DWT decomposition is performed to get the DWT
coefficient set LL3 .
p1
λ1 = round( ), (1)
δ1
where round is the function which rounds the element to the nearest integer. For
example, a = [-1.9, -0.2, 3.4, 5.6, 7.0], round [a] = [-2.0, 0, 3, 6, 7]. δ1 is the
quantization step.
658 C. Lin, J.-S. Pan, and Z.-M. Lu
⎧ 1
⎪ (λ1 − )δ1 , λ1 + wij ≡ 1 (mod 2)
⎪ 2
p1w = ⎨ , (2)
1
⎪(λ + )δ , λ1 + wij ≡ 0 (mod 2)
⎪⎩ 1 2 1
Step 4. The IDWT operation is performed to get the watermarked image A1w .
Step 1. For the watermarked image A1w , the 3-level DWT decomposition is performed
to get the DWT coefficient set LLw3 .
p1w
λ1w = floor ( ), (3)
δ1
where floor is the function which rounds the element to the nearest integers less than or
equal to itself. For example, a = [-1.9, -0.2, 3.4, 5.6, 7.0], floor [a] = [-2, -1, 3, 5, 7].
Step 3. Judge the value of the watermark bit wij′ according to the parity of λ1w . If λ1w
is the odd number, the watermark bit wij′ should be 1; if λ1w is the even number, the
watermark bit wij′ should be 0. That is:
⎪⎧1
wij′ = ⎨
, λ1w ≡ 1 (mod 2)
, (4)
⎪⎩0 , λ1 ≡ 0 (mod 2)
w
p2
λ2 = round( ),
δ2 (5)
where round is the function which rounds the element to the nearest integer , and δ 2
is the quantization step.
Step 3. According to the watermark to be embedded, change p2 to p2w .
⎧ 1
⎪⎪(λ2 − 2 )δ 2 , λ2 + wij ≡ 1 (mod 2)
p =⎨
w
2
, (6)
⎪(λ + 1 )δ , λ2 + wij ≡ 0 (mod 2)
⎪⎩ 2 2 2
When all the pixels are modified by the above method, we get the watermarked
image A2w .
The watermarked image A2w may be attacked, so we use A2w* to represent the attacked
watermarked image.
Step 1. Divide A2w* into pixel blocks with size k × k , the quantity of the pixel blocks is
decided by the actual size of A2w* .
Step 2. Extract the watermark from the pixel blocks. One pixel block corresponds to
one watermark bit wij′ . wij′ = ( w′) , w′ is the sub-watermark-bit, one pixel in the pixel
blocks corresponding to one sub-watermark-bit.
Step 2.1. In each pixel block, calculate the quantization value λ2w* of pixel p2w* by the
following formula.
pw
λ2w* = floor ( ), (7)
δ2
where floor is the function which rounds the element to the nearest integers less than
or equal to itself.
Step 2.2. Judge the value of sub-watermark-bit w′ according to the parity of λ2w* .
⎪⎧1
w′ = ⎨
,
λ2w* ≡ 1 (mod 2)
. (8)
⎪⎩0 , λ2 ≡ 0 (mod 2)
w*
4 Experimental Results
In our experiment, the target image is the popular Lena image with size 512 × 512
and 8 bits-per-pixel resolution. The input watermark is a bird image with size 64 × 64
and 1 bit-per-pixel resolution, each pixel bit in the watermark image is used as a
watermark bit. In the experiment, the value of the quantization step δ1 and δ 2 is 34.43
and 4.43, respectively. The wavelet base is Haar wavelet. BCR is the abbreviation of
bit-correct-rate of the extracted watermark.
The watermark is extracted twice, firstly the watermark is extracted from the pixel
domain, and then the watermark is extracted from the DWT domain. Select the
watermark which has better visual quality as the final result.
Fig. 1 is the original Lena image, the watermark image, the watermarked Lena
image, and the extracted watermark under no attacks. The PSNR value of the
watermarked Lena image is 37.4552dB, the watermark is embedded in the DWT and
the pixel domain.
Fig. 1. (a1) Original Lena image. (a2) Original watermark. (b1) Watermarked Lena image. (b2)
The extracted watermark.
(b2) from the pixel domain (c2) from the pixel domain
(a2) from the pixel domain
(b3) from the DWT domain (c3) from the DWT domain
(a3) from the DWT domain
(a1) under rotation attack (b1) under rotation attack (c1) under inclining attack
(a2) from the pixel domain (b2) from the pixel domain (c2) from the pixel domain
(a3) from the DWT domain (b3) from the DWT domain (c3) from the DWT domain
(a1)width70%, height70%
(b1)width 120%,height 120% (c1)width 100%,height 130%
(a2) from the pixel domain (b2) t from the pixel domain (c2) from the pixel domain
(a3) from the DWT domain (b3) from the DWT domain (c3) from the DWT domain
Fig. 2 to Figure 4 are the experimental results under geometrical attacks. Fig. 2 is
the experimental results under cropping attacks. From the experimental results we can
see that, when the watermarked Lena image is cropped, the watermark is also cropped
with it. The remained part of the watermark can also be used to insure the copyright.
We select the watermark images which have the better visual quality as the final
results. Fig. 3 is the experimental results under rotation, inclining, distortion, and
perspective attacks. We can see that the extracted watermark is also changed with the
662 C. Lin, J.-S. Pan, and Z.-M. Lu
Fig. 5. Experimental results under JPEG compression attack with different quality factor
(a1) Salt and Pepper noise (b1)Gaussian white noise (c1) Gaussian low pass filter
5 Conclusions
The paper proposed a double domain based watermarking scheme, the concrete
embedding and extracting algorithm is fulfilled by the parity modulation method. By
carefully regulating the values of the parameters and selecting the embedding
sequence, the watermarks in the two domains can smoke the calumet together and
make out contributions in different attacks. The experimental results show that the
proposed method is very robust to the geometrical attacks and common image
processing attacks.
References
1. Yin, H., Lin, C., Qiu, F., et al.: A survey of digital watermarking. Journal of Computer
Research and Development 42(7), 1093–1099 (2005)
2. Joseph, J.K., Ruanaidh, Ó., Pun, T.: Rotation, scale and translation invariant spread
spectrum digital image watermarking. Signal Processing 66(3), 303–317 (1998)
3. Lin, C.-Y., Wu, M., Bloom, J.A., Cox, I.J., Miller, M.L., Lui, Y.-M.: Rotation, Scale, and
Translation Resilient Watermarking for Images. IEEE Trans. on Image Processing 10(5),
767–782 (2001)
4. Pereia, S., Pun, T.: Robust Template Matching for Affine Resistant Image Watermarks.
IEEE Trans. on Image Processing 9(6), 1123–1129 (2000)
5. Kang, X., Huang, J., Shi, Y.Q., Lin, Y.: A DWT-DFT Composite Watermarking Scheme
Robust to Both Affine Transform and JPEG Compression. IEEE Trans. on Circuits and
Systems for Video Technology 8, 776–786 (2003)
6. Braudaway, G.W., Minter, F.: Automatic recovery of invisible image watermark from
geometrically distorted images. In: Proceedings on SPIE Security and Watermarking of
Multimedia Contents, CA, USA, vol. 3971, pp. 74–81 (2000)
ABF Based Face Texturing
1 Introduction
Texture mapping is one of the oldest techniques in Computer Graphics [7].
Originally, it is used to convey realism of objects, see [4][5]for an overview . As the
methods of texture mapping are throughout computer graphics and image processing
[4], its popularity is undoubted. In recent years, it is fast developed and powerfully
used in all kinds of modeling not only because it makes the model, like a face, more
vivid, but also because the low -cost hardware support for this technique is available.
It has already achieved great success in high quality image synthesis [4] and attracts
more and more focus from researchers.
In face texturing, there are mainly two kinds of ways to get the texture: 3D
scanner, like Cyberware, and image-based texture mapping [5][6].The former one
generates both the model and the texture by scanning the real object. It needs a large
database which is costly and the generated texture is not so satisfying [3]. A more
common way is to use input images. We focus our attentions on these previous works
and represent our method in section 1.1..
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 664–674, 2008.
© Springer-Verlag Berlin Heidelberg 2008
ABF Based Face Texturing 665
We define 41face feature points. 25 of which are totally anatomy based [16], denoting
eyes, mouth, nose and the face contour. In front image, as the triangulation of these 25
feature points can’t cover the whole face, we need another 8 more. In each side
image, we define 4 more on each ear. See the defined feature points in figure 1 and 2,
both in 3D model and 2D image domain.
666 X. Zhou et al.
2.2 Feature Point Detection Both in Face Images and 3Dface Model
To get the complete face texture, we prepare three face images: front image, left and
right. In the front face image, we use Active Apparent Model (AAM[14]) to get all 33
feature points automatically. AAM combines face shape and texture information,
giving effective and precise face alignment result. In this paper, we use the result of
our other work in lab ([15]) directly, selecting 25 of the 87 feature points generated.
In both left and right image, only half of the feature points can be seen, we pick 9 in
each manually, see figure 1.
Fig. 1. (a) Green dots are the AAM results, red dots are selected as our feature points (b) Red
dots are from AAM, the yellow ones can be inferred by the 25 features geometrically. (c)(d)
Green and red dots are manual, and yellow ones are the middle of reds.
3 Surface Parameterization
Since surface parameterization can generate mip-mapped texture mapping, it’s widely
used in this field. A large amount of surface parameterization method has been
proposed, for example, Floater[8] uses the solution of a convex combinations based
linear system to computer the vertex positions in parameterization plane; Eck et
al’s[9] proposed harmonic mapping; A. Sheffer presented angle based
flattening(ABF) in [10], and so on. Among all the work, we notice that the result of
ABF is free boundary and easy to converge. So, we focus on the work on ABF.
(b)
Vint is the set of interior vertices, and v* is the set of angles incident on vertex v.
(c)
k ⊕ 1 k Θ1 indicate the next and the previous angles in the triangles respectively.
See figure 2 to know what the symbols stand for, and the formulas are from [11]
We implement ABF in this paper, and the result can be seen in figure 7. The
parameterization result is continuous. In our use, we discretize it into pixels, to be the
texture domain. See details in the next section.
668 X. Zhou et al.
Table 1. The symbol defined for each kinds of vertexes, pixels and texels
Fig. 3. (a) The face model. (b) In texture domain, green dots are TFv , matching Fv in face
model and pFv in images; Blue dots mean TCv , matching Cv and pCv ; Orange dots are TNv ,
matching pNv .
ABF Based Face Texturing 669
After parameterization, every Fv is tied to a TFv , and every Cv find its TCv . In
section 2, pFv is picked out for every Fv .
We now describe each step in more details.
In this part, we look for the matched pixel in both images and texture domain for
every vertex in face model, which means finding pFv , TFv for Fv , and pCv , TCv
for Cv .TFv , TCv , pFv are already known in the above work, then, so we only have to
find pCv for Cv . We ask TCv for help and two steps are taken:
Step 1: In texture domain, we find the feature triangle, which TCv is in, as well as the
precise position of TCv in this feature triangle by barycentric coordinate.
Step 2: Based on the barycentric coordinate, we locate the pCv .
The barycentric coordinate is used here, see details as follow:
Barycentric Coordinate: For a point and a triangle in the same plane, the coordinates
of the point coordinates can be linearly denoted by the three vertices’ coordinates, and
the coefficients are the barycentric coordinates of the point to this triangle. See
figure 4:
si
p = c1 A1 + c2 A2 + c3 A3 , ci =
∑
i =1,2,3
si (5)
Based on this, we can find pCv . In section 2, feature triangles are formed by
feature vertices. For every TCv , we compute ci to every feature triangle, find the one
with all ci >0, named +tri _ f , and record the ci as ci _ tri _ f .The +tri _ f has a
matched feature triangle in images, we find this triangle and use the same ci _ tri _ f to
get the exact pCv . So, when finding the +tri _ f , pCv is also found.
However, this way to find the +tri _ f is low efficient obviously. We used an
improved way instead. See figure 5. We compute the ci of a TCv to an arbitrary
feature triangle. If all the ci >0, +tri _ f is found; else, using ci , we still can know
TCv outside which edge , and tag it edge E. Then we compute the ci of TCv to the
feature triangle sharing edge E with the first feature triangle. Repeating the operation
until finding the +tri _ f . The way works effectively in the experiment.
Obviously, to each face model vertex, the corresponding pixel in images is not
unique. There may be two pixels, one in front image, the other in one side image, tied
to a vertex. We use an area-weighted method to handle with this situation. See details
in section 4.2.
Fig. 5. Find a TCv ’s location in texture domain: black, gray and red triangles are all feature
triangles. Compute ci of TCv to black triangle, and know TCv is outside yellow edge, then
find the grey one. Repeat the operation, then we find the +tri _ f for the TCv , that is the red
triangle in the figure.
In this section, we first attach pixel values to TFv and TCv from multi-view images,
then set the pixel values to TNv .
For TFv and TCv , the corresponding pFv and pCv are already found in multi-view
images after section 4.1. We use an area-weighted method to handle with the situation
that one texel in texture domain has two matched pixels in images. For example, an
TFv has one matched pixel in front image, pFv 1, in feature triangle + F , another
matched pixel in left image, ,
pFv 2 in + L . S stands for the triangle’s area. We
ABF Based Face Texturing 671
Fig. 6. The blue point is the first selected incenter, the red ones are the second selected
We choose the incenter of the + para as the first selected sample. This incenter
separates the + para into three triangles. The second selected samples are the
incenters of the three small triangles. Four times recursion generate 40 TNv , which is
proved enough to represent the + para in our experiment. Then the pixel value of left
TNv in the + para can be set the same as its adjacent sample TNv .It works fast, and
the result turns to be legible. See result in Figure 7.
We discretize the parameterization result into pixels, and copy the pixel values
from multi-view images one by one. This operation minimizes the norm effect on
texture, and generates a seamless texture, without considering the illumination
variation.
5 Results
We experimented on several individuals. All the experiments are done on a common
computer with PC 1.7 GHz Pentium 4.The 3D face model feature points detection is a
pure interactive step, taking about 5 minutes. But we need to do this only once. It is a
step in system design, not in the application. In 2D image feature detection, as AAM
is real-time, left and right side images interaction cost all of the time in this step,
about 2 minutes each. ABF takes 4 seconds for a mesh of 4096 triangles. Getting the
pixel value for every pixel in the texture domain (616 × 496 pixels) from images needs
12 seconds. The whole procedure costs about 4.25 minutes from inputting three
images to getting the textured face. See the result in figure 7:
672 X. Zhou et al.
(b) The face model with feature points(c) The parameterization result
Fig. 7. (a)The three input images (b)The face model(c)the ABF result(d)The generated texture
on parameterization result(e)The textured face
ABF Based Face Texturing 673
References
1. Tarini, M., Yamauchi, H., Haber, J., Seidel, H.-P.: Texturing Faces. In: Proc. of Graphics
Interface, Calgary, Canada, pp. 89–98 (2002)
2. Rocchini, C., Cignoni, P., Montani, C., Scopigno, R.: nultiple Textures Stitching and
Blending on 3D Objects. In: Proceedings of Eurographlcs Workshop on Rendering,
Granada, pp. 173–180 (1999)
3. Blanz, V., Vetter, T.: A Morphable Model For The Synthesis Of 3D Faces. In:
Proceedings of ACM SIGGRAPH99 Conference, Los Angeles, pp. 187–194 (1999)
4. Heckbert, P.S.: Survey of Texture Mapping. IEEE Comput. Graph. Appl. (Sept.), 56–67
(November 1986)
5. Ebert, D.S., Musgrave, F.K., Peachey, D., Perlin, K., Worley, S.: Texturing and Modeling:
A Procedural Approach, 2nd edn. Academic Press, London (1998)
6. Yamauchi, H., Lensch, H.P.A., Haber, J., Seidel, H.-P.: Textures revisited. The Visual
Computer 21(4), 217–241 (2005)
7. Blinn, J.F., Newel, M.E.: Texture and Reflection in Computer Generated Images.
Communications of the ACM 19(10), 542–546 (1976)
8. Floater, M.S.: Parameterization and Smooth Approximation of Surface Triangulations.
Computer Aided Geometric Design 14, 231–250 (1997)
9. Eck, M., DeRose, T., Duchamp, T., et al.: Multiresolution Analysis of Arbitrary Meshes.
In: Computer Graphics. SIGGRAPH’s 1995. Annual Conference Series, pp. 173–182
(1995)
10. Sheffer, A., de Sturler, E.: Parameterization of Faceted Surfaces For Meshing Using
Angle-Based Flattening. Engineering with Computers 17(3), 326–337 (2001)
11. Sheffer, A., Lévy, B., Mogilnitsky, M., Bogomyakov, A.: ABF++: Fast and Robust Angle
Based Flattening. ACM Transactions on Graphics (TOG) 24(2), 311–330 (2004)
12. Kraevoy, V., Sheffer, A., Gotsman, C.: Matchmaker: Constructing Constrained Texture
Maps. ACM Transactions on Graphics 22(3), 326–333 (2003)
674 X. Zhou et al.
13. Ding, B., Wang, Y., Yao, J., Lu, P.: Fast Individual Face Modeling and Facial Animation
System. In: The 1st International Conference of E-Learning and Games, Hangzhou, China
(April 2006)
14. Cootes, T.F., Edwards, G.J., Taylor, C.J.: Active Appearance Models. In: Proc. European
Conf. Computer Vision 2, pp. 484–498 (1998)
15. Wang, S., Wang, Y., Chen, X.: Weighted Active Appearance Models. In: Huang, D.-S.,
Heutte, L., Loog, M. (eds.) ICIC 2007. LNCS, vol. 4681, pp. 1295–1304. Springer,
Heidelberg (2007)
16. EDU Resources.NET: http://www.eduresources.net/science/anatomy/bsh00.htm
Tile-Based Interactive Texture Design
1 Introduction
Textures have been a research focus for many years in human perception, computer
graphics and computer vision. Recently, research activities in this area emphasize on
texture synthesis. Given an example texture, a texture synthesis algorithm generates a
new one bearing the same visual characteristics. In spite of the fact that numerous
methods have been proposed for texture synthesis, how to design a variety of large
textures from a single small example texture is still a challenging problem.
Recently, Matusi et al. [1] developed a system for designing novel textures in the
space of textures induced by an input database. However, their texture interpolation
technique is based on a single one-to-one warping between the pairs of texture examples,
which might be too restrictive for textures with highly irregular structures, causing
discontinuous mappings of the patches to the original image. Shen et al. [2, 3] proposed
deformation-based texture design techniques for producing a variety of textures by
applying deformations and energy optimizations to the extracted patches of texture
elements. The main limitations of Shen et al.’s methods, lie in the facts that they have no
reusability for the resulting texture elements and the run-time synthesis speed is low.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 675–686, 2008.
© Springer-Verlag Berlin Heidelberg 2008
676 W. Dong, N. Zhou, and J.-C. Paul
Fig. 1. Our tile-based interactive texture design algorithm. The texture element of the yellow
flower is added into some tiles by user interactions. We can see that the designed tiles can
increase the variations of the result textures.
with little user interaction, and hence drastically broadens the variations of textures that
can be synthesized with the existing tile-based methods. As shown in Fig. 1, from a
single small example texture, our technique can create a variety of versatile textures,
with increased or decreased density of texture elements in the tiles. The main
contributions of our work consist of the following three aspects:
(1) A novel framework for designing a large variety of textures by integrating the
techniques of 1) tile-based texture synthesis, 2) interactive image editing, 3) genetic
algorithm (GA) based optimization, and 4) gradient-based Poisson image composition.
(2) An effective GA-based method for automatically extracting optimized sample
patches from the input example texture.
(3) A new composition based algorithm for synthesizing texture tiles from multiple
sources.
In the rest of the paper, we first introduce the related work on texture synthesis and
interactive image manipulation tools in Sec. 2. Then, in Sec. 3, we discuss the details of
our tile-based interactive texture design scheme. The extension of the existing tile
optimization algorithm using genetic algorithm is also described in Sec. 3. The method
for synthesizing textures from multiple sources using Poisson image composition is
presented in Sec. 4. After showing the experimental results in Sec. 5, we conclude the
paper and show some directions for future work in Sec. 6.
2 Related Work
Nowadays local region-growing methods are popularly used in texture synthesis. These
methods generate the texture by growing one pixel [9–12] or patch [13–18] at a time
with the constraint of maintaining coherence of neighboring pixels in the grown region.
Such approaches always suffer the time-consuming neighborhood matching in the
example and do not sufficiently meet real-time applications. On the other hand, some
near real-time texture synthesis methods usually achieved low quality results for the
lack of optimization in the pre-processing [19, 20] or needed very complex
pre-computation and data structures [21]. Recently efficient GPU-based texture
synthesis techniques [22, 23] have also been proposed, however they always demand a
high performance graphics hardware, and their methods suffer from the pixel-based
synthesis issue of performing poorly on textures with semantic structures not captured
by small neighbor-hoods.
An alternative approach is the tile-based methods. Cohen et. al. [4] developed a
stochastic algorithm to non-periodically tile the plane with a small set of Wangtiles at
run time. Wei [5] extended this work with GPU to improve tile-based texture mapping.
Ng et al. [6] presented another approach to generate a set of small texture tiles from an
input example. These tiles could also be tiled together to synthesize large textures.
Dong et. al [8] extend the algorithm in [6] to derive new tile sets by increasing the
number of sample patches. Our technique uses their ω-tiles as the tile set pattern.
678 W. Dong, N. Zhou, and J.-C. Paul
Interactive image manipulation and editing packages, such as Adobe Photoshop, are
commonly utilized by digital photographers. In their workflow, images are manipulated
directly and immediate visual feedback is provided.
Recently, many researchers proposed several interactive digital image editing tools
by using region-based methods, e.g., the magic wand in Photoshop, intelligent paint
[24], interactive graph-cut image segmentation [25], Poisson image editing [26],
GrabCut [27], lazy snapping [28], interactive image photomontage [29], drag-and-drop
pasting [30] and photo clip art [31].
The tile design process in our framework is most closely related to the method of
drag-and-drop pasting [30], where users use brushes to indicate which parts of a set of
photographs should be combined into a composite tile. By allowing the user to interact
with local texture elements, our proposed algorithm has the ability to increase or
decrease the density of texture elements interactively, which is suitable for designing a
variety of versatile textures from a single small example texture.
3 Our Approach
The goal of our algorithm is to enable the texture designer to easily create a set of
ω-tiles, from single or multiple sources.
Our proposed work flow is summarized as follows:
1. Load a small example texture image I .
2. Select a group of sample patches from I . The number of sample patches is
specified by the pattern of the ω-tile set which is going to be used [8]. Then construct
the tile set To1 , To 2 ,L, Tom with the extracted sample patches, as described in Sec. 3.2.
3. Design the ω -tiles by copying texture elements or other objects from the input
sources.
4. Repeat Step (3) until a satisfactory set of tiles Td 1 ,Td 2 ,LTdm is obtained.
5. Synthesize the final results by texture tiling [6, 8].
This work flow is illustrated by the sequence of images in Fig. 1. Given an input
example texture (Fig. 1(a)), a set of ω-tiles are produced (Fig. 1(b)) after applying the
sample patches selection and tile construction operations. Then the user uses brushes to
paint some texture elements interactively and the corresponding regions of those
texture elements are automatically calculated. These texture elements are stitched into
the tiles with the gradient-based Poisson image composition [26, 30], in order to obtain
the textures with varying local properties (Fig. 1(c)). Finally, by tiled the ω-tiles
together, large textures are synthesized (Fig. 1(d)).
In order to make the above work flow effective, several requirements should be met,
such as quickly generated previews of the overall result, a simple, intuitive and easy to
Tile-Based Interactive Texture Design 679
use mechanism for performing the local modification, and an undo function allowing
the user to modify previously specified adjustments. Our prototype implementation is
based on the interactive digital photomontage technique [29] and the
deformation-based interactive texture design system [3]. Besides the above
requirements, another important aspect we should notice is the construction of the
original ω-tile set. We should assure to generate a high quality tile set. For this, the most
important step is the selection of sample patches [8].
We use a similar framework of Dong et al. [7, 8] to select a set of optimized sample
patches from the input example. The algorithm is essentially based on genetic
algorithm (GA) [32–36]. GA starts with an initial set of randomly generated
chromosomes called a population where each chromosome encodes a solution to the
optimization problem. All chromosomes are evaluated by an evaluation function which
is some measure of fitness. A selection process based on the fitness values will form a
new population. The cycle from one population to the next is called a generation. In
each new generation, all chromosomes will be updated by the crossover and mutation
operations. Then the selection process selects chromosomes to form a new population.
After performing a given number of cycles, or other termination criteria is satisfied, we
denote the best chromosome into a solution, which is regarded as the optimal solution
of the optimization problem.
The main limitation of GA is sometimes it plunges the objective into a local optimal
solution. So we improve the GA-based optimized sample patches selection framework
in [8] by adding a ”genetic-dominance obtaining strategy” before the crossover process
of GA. The concept of our technique is as similar as the immuno-dominance obtaining
strategy in artificial immune system [37–39]. We denote the sets of sample patches
(chromosomes) as Α = [ A1 , A2 ,L Ak ] ,
⎛ a11 K a1n ⎞
⎜ ⎟
A=⎜ M O M ⎟
⎜a L a ⎟
⎝ k1 kn ⎠
⎧ 1 n
⎪1, ∑ a jj ≥0.25
ki = ⎨ n j =1
⎪0 otherwise
⎩
Denote Ai ∈ A is the best chromosome in the population, if F (c ) > F ( Ai ) , then
exchange each other. Note here F is the fitness function of the genetic algorithm [8].
We make Aj ∈ A, j = 1, 2,..., k , j ≠ i to get the genetic-dominance referring to a
possibility pid . Specifically, we set a 'j = H (a j + ai − c − 1) , where
⎧1 x > 0
H ( x) = ⎨
⎩0 otherwise
680 W. Dong, N. Zhou, and J.-C. Paul
Fig. 2. Compariation of the texture synthesis results using our algorithm and the algorithm in [8].
Form left to right: input example, feature mask, the result using the algorithm in [8], our results.
As shown in Fig. 2, we can see that our algorithm can generate better results than [8]
for the examples with some structural texture elements.
At Step (3) of our algorithm, the local change of an ω -tile is realized by replacing its
local regions with the texture elements from the input example. We call the ω -tile to be
locally designed the base tile Tbase (Tbase ∈ {Tb1 , Tb 2 ,..., Tbm }) and the source providing the
texture elements the reference texture I ref ( I ref = I in our experiments). As shown in
Fig. 1, the user does not need to precisely specify the region including the texture
Tile-Based Interactive Texture Design 681
elements (”yellow flowers”) in I ref . The corresponding region including the texture
elements is calculated automatically with the graph-cut-based energy optimization
technique [14]. The obtained texture elements are then embedded into to the base tile
Tbase seamlessly by the gradient-based Poisson optimization method [26, 30]. Such
local design is repeated several times, while at each step the user is allowed to choose
new texture elements by painting new strokes according to his creation.
Fig. 3. Examples of spatially varying designed textures using our tile-based method. Left two
columns are the input textures and results without tiles design, the others are the synthesized
textures with tiles design.
Our texture design method from multiple sources using image composition is extend
from [30]. The goal of multi-source texture design is to synthesize new textures that
capture the combined characteristics of several input images. As illustrated in Fig. 4, a
desert texture is selected as the background. This kind of textures can be frequently
explored in many computer games. Without losing generality, we use the basic 8-tiles
ω-tile set in [6, 8] as the bearer in our demos so that more ”designed” patterns can be
shown in a normal size tiling. To embed objects in some tiles, we first use GrabCut [27]
to produce the rough boundaries of the interested objects from the source images, then
the algorithm in [30] is proceeded to find the optimized boundaries. Finally we simply
use Poisson image composition to embed the objects in the specific tiles, as shown in
Fig. 4(b). In Fig. 4(c), we can see an image which illustrates a desert with pyramids,
stones, desert plants and sand dunes. It is generated in real-time with the tiles in
Fig. 4(b) and will be more vivid than a simple bare desert when it appears in a computer
game. We can choose ω-tile set with more tiles in it [8] to construct designed tiles when
682 W. Dong, N. Zhou, and J.-C. Paul
synthesizing large size images, especially for certain applications where continuous
patterns are required.
For some objects of the input sources, the sizes may be too large to be embedded in a
single tile of smaller size. In this case, we can compose the interested object with
multiple tiles, i.e. with a ”small” tiling of the tile set. Fig. 6 shows a demo of this
instance, here we embed Tintin with 4 tiles which can be tiled together to form a
relatively larger background. Then in the tiling process, we just use the Tintin pattern
when the same tiles tiling sequence is detected.
Fig. 4. A desert with pyramids, stones, desert plants and sand dunes. (a) Background example
texture. (b) Tiles design with poisson image editing from multiple sources. The blue curves are
the rough boundaries of the objects. (c) Image synthesis.
Our algorithm has been applied to a variety of example texture images. All the
experiments shown in this paper were run on a PC with Core2 Duo E6550 2.33GHz
CPU + 2GB RAM.
In Fig. 2, we compare our approach with the existing tile-based texture synthesis
technique in [8]. The synthesis quality of some structural texture examples could be
effectively improved using our enhanced GA-based ω-tile construction algorithm.
Fig. 3 gives examples demonstrating the capability of our technique for creating a
large variety of textures from a small examples, while maintaining the continuity of
texture features as well as the shapes of individual texture elements. Our methods
changes the density of texture elements (white flowers in the first row, the flowers and
grass in the second row) interactively according to the designer’s need. Fig. 5 shows
another result of image synthesis with tiles design using multiple sources. It is a busy
water surface with bears, water birds and boat models. We can see that our technique
can generate versatile images with textural backgrounds.
Tile-Based Interactive Texture Design 683
Fig. 5. Water surface with bears, water birds and boat models. (a) Background example texture.
(b) Tiles design. (c) Real-time tiling.
Fig. 6. A stone wall painted with Tintin, Doreamon and Mickey Mouse. The image is generated
in real-time with our tile-based interactive texture design algorithm.
684 W. Dong, N. Zhou, and J.-C. Paul
Acknowledgement
We thank Jiaya Jia, Vivek Kwatra and Sylvain Lefebvre for sharing their results and
texture examples on the web sites. This work is supported by National Natural Science
Foundation of China projects No. 60073007, 60473110; by National High-Tech
Research and Development Plan 863 of China under Grant No. 2006AA01Z301; and
by the MOST International collaboration project No. 2007DFC10740.
References
[1] Matusik, W., Zwicker, M., Durand, F.: Texture design using a simplicial complex of
morphable textures. ACM Trans. Graph. 24(3), 787–794 (2005)
[2] Shen, J., Jin, X., Mao, X., Feng, J.: Completion-based texture design using deformation.
The Visual Computer 22(9-11), 936–945 (2006)
[3] Shen, J., Jin, X., Mao, X., Feng, J.: Deformation-based interactive texture design using
energy optimization. The Visual Computer 23(9-11), 631–639 (2007)
[4] Cohen, M.F., Shade, J., Hiller, S., Deussen, O.: Wang tiles for image and texture
generation. ACM Trans. Graph. 22(3), 287–294 (2003)
[5] Wei, L.Y.: Tile-based texture mapping on graphics hardware. In: HWWS 2004:
Proceedings of the ACM SIGGRAPH/EUROGRAPHICS conference on Graphics
hardware, pp. 55–63. ACM Press, New York (2004)
[6] Ng, T.Y., Wen, C., Tan, T.S., Zhang, X., Kim, Y.J.: Generating an ω-tile set for texture
synthesis. In: Proceedings of Computer Graphics International 2005 (CGI 2005), Stone
Brook, NY, USA, pp. 177–184 (2005)
[7] Dong, W., Sun, S., Paul, J.C.: Optimal sample patches selection for tile-based texture
synthesis. In: CAD-CG 2005: Proceedings of the Ninth International Conference on
Computer Aided Design and Computer Graphics (CAD-CG 2005), pp. 503–508. IEEE
Computer Society, Washington (2005)
Tile-Based Interactive Texture Design 685
[8] Dong, W., Zhou, N., Paul, J.C.: Optimized tile-based texture synthesis. In: GI 2007:
Proceedings of Graphics Interface 2007, pp. 249–256. ACM, New York (2007)
[9] Bonet, J.S.D.: Multiresolution sampling procedure for analysis and synthesis of texture
images. In: SIGGRAPH 1997: Proceedings of the 24th annual conference on Computer
graphics and interactive techniques, pp. 361–368. ACM Press/Addison-Wesley Publishing
Co., New York (1997)
[10] Efros, A.A., Leung, T.K.: Texture synthesis by non-parametric sampling. In: ICCV 1999:
Proceedings of the International Conference on Computer Vision, vol. 2, p. 1033. IEEE
Computer Society, Washington (1999)
[11] Wei, L.Y., Levoy, M.: Fast texture synthesis using tree-structured vector quantization. In:
SIGGRAPH 2000: Proceedings of the 27th annual conference on Computer graphics and
interactive techniques, pp. 479–488. ACM Press/Addison-Wesley Publishing Co., New
York (2000)
[12] Ashikhmin, M.: Synthesizing natural textures. In: SI3D 2001: Proceedings of the 2001
symposium on Interactive 3D graphics, pp. 217–226. ACM Press, New York (2001)
[13] Efros, A.A., Freeman, W.T.: Image quilting for texture synthesis and transfer. In:
SIGGRAPH 2001: Proceedings of the 28th annual conference on Computer graphics and
interactive techniques, pp. 341–346. ACM Press, New York (2001)
[14] Kwatra, V., Schodl, A., Essa, I., Turk, G., Bobick, A.: Graphcut textures: image and video
synthesis using graph cuts. ACM Trans. Graph. 22(3), 277–286 (2003)
[15] Nealen, A., Alexa, M.: Hybrid texture synthesis. In: EGRW 2003: Proceedings of the 14th
Eurographics workshop on Rendering, Aire-la-Ville, Switzerland, Switzerland,
Eurographics Association, pp. 97–105 (2003)
[16] Wu, Q., Yu, Y.: Feature matching and deformation for texture synthesis. ACM Trans.
Graph. 23(3), 364–367 (2004)
[17] Liu, Y., Lin, W.C., Hays, J.: Near-regular texture analysis and manipulation. ACM Trans.
Graph. 23(3), 368–376 (2004)
[18] Kwatra, V., Essa, I., Bobick, A., Kwatra, N.: Texture optimization for example-based
synthesis. ACM Trans. Graph. 24(3), 795–802 (2005)
[19] Zelinka, S., Garland, M.: Towards real-time texture synthesis with the jump map. In:
EGRW 2002: Proceedings of the 13th Eurographics workshop on Rendering, Aire-la-Ville,
Switzerland, Eurographics Association, pp. 99–104 (2002)
[20] Zelinka, S., Garland, M.: Jump map-based interactive texture synthesis. ACM Trans.
Graph. 23(4), 930–962 (2004)
[21] Liang, L., Liu, C., Xu, Y.Q., Guo, B., Shum, H.Y.: Real-time texture synthesis by
patch-based sampling. ACM Trans. Graph. 20(3), 127–150 (2001)
[22] Lefebvre, S., Hoppe, H.: Parallel controllable texture synthesis. ACM Trans. Graph. 24(3),
777–786 (2005)
[23] Lefebvre, S., Hoppe, H.: Appearance-space texture synthesis. ACM Trans. Graph. 25(3),
541–548 (2006)
[24] Barrett, W.A., Cheney, A.S.: Object-based image editing. In: SIGGRAPH 2002:
Proceedings of the 29th annual conference on Computer graphics and interactive
techniques, pp. 777–784. ACM, New York (2002)
[25] Boykov, Y., Veksler, O., Zabih, R.: Fast approximate energy minimization via graph cuts.
IEEE Transactions on Pattern Analysis and Machine Intelligence 23(11), 1222–1239
(2001)
[26] Pérez, P., Gangnet, M., Blake, A.: Poisson image editing. ACM Trans. Graph. 22(3),
313–318 (2003)
686 W. Dong, N. Zhou, and J.-C. Paul
[27] Rother, C., Kolmogorov, V., Blake, A.: “grabcut”: interactive foreground extraction using
iterated graph cuts. ACM Trans. Graph. 23(3), 309–314 (2004)
[28] Li, Y., Sun, J., Tang, C.K., Shum, H.Y.: Lazy snapping. ACM Trans. Graph. 23(3),
303–308 (2004)
[29] Agarwala, A., Dontcheva, M., Agrawala, M., Drucker, S., Colburn, A., Curless, B.,
Salesin, D., Cohen, M.: Interactive digital photomontage. In: SIGGRAPH 2004: ACM
SIGGRAPH 2004 Papers, pp. 294–302. ACM, New York (2004)
[30] Jia, J., Sun, J., Tang, C.K., Shum, H.Y.: Drag-and-drop pasting. ACM Trans. Graph. 25(3),
631–637 (2006)
[31] Lalonde, J.F., Hoiem, D., Efros, A.A., Rother, C., Winn, J., Criminisi, A.: Photo clip art.
ACM Trans. Graph. 26(3), 3 (2007)
[32] Holland, J.H.: Adaptation in natural and artificial systems. University of Michigan Press,
Ann Arbor (1975)
[33] Koza, J.R.: Survey of genetic algorithms and genetic programming. In: Proceedings of
1995 WESCON Conference, pp. 589–594. IEEE, Los Alamitos (1995)
[34] Pan, J., McInnes, F., Jack, M.: Application of parallel genetic algorithm and property of
multipleglobal optima to vq codevector index assignment for noisy channels. Electronics
Letters 32(4), 296–297 (1996)
[35] Liu, B., Liu, B.: Theory and Practice of Uncertain Programming. Physica-Verlag (2002)
[36] Sun, H., Lam, K.Y., Chung, S.L., Dong, W., Gu, M., Sun, J.: Efficient vector quantization
using genetic algorithm. Neural Comput. Appl. 14(3), 203–211 (2005)
[37] de Castro, L.N., Zuben, F.J.V.: The clonal selection algorithm with engineering
applications. In: Proceedings of GECCO 2000: Workshop on Artificial Immune Systems
and Their Applications, Las Vegas, Nevada, USA, August 2000, pp. 36–39 (2000)
[38] de Castro, L.N., Timmis, J.: Artificial Immune Systems: A New Computational
Intelligence Approach. Springer, Heidelberg (2002)
[39] de Fran¸ca, F.O., Zuben, F.J.V., de Castro, L.N.: An artificial immune network for
multimodal function optimization on dynamic environments. In: GECCO 2005:
Proceedings of the 2005 conference on Genetic and evolutionary computation, pp.
289–296. ACM Press, New York (2005)
Efficient Method for Point-Based Rendering on GPUs
1 Introduction
Physically based modeling and animation of deformable objects has been an ongoing
research topic in the field of computer graphics since the 80's. We believe that al-
though great achievements have been made, the full potential of alternative modeling
primitives and also the possible areas of application have not yet been fully explored.
In this research project, we investigate point primitives for modeling and rendering
deformable objects, which simplify the simulation of topological changes due to,
fracture, melting, clay-like splitting and merging and other possible phenomena.
In the last years point-based rendering has been shown to offer the potential to
outperform traditional triangle based rendering both in speed and visual quality when
it comes to processing highly complex models. Existing surface splitting techniques
achieve superior visual quality by proper filtering but they are still limited in render-
ing speed.
When considering point-based surface representations in general, we further dis-
,
tinguish between a piecewise constant point sampling [1 2] and piecewise linear
surface splats [3]. In this paper we focus on surface splats, since besides providing a
higher approximation order, they also allow for more efficient rendering and achieve
a higher visual quality by sophisticated anti-aliasing techniques [4].
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 687–694, 2008.
© Springer-Verlag Berlin Heidelberg 2008
688 L. Yan and Y. Yuan
2 Related Work
Most of the work in direct volume visualization in recent years has been focused on
texture-based approaches. Cabral B[ 5] and Stegmaier S [6] presented a flexible
framework for single pass GPU-raycasting that takes advantage of the easily extensi-
ble raycasting approach to demonstrate a number of non-standard volume rendering
techniques, including translucent material and self-shadowing isosurfaces, as well as
an acceleration technique based on exploiting inter-frame coherence. Hadwiger et al
[7] presented a GPU-raycasting system for isosurface rendering. They employ a two-
level hierarchical representation of the volume data set for efficient empty space skip-
ping and use an adaptive sampling approach for iterative refinement of the isosurface
intersections.
To improve robustness, Ohtake and Belyaev advocated moving the triangle cen-
troids to the zero iso-contour instead of the nodes, and matching the triangle normals
with the implicit surface normals [8].
The main source for real-time point-cloud refinement methods is the work [9] by G.
Guennebaud. The rendering of point splats involves several sub-tasks: first the size
and shape of the splats have to be determined from the current viewing parameters so
that we get a holefree image[10] .Using these techniques alone already results in mid-
quality elliptical but still unfiltered surface splats. Nevertheless it provides a much
Efficient Method for Point-Based Rendering on GPUs 689
3d scan datasets
Initial mesh
Generate/update
volume mesh
Analysi
No
Shape and
Errors Acceptable
Yes
Volume rendering
Meshes Extraction
better representation of the geometry than fixed splat shapes, especially noticeable
near contours. Each point in the point set p, have a position, a normal and a radius.
The refined point set is iteratively defined. Every weighted point gives rise to a dis-
tance function, namely the power distance function,
π p : R 3 → R, x a x − z − r
2
(1)
Let P be a set of weighted point in R3.The power diagram of P is a decomposition of
R3 into the power cells of the points in P[11].
Theorem 1. The natural coordinates satisfy the requirements of a coordinate system,
∈
namely, for any p P, λP (q) = δ pq where δ pq is the Kronecker symbol and the point x is
the weighted center of mass of its neighbors. That is,
x = ∑ λ p ( x) p, with ∑ λ p ( x) = 1
p∈P p ∈P (2)
Induced distance function.
690 L. Yan and Y. Yuan
Using point-based methods to solve a visualization problem may seem like the wrong
way to go since polygonal methods is well known and understood. The graphics
hardware is also designed and highly optimized for polygons. We propose the use of
point primitives in the context of 3D shape modeling. We split and merge to dynami-
cally adapt the surface sampling density during simulation. To maintain a close con-
nection between physical particles and surface samples, we use a space warping ap-
proach, similar to the free-form shape deformation scheme[12] .We use a linear ver-
sion of the Moving Least Squares projection for dynamic surface reconstruction.
We define a point-based object Pi 0 = Pi , i = 1,..., n as a cloud of surfels and phyxels.
where k≥1 [14].
P2ki −1 = ω1 (k ) Pi k −1 + ω0 (k ) Pi +k1−1 , (3)
P = ω0 (k ) Pi
k
2i
k −1
+ ω1 (k ) P k −1
i +1
; (4)
, ω1 (k ) = 1 + 2 cos(l 2 ) , ω0 (k ) + ω1 (k ) = 1 .
k
where, ω0 (k ) = 1
2(1 + cos(l 2 ))
k
2(1 + cos(l 2 ))
k
The point set (point-cloud) P 0 , serves as a starting point for the algorithm. The solu-
tion is iterative, and for each iteration, points in Pn is either accepted as part of the
final point-set S, or refined and added to Pn+1, which serves as input to the next
iteration. Each point P ∈ P n is locally refined in each iteration[13]. This procedure
contains several steps.
First, the particle p is moved closer to the real isosurface with help of an approxi-
mation in form of one Newton-Raphson step
f (p n )∇f (p n )
pn +1 =p n −
∇f (p n )
2
(5)
Where
f (p) = cS (p) − kisocolor (6)
At this stage, we have a triangle mesh with the correct connectivity and vertices in
screen space, e.g. with coordinates [xp,yp, zp] T . In order to render this mesh and to
compute reflections and refractions of the 3D environment, the vertices are projected
back into world space while the connectivity is kept fixed. We, therefore, need to
invert the transformation given in Eq. (1) and Eq. (2). Let Q ∈ R 4×4 be the inverse of
Efficient Method for Point-Based Rendering on GPUs 691
the projection matrix, i.e. Q = P−1. With Q, the world coordinates [xp,yp, zp] T can
be computed via the inverse projection equation.
⎡ x⎤ ⎡ ( −1 + 2 x p / W ) w ⎤
⎢ y⎥ ⎢ ⎥
⎢ ⎥ = Q ⎢(−1 + 2 y p / W ) w⎥ (7)
⎢z⎥ ⎢ zp ⎥
⎢ ⎥ ⎢ ⎥
⎣1⎦ ⎣ w ⎦
At this point, we do not have the projective divisor w. The GPU stores internally a
transformed version of the true depth to achieve better accuracy for near objects. The
exact depth can be computed as d = za /( z − zb ) , where z is the transformed depth
from the GPU
za =
d far , z = d near d far and d near and d far are the distances of the near
d far − d near d near − d far
b
and far clipping planes, respectively. To maximize depth accuracy, the near and far
planes are set tightly around the interval of possible depths. The near plane is set
slightly closer as the closest point-to-triangle distance of the object vertices, and
the far plane is set slightly further away than the maximum distance between cage
vertices[14].
With lM (ei ) the edge length in the metric M . Here, an efficiency index close to 0.86
is obtained in each case. d = x 2 + y 2 + z 2
⎧ -π
⎪0.1sin(50 x) if x ≤ 50
⎪ (10)
⎪ -π 2π
f3d ( x, y, z ) = ⎨sin(50x) if <x≤
⎪ 50 50
⎪ 2π
⎪0.1sin(50 x) if 50 < x
⎩
See figure 2 for an example rendered with this technique at real-time rates.
(a) (b)
Fig. 2. Triceratops point-based Animation and Real-time Rendering(Vertices number : 4209;
triangles numbers:5008; tetrahedral:16910) (a) mesh generation from point-sets (b) Hard-
ware accelerated (>60Hz) rendering.
0.4
0.3
0.2
0.1
0.0
-0.1
0 10 20 30 40 50 60 70
Frame
Fig. 3. Plot of changes in maximum aspect ratios during simulation of isometric contraction for
dynamics and optimization based meshes.
Efficient Method for Point-Based Rendering on GPUs 693
Figure 3 shows the relative change in maximum aspect ratio observed during an
isometric contraction of the biceps for meshes created using the optimization algo-
rithm and using the dynamics algorithm. Similar results were observed for the triceps
and during isotonic contraction [16]. These results suggest that initial mesh quality
may be misleading and not sufficient to guarantee performance of a mesh throughout
simulation. In all of our comparisons, the optimization based meshes were of higher
quality initially, but tended to undergo as much as a 67% change in maximum aspect
ratio during muscle contraction, whereas the dynamics based meshes tended to de-
grade by only 20.02%.
References
1. Pfister, H., Zwicker, M., Van Baar, J., Gross, M.: Surfels: Surface Elements as Rendering
Primitives. In: Proc. of ACM SIGGRAPH 2000, pp. 335–342 (2000)
2. Alexa, M., Behr, J., Cohenor, D., Fleishman, S., Levin, D., Silva, C.T.: Point Set Surfaces.
In: Proc. of IEEE Visualization 2001, pp. 21–28 (2001)
3. Boulic, R., Magnenat-Thalmann, N., Thalmann, D.: A Global Human Walking Model with
Real Time Kinematic Personification. The Visual Computer 6, 344–358 (1990)
4. Kobbelt, L., Botsch, M.: A Survey of Point-based Techniques in Computer Graphics.
Computers & Graphics 28(6), 801–814 (2004)
5. Cabral, B., Cam, N., Foran, J.: Accelerated Volume Rendering and Tomographic Recon-
struction using Texture Mapping Hardware. In: Proceedings of the 1994 Symposium on
Volume Visualization, pp. 91–98 (1994)
6. Stegmaier, S., Strengert, M., Klein, T., Ertl, T.: A Simple and Flexible Volume Rendering
Framework for Graphics-Hardware based Raycasting. In: Proceedings of the International
Workshop on Volume Graphics 2005, pp. 187–195 (2005)
7. Hadwiger, M., Sigg, C., Scharsach, H., Bhler, K., Gross, M.: Real-Time Ray-Casting and
Advanced Shading of Discrete Isosurfaces. In: Proceedings of Eurographics 2005, pp.
303–312 (2005)
694 L. Yan and Y. Yuan
8. Ohtake, Y., Belyaev, A.G.: Dual/Primal Mesh Optimization for Polygonized Implicit Sur-
faces. In: Proc. of the 7th ACM Symp. on Solid Model. and Appl., pp. 171–178. ACM
Press, New York (2002)
9. Guennebaud, G., Barthe, L., Paulin, M.: Interpolatory Refinement for Real-Time Process-
ing of Point-Based Geometry. In: Computer Graphics Forum, Eurographics 2005 confer-
ence proceedings, September 2005, vol. 24, pp. 657–667 (2005)
10. Frey, P., Alauzet, F.: Anisotropic Mesh Adaptation for CFD Computations. Comput.
Methods Appl. Mech. Engrg. 194, 48–49 (2005)
11. Guennebaud, G., Barthe, L., Paulin, M.: Dynamic Surfel Set Refinement for High Quality
Rendering. In: Computer & Graphics, pp. 827–838. Elsevier Science, Amsterdam (2004)
12. Alexa, M., Adamson, A.: On Normals and Projection Operators for Surfaces Defined by
Point Sets. In: Proceedings of Eurographics Symposium on Point-Based Graphics 2004,
pp. 150–155 (2004)
13. Frey, P.J.: About Surface Remeshing. In: Proc.of 9th Int. Meshing Roundtable, New Or-
leans, LO, USA, pp. 123–136 (2000)
14. Ganovelli, F., Cignoni, P., Montani, C., Scopigno, R.: A Multiresolution Model for Soft
Objects Supporting Interactive Cuts and Lacerations. In: Eurographics, pp. 271–282
(2000)
15. You-wei, Y., La-mei, Y.: A Neural Network Approach for 3-D Face Shape Reconstruc-
tion. In: Proceedings of 2002 International Conference on Machine Learning and Cyber-
netics, vol. 5, pp. 2073–2077. IEEE, Beijing (2002)
16. Bordeux, C., Boulic, R., Thalmann, D.: An Efficient and Flexible Perception Pipeline for
Autonomous Agents. In: Proceedings of Eurographics 1999, Milano, Italy, pp. 23–30
(1999)
Efficient Mushroom Cloud Simulation on GPU
1 Introduction
Most natural sceneries effects, such as clouds, fire, rain, snow, smoke, sparks,
blood, etc, are full of motion, full of chaos and full of fuzzy objects, and change
with time past. These natural sceneries effects are routinely created in todays
video games. So the simulation of natural sceneries becomes a hot topic in the
research field of computer graphics.
Usually, the particle systems could be used to simulate these natural sceneries
on CPU [1, 2, 3, 4, 5]. However, if the number of particles is above 10K, the
particle system on CPU is difficult to run on real-time. It is required that plenty
of particles more than 10K should be need in simulate system of photorealist
natural sceneries effects. Today, with the development of GPU, we could deal
with complex computing and programming on GPUs.
In this paper, we present a new method to simulate the Mushroom Cloud
efficient on GPU using advanced particle system, and our particle system is a
state-preserving simulation system. We provide the visual-only model of Mush-
room Cloud and we divide the Mushroom Cloud into five portions. Then we
present our advanced particle system method in detail. Our particle system
method processes the birth and death of particles via index on CPU and uses a
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 695–706, 2008.
c Springer-Verlag Berlin Heidelberg 2008
696 X. Cai, J. Li, and Z. Su
pair of Floating Point Textures on GPU to store the dynamic attributes of par-
ticles. This method also updates the dynamic attributes of particles, handles the
collision between particles and other models and renders the system on GPU. We
also provide a three-layers hierarchical structure to manage the particle system
and batch rendering the particles having the similar attributes.
In this paper, after exploring the related work of particle system, we present
our visual-only model of Mushroom Cloud. Then we introduce our particle sys-
tem method on GPU. In Section 5, we give the three-layers hierarchical struc-
ture of our advanced particle system. In Section 6, we show the results using our
method before we draw the conclusion in Section 7.
2 Related Work
Particle system has a long history in video games and computer graphics. Very
early video games in the 1960s already used 2D pixel clouds to simulate explo-
sions. In 1983, Reeves [6] first described the basic motion operations and the
basic data representing a particle, which both have not been altered much since
being presented. In 1990, an implementation on parallel processors of a super
computer has been done by Sims [7]. In 2000, McAllister [8] also described many
of the velocity and position operations of the motion simulation which are used
in this paper. The latest descriptions of CPU-based particle system for using
in video games and photorealist natural sceneries have been done by Wang et
al [1], Liu et al [4], Burg [9].
With the development of GPU, several forms of physical simulation have
recently been developed for modern GPU. In 2003, Harris [10] has used GPU
to perform fluid simulations and cellular automata with similar texture-based
iterative computation. Green [11] describes a cloth simulation using simple grid-
aligned particle physics, but does not discuss generic particle systems’ problems,
like allocation, rendering and deallocation. However, their algorithms do not
show the necessary properties to exploit the high frame-to-frame coherence of
the particle system simulation. Recently, Schneider et al [12], Li et al [13], Livny
et al [14], and Eric et al [15] have used GPU to render large scale terrain scene.
Christopher et al [16] also provide the method of real-time mesh simplification
using GPU. As GPU could deal with complex computing so fast, we want to
implement particle system on GPU.
Some particle systems have been implemented with vertex shaders (also called
vertex programs) on programmable GPUs in NVIDIA SDK [17]. However, these
particle systems are stateless. They do not store the current attributes of the
particles, including current position, current velocity, etc. To determine a par-
ticle’s position, the system needs to find a closed form function for computing
the current position only from initial values and the current time. Stateless par-
ticles are not meant to collide with the environment. They are only influenced
by global gravity acceleration and could be simulated quite easily with a simple
function. As a consequence, such particle system can hardly react to a dynamic
Efficient Mushroom Cloud Simulation on GPU 697
The Column portion is the smoke column of Mushroom Cloud from the terrain
surface to the air. It’s shape like a curved column, so we use an inverse pro-
portion function to create it. The Column portion joints the Core portion and
the Bottom Wave portion. In Fig. 4, as the Core portion is rising, the height of
Column portion becomes higher and the radius is outspread. Then the radius is
shrinking.
At the beginning of explosion, the Core portion is not the mushroom shape.
It is close to a hemisphere. And its position is very low. As the time past, the
hemisphere shape Core portion is moving up and changing to the Mushroom
Cloud that we all know. We use the hemi-ellipse function and the Double Folium
function to create the Core portion. As the Core portion rising, the Mushroom
Cloud is shrinking. Fig. 6 shows the changing process.
The particles in a system can either exist permanently or only for a limited time.
A static number of permanently existing particles represents the simplest case
for the simulation, because it only requires uploading all initial particle data
to the particle attributes textures once. As this case is rather rare, we assume
Efficient Mushroom Cloud Simulation on GPU 701
a varying number of short-living particles for the rest of the discussion. The
particle system must then process the birth of a new particle like its allocation,
the death of a particle and its deallocation.
The birth of a particle requires associating new data with an available index
in the attribute textures. Since allocation problems are serial by nature, this can
not be done efficiently with a data-parallel algorithm on the GPU. Therefore an
available index is determined on the CPU via traditional fast allocation schemes.
The simplest allocation method uses a stack filled with all available indices.
In our method, the particle’s death is processed independently on the CPU
and GPU. The CPU registers the death of a particle and adds the freed index
to the allocator. The GPU does an extra pass over the particle data: The death
of a particle is determined by the time of birth and the computed age. The dead
particle’s position is simply moved to invisible areas, e.g. infinity. As particles
usually fade out or fall out of visible areas anyway at the end of their lifetime,
the extra pass rarely really needs to be done. It is a basically clean-up step to
increase rendering efficiency.
The most important attributes of a particle are its position and velocity. So we
just deal with particle position and velocity. The actual program code for the
attributes simulation is a pixel shader which is used with the stream processing
algorithm. The shader is executed for each pixel of the render target by rendering
a screen-sized quad. The current render target is set to one of the double buffer
attribute textures. The other texture of the double buffer is used as input data
stream and contains the attributes from the previous time step. Other particle
data, either from inside the attribute textures or as general constants, is set
before the shader is executed.
Update Velocities: There are several velocity operations that can be combined
as desired: global forces (e.g. gravity, wind), local forces (attraction, repulsion),
velocity dampening, and collision responses. For our GPU-based particle system
these operations need to be parameterized via pixel shader constants. Their
dynamic combination is a typical problem of real-time graphics. Comparable
to the problem of light sources and material combinations, it could be solved
in similar ways. Typical operation combinations are to be prepared in several
variations beforehand. Other operations could be applied in separate passes, as
all operations are completely independent.
Global and local forces are accumulated into a single force vector. The accel-
eration can then be calculated with Newtonian physics as Formula 1. In Formula
1, a is the acceleration vector, F is the accumulated force and m is the mass of
the particle. If all particles have unit mass and forces have the same value, the
accelerations can be used without further computation.
F
a= . (1)
m
702 X. Cai, J. Li, and Z. Su
v = v + a · t. (2)
The velocity is then updated from the acceleration with a simple Euler inte-
gration in the form of Formula 2. In Formula 2, v is the current velocity, v is the
previous velocity and t is the time step.
p = p + v · t. (3)
Update Positions: Euler integration has already been used in the previous
section to update the velocity by using the acceleration. The computed velocity
can be applied to all particles in just the same way. We use Formula 3 to update
the position. In Formula 3, p is the current position and p is the previous position.
Before rendering particles, we should copy the particle data from the floating
point texture to vertex data. Copying the particles data from a texture to vertex
data is a hardware feature that is only just coming up in PC GPUs. OpenGL [21]
offers vertex textures with ARB vertex shader extension. OpenGL also provides
two functions, glReadBuffer and glReadPixels. These functions could copy the
particle data from the floating point texture to vertex data.
The particles can be rendered as point sprites, triangles or quads. If the particle
is rendered as triangles or quads, the particle may have three vertices or more
vertices. So it is required that we should recompute the vertices position of
particle before rendering. To avoid this overhead, our implementation uses point
sprites.
Fig. 7. Three-layers hierarchical structure particle system (’Par’ stands for ’Particle’)
for the birth of a new Particles Cluster, the death of a Particles Cluster and its
deallocation. In this way, we could use the ability of GPUs rendering triangles
in batches and implement plenty of particles in video games and photorealist
natural sceneries.
6 Results
Fig. 8. Comparison between particle system on CPU and particle system on GPU
704 X. Cai, J. Li, and Z. Su
Cloud and we divide the Mushroom Cloud into five portions. Then we present our
advanced particle system method in detail. Our particle system method processes
the birth and death of particles via index on CPU and uses a pair of Floating
Point Textures on GPU to store the dynamic attributes of particles. This method
also updates the dynamic attributes of particles, handles the collision between
particles and other models and renders the system on GPU. We also provide
a three-layers hierarchical structure to manage the particle system and batch
rendering the particles having the similar attributes. Finally, the experiments
prove that our method is feasible and high performance.
As a future possibility, we are working on using our method to implement
other complex natural phenomenon, and developing the method to deal with
the collision between particles and other particles.
References
1. Changbo, W., Zhangye, W., Qunsheng, P.: Real-time Snowing Simulation. The
Visual Computer 22(5), 315–323 (2006)
2. Huagen, W., Xiaogang, J., Qunsheng, P.: Physically based real time fountain sim-
ulation. Chinese Journal of Computers 21(9), 774–779 (1998)
3. Ruofen, T., Lingjun, C., Guozhao, W.: A method for quick smog simulation. Jour-
nal of Software 10(6), 647–651 (1999)
4. Xiaoping, L., Ye, Y., Hao, C., et al.: Real-time simulation of special effects in
navigation scene. Journal of Engineering Graphics (3), 44–49 (2007)
5. Yu, G., Lin-can, Z., Wei, C., Qun-sheng, P.: Real Time Waterfall Simulation Based
Particle System 16(11), 2471–2474 (2004)
6. Reeves, W.T.: Particle Systems-Technique for Modeling a Class of Fuzzy Objects.
In: Proceedings of SIGGRAPH 1983 (1983)
7. Karl, S.: Particle Animation and Rendering Using Data Parallel Computation.
Computer Graphics 24(4), 405–413 (1990)
8. McAllister, D.K.: The Design of an API for Particle Systems. Technical Report,
Department of Computer Science, University of North Carolina at Chapel Hill
(2000)
9. Burg, V.D.: Building an Advanced Particle System. Game Developer Magazine
(2000), http://www.gamasutra.com/features/20000623/vanderburg pfv.htm
10. Harris, M.: Real-Time Cloud Simulation and Rendering. PhD thesis, University of
North Carolina at Chapel Hill (2003)
11. Green, S.: Stupid OpenGL Shader Tricks (2003),
http://developer.nvidia.com/docs/IO/8230/
GDC2003 OpenGLShaderTricks.pdf
706 X. Cai, J. Li, and Z. Su
Hanwen Guo1,*, Minyong Shi2, Zhiguo Hong1, Rui Yang1, and Li Zhang1
1
Computer School, Communication University of China, Beijing 100024, China
2
Animation School, Communication University of China, Beijing 100024, China
Eniak_19@msn.com, myshi9613@hotmail.com
Abstract. This paper presents some algorithms for a novel simulation of the
folding and cut crafts in Chinese Paper-cut. All algorithms in simulation are de-
signed by consideration of the real Paper-cut’s whole process. We applies area
filling algorithm, polygon clipping algorithm, contour extraction algorithm and
subdivision curve algorithm, with the proper improvement that meets the tech-
nique of the art in the nature, to yield some vivid Paper-cut works. Our ap-
proach is feasible to anyone with the desire, due to Paper-cut illumines itself in
2D and all the operations are also in 2D.Moverover, recent compelling graphic
systems, such as Flash, Photoshop, Maya and 3DMax, did not provide a similar
interface for paper’s fold and decorative pattern which are primary elements in
Paper-cut. However, our approach not only meets the interest of the Paper-cut,
but also holds an interface for these compelling systems because of measuring
up the SVG.
1 Introduction
The Chinese civilization can be traced far back in ancient times, with a long history of
5,000 years. As the traditional Chinese painting and Chinese paper-cut branded with
"Chinese style", embed by lasting appeal, they are widely recognized and well re-
ceived by the world arts community and academia. However, because of the high cost
and long production cycle, it’s unable to meet the development requirements by using
the Chinese traditional art methods to produce works of art.
Paper-cut is one of the most popular folk art in China; according to the archae-
ology, its history can go back to the sixth century AD. As a form of folk art, paper-cut
inherits the traditional skills and modeling, and absorbs the experience of skills and
techniques accumulated by great people in China, flickering aesthetic concept for
working people.
*
Hanwen Guo was born in 1983 in China. Currently, he is a postgraduate student is Computer
Graphics in Communication University of China. His specific research area is computational
geometry and motion retargeting technique. Professor Minyong Shi, who was born in 1964 in
China, he received his PhD degree in Beijing Institute of Technology, his major research area
is Graph Theory.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 707–718, 2008.
© Springer-Verlag Berlin Heidelberg 2008
708 H. Guo et al.
Three main algorithms, as present in this paper, are Folding Algorithm, Cutting Algo-
rithm and Unfolding algorithm. The Reform Outline algorithm, which considered as
incidental, is employed by Unfolding Algorithm. The algorithms we have designed
are the improvements of many kinds of algorithms in the Computer Graphics, which
are belong to four main research directions, area filling algorithm, polygon cutting
algorithm, contour extraction algorithm and curve subdivision algorithm.
In the traditional raster display, polygon area filling is an important research area.
The representative filling algorithm based on the scanning line is [3].
The Sutherland[4] algorithm and Weiler-Atherton [5] algorithm are the famous
algorithms in the area of polygon clipping. Algorithm [4] is applied to rectangle poly-
gon, while algorithm [5] is applied to any convex and concave polygon.
Virtual Artistic Paper-Cut 709
The contour extraction algorithm can be roughly summed up as the following two
types: one is the Snake [6] algorithm and another one is Active Contour Models [7]
algorithm. This kind of algorithms first need the initial outline and then iteration so
that the contour gets close along the direction of the energy reduction, finally get an
optimization border.
Over the past few decades, researchers has proposed a lot of mature curve subdivi-
sion algorithms, such as the angle cutting curve subdivision algorithm early proposed
by Chaikin [8], de Casteljau algorithm [9], de Boor algorithm [10 ], four points curve
subdivision algorithm [11] proposed by Dyn and other researchers and Hermite sub-
division algorithm [12].
In our algorithms, we make some improvement of the polygon filling algorithm
base on the scanning line and apply it to the Folding Algorithm. Polygon clipping
algorithm, Weiler-Atherton, as the definition of the Cutting Algorithm, is improved to
be more suitable for the actual situation in the paper-cut. At the same time, we employ
contour extraction algorithm and curves subdivision algorithm for management of the
decorative patterns and complex shapes. The Unfolding Algorithm, in some words, is
the converse process of Folding Algorithm and the combination of the improved
Weiler-Atherton.
As shown in figure 1, the classes in our algorithms are: Paper, Polygon, FlodPolygon
and CutPolygon.
Fig. 1. Relation
Paper is a class equivalent to a container and it has a pointer pointing to the root
node of a polygon’s multi-tree. Class Paper has many member methods, such as Init
(Polygon * pg), Folding (Axis * axis), Unfolding (), and someone else.
In Paper::Init (Polygon * pg), Paper:: m_root_polygon will automatically point to
the pointer of polygon, and set itself as the root node of the folding polygon’s tree.
710 H. Guo et al.
Fig. 2. Mutil-Tree
c) Form the first point in InsertP, all point which are not visited and intersec-
tion points must be added as a circuit for generating new polygon P. New
polygon is stored in m_prChidArray. This description is reflected in Figure
3 step3 to setp4;
As depicted in Figure 3, 1 to 2 is logic rotation;2 to 3 is generation of InsertP: A B
C I5 D I4E F(I3) G I2 H I1 I and CrPntlist:I1 I2 I4 I5grouping into pairs;3 to 4 is new
polygons generation.
The method Folding(Axis* axis) of FoldPolygon and CutPolygon inherit form
Polygon::Folding(Axis* axis).Whenever there are new CutPolygon generated by fold,
a specific relation between new FoldPolygon and new CuttPolygon are estimated by
logic.
In the real situation of the Paper-cut, it is pretty obviously that one polygon’s edge
will overlap with another one’s edge or vertex will lies on the edge of another. WA
also gives out salvation aimed at this confused situation. Special vertexes and edges
need two more consideration listed below:
1. Edges belonged to subject polygon and overlapped with clip polygon need not be
taken into the calculation of intersection points;
2. Vertexes belong to edge of the subject polygon but lies on the edges of clip poly-
gon, if the edge they belong to is in the polygon, then it must be taken as intersec-
tion points, or else, they are not.
As described in Figure 4, the subject polygons are colored by red, and the clip
polygon is colored by blue. Taking point A as the intersection point, and not regard-
ing point B as intersection point, Algorithm would manage two polygons correctly.
Rectangle would be disposed by regarding points C and D as the intersection points.
Also, triangle would be disposed correctly, as if the condition of regarding E as inter-
section point were taken.
In our Algorithm, clipping method of WA is applied between the instances of
FoldPolygon and CutPolygon.
We considered instance of FoldPolygon as the subject polygon and instance of
CutPolygon is clip polygon. Still, new polygons are generated via the WA are regard
as the instances of CutPolygon. The new ones are stored in the m_CpgArray of the
subject polygon.
Combination method of WA is applied between CutPolyongs. In some peculiar
situation, after the combination, there would be some polygons appended “islands”.
Therefore, we add one step to handle this strange situation which never occurs in Pa-
per-cut. We delete polygon oriented in clockwise and only reserve one polygon in
counterclockwise. Figure 5 reflects what we describe above. Polygon i1c1c6i4s7s8s1s2
needs to be reserve and polygon i2s3s4s5i3c4c4 must be deleted.
Fold, Cutting and Unfolding are the three steps of the Paper-cut. They work in se-
quence in practice. Accordingly, all the Algorithms are designed in this order. First,
through the Folding Algorithm, it generates a multi-tree. Second, it is Cutting Algo-
rithm. The leaf nodes of the tree and instances of CutPolygon are involved by this
algorithm to generate some new polygons which regarded as instances of CutPolygon.
Last, it is Unfolding Algorithm. In Unfolding Algorithm, some of the instances of
714 H. Guo et al.
Fig. 5. Island
CutPolygon need the affine transforms due to the fold and they are not suitable for
Reform Outline Algorithm directly. It is because that some of them occupy the
edges of initial polygon described paper’s shape, and some does not. In another
word, some of the polygons could reshape the paper’s outline directly, while some
of them are regarded as decorative patterns called genus of the initial shape in glos-
sary of topology.
According Fig 6, the red rectangle is the initial outline of the paper. Blue polygons
are the ones which occupied the initial outline’s edge. Green ones and black ones are
the genus according to initial polygon. The figuration of the final artistic work, in
some ways, is the XOR between initial outline and Cutpolygon, actually. Some un-
wanted step could be skipped after the genus and none-genus are sorted.
Upon these reasons, we divide the results into two categories which are set A and
set B. The set A contains polygon which are not genus according to initial outline.
The genuses are aggregated in the set B. In Figure 6, the blue ones are in the set A,
the rest of them belong to set B.
Unfolding Algorithm:
step I Producing the set A and set B;
step II Taking each element in B for combination with every element in A, if the
combination is successful, redefine the elements both in A and B. Replacing the
polygon inhere in A by the new one generated after the combination deleting the
polygon inhere in B. Repeating this step until no more combination occurs, thus,
two new sets A* and B* are gave out;
step III Do the Reform Outline algorithm until all elements in A* are taken in;
step IV Reserving the elements of set B appears in final outline.
Virtual Artistic Paper-Cut 715
Fig. 7. Result
In Figure 6, after the setp II, the black polygon is absorbed in set A, then A* comes
out. After setp IV, the smaller green genus at the left bottom is deleted, and the left
one is reserved. Figure 7 is the final result.
It is Reform Outline Algorithm in setp III. This algorithm shares some inspirations
with the WA. As we write below, taking the outline P as the subject polygon and all
elements in A* as the clipping polygon :
step I Building two vertex lists of the two polygon respectively;
step II Calculating intersection points of the two polygons. Intersection points list,
two insert lists that are InsertP and InsertC, and one bidirectional list are gener-
ated.
step III Reshape the outline:
a) Building a new temporary vertex list;
b) Selecting one vertex that has never been visited and is not the intersection
point in InsertP, then add it into the temporary vertex list;
c) Adding the points in InsertP in the positive order (counter-clockwise), convert-
ing to InsertC when meet an intersection point, and then still add points but in
negative order (clockwise); whenever meets an intersection points, current in-
sert list must be converted. This step ended until the first point comes out
again. Marking the current polygon generated by these points as the P’;
d) Repeating the step a b c, until all the candidates suit step b are visited.
step IV For all P ′ , convert the original P by P ′ which has the largest area.
If there is any special situation, such as vertexes or edges occupy other edge should
follow the instructions gave out by WA.
In this section, we introduce our Unfolding Algorithm. Unfolding Algorithm is
consisted by improved WA applied in combination between CutPolygon’s instances
and Reform Outline Algorithm. The initial outline was decided by the point in Pa-
per::Init(Polygon* pg), alternatively, could be obtained through hand sketch or [5][6].
By [5][6], a consecutive points array will be obtained. We will refine consecutive
points array by applying [8-12]. Some of the parametric curves are also refined by
those algorithms. Vignettes are made by Bezier in manual or via automatic method.
716 H. Guo et al.
Fig. 8. On top of the interface is a control panel, In the middle is a work area, on the left side is
the finished modeling, on the right side is the decorative patterns
Following is the works created by artists using our paper folding and cutting system.
It is an Olympic Mascot in Figure 10.First, as we introduced above, the artist use
our system to make a semi finished articles, except the nose and eyes of the character,
due to unwanted “island” in our algorithms. Secondly, with the second disposal in
SVG Developer, the eyes and nose of this character are added as the pattern. All these
efforts make it more vivid than before.
In the following researches, we will focus on the folding and cutting algorithms of
paper with different materials. At present, our system can fold and cut papers of dif-
ferent colors. The system needs further improvement and expansions, so it can be
polished to manage papers of more than one material effect and manage paper by
more simulation algorithms. It means that we should take the needs of artists and
ordinary users into consideration. From the following aspects we can try our best to
achieve the perfect unity of art and computer simulation.
1. The realization of the paper coloring: including chromatic, coloring, color
separation, color contrast, dislocation stack liner, overturned and spray color,
and so on.
2. The realization of paper-cut and pasting. Those have not been achieved in our
system.
Therefore, the realization of paper coloring, paper-cut and pasting will be the goals
of my next step’s work.
Virtual Artistic Paper-Cut 717
Acknowledgement
The author would like to acknowledge the staff of Digital Technology and Digital Art
R&D Center and his supervisor prof. Minyong Shi. Also, this program is funded by
Technology Department, State Administration of Radio Film and Television (Grant
No. 2006-11).
References
1. Yu, J., Luo, G., Peng, Q.: Image Based Synthesis of Chinese Landscape Painting [J]. Jour-
nal of Computer Science and Technology 18(1), 22–28 (2003)
2. Rogers, D.F.: Procedure Elements for Computer Graphics. McGraw Hill, New York
(1985)
3. Sutherland, I.E., Hodgman, G.W.: Reentrant polygon clipping. Comm. ACM 17(1), 28–
352 (1974)
4. Weiler, K.J., Atherton, P.R.: Hidden surface removal using polygon area sorting. Com-
puter Graphics 11(2), 214–222 (1977)
5. Donna, W., Mubarak, S.: A fast algorithm for active contours and curvature estimation.
CV GIP: Image Understanding 55(1), 14–26 (1992)
6. Kass, M., Witkin, A., Terzopoulos, D.: Snake: active contour models. International Journal
of Computer Vision 1(4), 321–331 (1998)
7. Riesenfeld, R.F.: An algorithm for high speed curve generation[A]. Computer Graphics
and Image Processing[C] 3(4), 346–349 (1974)
8. Carstensen, C., Mühlbach, G., Schmidt, G.: De Casteljau’s algorithm is an extrapolation
method [J]. Computer Aided Geometric Design 12(4), 371–380 (1995)
9. de Boor, C.: Cutting corners always works [J]. Computer Aided Geometric Design 4(122),
125–131 (1987)
10. The, C.-H., Chin, R.T.: On the Detection of Dominant Pointson Digital Curves [J]. Trans
on Pattern Analysis and Machine Intelligence 11(8) (1989)
A Sufficient Condition for Uniform Convergence of
Stationary p-Subdivision Scheme
1 Introduction
Stationary subdivision schemes arise from modeling and interrogation of curves and
surfaces, image decomposition and reconstruction, and the problems of constructing
compact supported wavelet basis etc. [1, 9]. These schemes are being developed in
geometric modeling with great potentiality in CAD/CAM, computer graphics, image
processing, etc. [1-11, 14, 15]. Stationary subdivision schemes are widely used in
mechanical CAD, garment CAD, jewellery CAD, and applied in computer graphics.
They also play important roles in image coding, signal processing, and the construc-
tion of basis function of compact supported orthogonal wavelets by using multiresolu-
tion analysis [1-3, 17-23]. They are also important in fractal and its generation by
computer in particularly [1, 2, 4, 12]. Stationary subdivision schemes are used to
construct the required curves and surfaces from scattered data directly through stated
subdivision rules. Moreover the theoretical contribution of this approach consists in
their tight combination of three research disciplines: spline functions, wavelets and
*
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 719–727, 2008.
© Springer-Verlag Berlin Heidelberg 2008
720 Y.-K. Zhang et al.
fractals [1-4, 9, 10, 12]. Therefore, the research of stationary subdivision schemes,
especially its convergence, is significant in theoretical research and shape modeling
[2, 3, 5-13]. The idea and approaches of stationary subdivision schemes are still effec-
tive in subdivision surfaces [24, 25,26] and the constructions of compactly supported
orthogonal wavelets basis and fractal [11, 12, 13, 16, 24].
The systematic development of the basic mathematical principles and concepts as-
sociated with stationary 2-subdivision schemes is presented in [1]. The structure of
these algorithms in a multidimensional setting and convergence issue are researched
systematically. The complete theoretical system is constructed. The analytic structure
of limit curves and surfaces generated by these algorithms is revealed [1, 16].
The extension of stationary 2-subdivision to stationary p-subdivision scheme is
presented in [9]. Some properties of convergence of such schemes are described
through Fourier analysis, functional analysis and spline function. A sufficient condi-
tion of the uniform convergence of the stationary p-subdivision scheme is discovered
in [10] through a special polygon, δ-control polygon.
The problem is important of how to use this kind of subdivision schemes to gener-
ate curves and surfaces in computer graphics [1-8, 18-21]. The convergence of sta-
tionary subdivision schemes is a key problem in the theory of stationary subdivision
scheme and their applications [1, 9, 10]. Finding the features of the support set of
nonnegative mask a = {aα : α ∈ Z s } [1, 9, 10] has an important value in theoretical
researches and practical applications, because the convergence of these algorithms
can be judged directly in the construction of curves and surface. So, the sufficient
conditions of the uniform convergence of stationary p-subdivision schemes based on
the supported set of mask may promote theoretical researches and practical applica-
tions [1, 9, 10].
A sufficient condition of the uniform convergence of stationary p-subdivision
schemes is presented in this paper using contractility and spline function. This work is
based on three aspects: the nonnegative mask and its support set of stationary p-
subdivision schemes, some definitions and properties of stationary p-subdivision
schemes presented in [9, 10], and the works of [1-5].
Here is the main theorem of this paper.
Theorem. The stationary p-subdivision scheme
λ0 = λ , λm = Sλm−1 , m = 1, 2, L
defined in (2) is uniformly convergent, if the positive mask a = {aα : α ∈ Z s } sup-
ported on Ω satisfies
∑ aα − pβ = 1, α ∈ Z ,
s
β ∈Z s
where
s
Ω := Z( A)I Z s , zonotope Z( A) := { Au : u ∈ ∏ [li , ui ]} , li + p − 1 < ui , i = 1, L , s .
i =1
Definition 1. Let s be a fixed natural number and Z s the integer lattice, and
a = {aα : α ∈ Z s } be the fixed real scalar sequences having finitely supported set
suppa= {α : aα ≠ 0} . A stationary p-subdivision operator S is defined as
S : l ∞ ( Z s ) → l ∞ (Z s ) (1)
by
( Sλ )α = ∑ aα − pβ λβ , λ ∈ l ∞ (Z s ) ,
β ∈Z s
where p > 1 is a fixed natural number, and λ is point sequence.
Definition 2. Let S be any stationary p-subdivision operator defined in (1), the follow-
ing iteration scheme
λ0 = λ , λm = Sλm −1 , m = 1, 2, L (2)
is defined as a stationary p-subdivision scheme. a = {aα : α ∈ Z s } is referred to as the
mask of the stationary p-subdivision scheme S. λ is called as the control polygon of
S. In fact, λα is a vertex of the control polygon λ .
Stationary p-subdivision schemes and their some basic convergent properties are
presented in [9]. A sufficient condition for uniform convergence of stationary p-
subdivision schemes is given in [10] by using a special control polygon of δ-control
polygon.
D : l ∞ (Z s ) → R + U {0} is thought as a no-trivial non-negative functional in the fol-
lowing description.
Definition 5. A stationary p-subdivision operator S is said to be contractive relative to
D if there exists a constant number γ (0<γ<1) for the subdivision operator S defined
by (1), such that
D( Sλ ) ≤ γD(λ ) , λ ∈ l ∞ (Z s ) . (5)
Dρ (λ ) := sup λα − λβ (6)
ρ (α − β ) < 2
α , β ∈Z s
is defined as a diameter of λ .
The convergent condition of stationary p-subdivision scheme will be discussed in
the following under the condition of that the support of mask a = {aα : α ∈ Z s } is the
s
union of Z s and a special zonotope Z ( A) := { Au : u ∈ ∏ [li , ui ]} , where A is a s × s
i =1
integer matrix, and det A = −1 . The following two propositions are used to prove the
main theorem.
∑ aα − pβ = 1, for ∀α ∈ Z ; s
(8)
β ∈Z s
C1 λ ∞
≤ ∑ λαψ (• − α ) ≤ C21 λ ∞
, (11)
α ∈Z s ∞
Where C2 := ∑ ψ (• − α ) .
α ∈Z s ∞
(ii) Stationary p-subdivision operator S defined by (1) is contractive relative to
functional D.
(iii) There exists a constant C, such that
S λ − Bλ ∞
≤ C ⋅ D (λ ) , λ ∈ l ∞ ( Z s ) . (12)
The following two conclusions can be obtained
(a) The stationary p-subdivision scheme determined by S is uniformly convergent.
(b) If the condition (11) is replaced by the following condition that there exist a
constant C such that
λα − λβ ≤ C ⋅ D (λ ) , β − α ∈ (sup pψ ) o , (13)
the stationary p-subdivision scheme determined by S is also uniformly convergent.
Proof: For the definitions in (1) and (2), we can obtain that [10]
λαm = ( S m λ )α = ∑ aαm− p m β λβ and aαm = ( Sa m−1 )α = ∑ a βm−1aα − pβ . (14)
β ∈Z s β ∈Z s
If define f λm as follows
f λm ( x) = ∑ λαmψ ( p m x − α ), m = 1, 2, L ,
α∈Z s
σ i − δi
ρ (σ − δ ) := 2 max <2.
1≤i ≤ s u1i − li1
Without loss of generality, let σ > δ , then 0 ≤ σ − δ < u1 − l1 , and so
σ − u1 < δ − l 1 . Now to solve the pβ from the expression (16), the result
σ − u1 ≤ pβ ≤ δ − l 1 is obtained. Therefore there always exists a integer in interval
[σ − u1 , δ − l1 ] , so that this integer is pβ , so the β can be find out according to the
each coordinate component, and such β ∈ Z s is satisfies all the inequalities (15).
Therefore the operator S has following property known from the conclusion (i):
D ( Sλ ) ≤ γ ρ Dρ (λ ) , and 0 < γ ρ < 1 .
A Sufficient Condition for Uniform Convergence 725
∑ ϕ1 ( xi − α i ) = 1 , so ∑ ϕ ( x − α ) = 1 x ∈ R
s
Considering (17)
α∈Z s α ∈Z s
Moreover if we let
⎧⎪1 − 1 p j , j ∈ {−1,0,1}
bα = bα1 bα 2 L bα s , α = (α1 , α 2 ,L , α s ) , b j = ⎨ ,
⎪⎩0 , others
then the ϕ satisfies the p-scale equation: ϕ ( x) = ∑ bα ϕ ( px − α ) , x ∈ R s . (18)
α ∈Z s
Now select η = (η1 ,η 2 , L , η s ) , such that li < η i < ui , i = 1, 2, L , s , and let
ψ ( x) = ϕ ( x − η ) , then sup pb ⊆ sup pa , and mask b is associated to ψ .
Since ∑ aα − pβ = 1 = ∑ bα − pβ , for α ∈ Z s , then there is following conclusion:
β ∈Z s β ∈Z s
(Sλ)α − (Bλ)α = ( ∑ aα− pβ λβ − ∑ Caα− pβ ) + ( ∑ Cbα− pβ − ∑ bα− pβ λβ ) ,
β∈Zs β∈Zs β∈Zs β∈Zs
References
1. Cavareta, A.S., Dahmen, W., Micchelli, C.A.: Stationary subdivision. In: Memoirs of
American Mathematical Society, vol. 93(453), pp. 1–185. American Mathematical Soci-
ety, Providence (1991)
2. Prautzch, H., Micchelli, C.A.: Computing curves invariant under halving. Computer Aided
Geometric Design 4, 140–321 (1987)
726 Y.-K. Zhang et al.
3. Micchelli, C.A., Prautzch, H.: Computing surfaces invariant under subdivision. Computer
Aided Geometric Design 4, 321–328 (1987)
4. Micchelli, C.A., Prautzch, H.: Uniform refinement of curves. Linear Algebra and Its Ap-
plications 114/115, 841–870 (1989)
5. Chaikin, G.M.: An algorithm for high speed curve generation. Computer Graphics and Im-
age Processing 3, 346–349 (1974)
6. Riesenfeid, R.E.: On Chaikin algorithm. Computer Graphics and Image Processing 4, 304–
310 (1975)
7. Zhijie, C.: Convergence error estimation and some properties off four-point interpolation
subdivision scheme. Computer Aided Geometric Design 12(15), 459–468 (1995)
8. Dyn, N., Gregory, J.A., Levin, D.: A 4-point interpolatory subdivision scheme for curve
design. Computer Aided Geometric Design 4, 257–268 (1987)
9. Zhang, Y., Zhang, X., Tian, J.: Stationary p-Subdivision Algorithm (in Chinese). Journal
of Engineering Graphics 22(3), 92–101 (2001)
10. Zhang, Y., Gong, S., Dou, J.: A Study on the Convergence of p-Subdivision Schemes (in
Chinese). Computer Applications and Software 19(9), 43–45 (2002)
11. Dyn, N., Gregory, J.A., Levin, D.: Analysis of uniform binary subdivision schemes for
curve design. Constructive Approximation 7(2), 127–147 (1991)
12. Schaefer, S., Levin, D., Goldman, R.: Subdivision Schemes and Attractors. In: Desbrun,
M., Pottmann, H. (eds.) Eurographics Symposium on Geometry Processing 2005. Euro-
graphics Association 2005, Aire-la-Ville, Switzerland. ACM International Conference
Proceeding Series, vol. 255, pp. 171–180 (2005)
13. Song, L., Xinlong, Z.: Multivariate refinement equation with nonnegative masks. Science
in China Series A: Mathematics 49(4), 439–450 (2006)
14. Hongqing, Z., Jun, H., Zhenglin, Y., Guohua, P., Shuili, R.: 4-point Subdivision Scheme
with Three Parameters. Progress in Intelligence Computation & Applications. In: The 1st
International Symposium on Intelligence Computation and Applications, ISICA 2005,
vol. 1, pp. 409–412. Springer, Heidelberg (2005)
15. Micchelli, C.A., Sauer, T.: Sobolev Norm Convergence of Stationary Subdivision
Schemes. In: Le Méhauté, A., Rabut, C., Schumaker, L.L. (eds.) Surfaces Fitting and Mul-
tiresolution Methods, pp. 245–260. Vanderbilt University Press, Nashville (1997)
16. Xinlong, Z.: Characterization of Convergent Subdivision Schemes. Analysis in Theory and
Applications 14(3), 11–24 (1998)
17. Liu, W., Kondo, K., Zhang, L.: An Approximating Subdivision Scheme by Cutting and
Upheaving the corner of Control Polygon. In: Proceedings of the 11th International Con-
ference on Geometry and Graphics, pp. 134–138. Moscow State University of Technology
Press, Moscow (2004)
18. Farin, G.: Curves and surfaces for computer aided geometric design: A Practical Guide,
2nd edn. Academic Press, Inc., Boston (1990)
19. Denis, Z., Schroder, P. (eds.): Subdivision for modeling and animation. SIGGRAPH 2000
Course #23 Notes. ACM Press, New York (2000)
20. Peter, S., Denis, Z.: Subdivision for modeling and animation. SIGGRAPH 1998 Course
#36 Notes. ACM Press, New York (1998)
21. Breen, D.E.: IEEE Computer Graphics and Applications. In: Computer Graphics in Tex-
tiles and Apparel, vol. 16(5), pp. 26–27. IEEE Computer Society, Washington, DC (1996)
22. Peter, S., Sweldens, W. (eds.): Wavelets in computer graphics. SIGGRAPH 1996, Course
#13 Notes. ACM Press, New York (1996)
A Sufficient Condition for Uniform Convergence 727
23. Felmi, C., lortiz, M., Schröder, P.: Subdivision surfaces: A new paradigm for thin-shell fi-
nite element analysis. International Journal for Numerical Methods in Engineering 47(12),
2039–2072 (2000)
24. Dyn, N., Gregory, J.A., Levin, D.: Analysis of uniform binary subdivision schemes for
curve design. Constructive Approximation 7(1), 127–147 (1991)
25. De Villiers, J.: On refinable functions and subdivision with positive masks. Advances in
Computational Mathematics 24(1-4), 281–295 (2006)
26. Chen, F., Ding, Y., Liu, J., Wei, D.: A Novel Non-stationary Subdivision Scheme for
Geometric Modeling. In: Proceedings of the Fourth International Conference on Computer
and Information Technology (CIT 2004), vol. 00, pp. 748–752. IEEE Computer Society,
Washington (2004)
Model and Animate Plant Leaf Wilting
1 Introduction
The literature of realistic modeling of leaves has a long history in computer graphics.
This is, in part, due to their either beautiful or colorful images. Partly they have a strong
visual effect on the audience. Many techniques have been proposed for modeling the
geometry or shape of leaves. Most of these methods, however, just describe the shape of
leaves in regular way.
Some researchers have endeavored to generate curled shapes of plant leaves. Prus-
inkiewicz et al. [1] provided a detailed representation of combining interaction and
parameterized algorithms for realistic plant modeling and scene creation involving
plants, including curled leaves. Mündermann et al. [2] proposed a method for modeling
lobed leaves. Effects of curled leaves could be generated by using free-deformation in
their framework. Recently, Sung et al. [3] proposed an interactive method for modeling
curled leaf surface. But it could involve excessive manual interactions for generating a
desired curled shape of leaf by using their method. Studies on curvature of plant leaves
from biophysical viewpoints have raised the question of what role, if any, genes play in
the control of curvature [4]. Yet others study wave or wrinkle pattern in leaves with
physical analysis [5]. But these may beyond our focus in this paper.
*
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 728–735, 2008.
© Springer-Verlag Berlin Heidelberg 2008
Model and Animate Plant Leaf Wilting 729
Additionally, there has been a great deal of previous work on the simulation motions
of plant, including plant growth, motion in the wind and so on, such as the work demon-
strated in [6]. Some physically-based models had been used to create natural plant shape
[7, 8]. Recently Wang et al. [9] had simulated physically the growth of a plant leaf, the
physical model used in the simulation is the governing equations of fluid mechanics –
the Navier-Stokes equations. But the leaf model in their simulator was 2D.
Basing on the fact that less work has focused on modeling leaf surface deformation
and simulating subtle behaviors of plant, such as wilting of leaves suffering from insuf-
ficient water supply, this paper presents a venation skeleton-based deformation method
for plant leaves, and aims to develop an approximately kinematic model of leaf for
simulating motions of plant leaves, especially wilting. The leaf skeleton plays two roles.
It is used to generate a venation skeleton for later deformation. A geometric mesh for
the leaf surface is also constructed from the leaf skeleton, and each vertex in the mesh is
mapped to its nearest vertex in the venation skeleton, as such the geometric mesh is
deformed according to the deformed venation skeleton. Applications of our approach to
simulate wilting plant leaves with realistic results illustrate that the proposed model is
flexible and effective.
techniques [10]. For our needs, we develop an interface for generating venation skele-
ton from a leaf skeleton interactively. As Fig.1 shows, the leaf skeleton can be ob-
tained from a scanned image, a venation skeleton is then generated from the leaf
skeleton. Each vein is segmented. Fig.1 (c) (the red line) shows a generated venation
skeleton consisting of one mid-vein and four secondary veins, the blue vertices string
segment each vein into several line sections. The process of generating venation
skeleton involve several interactive manipulation including defining the start point
and end point for each vein, and setting parameter for segmenting each vein. And
users can decide how the mid-vein crosses the leaf skeleton, and how many secondary
veins are attached to the mid-vein.
Skeleton-based methods have been widely used in computer animation and computer
modeling for mesh deformation. Traditional skeleton-based methods commonly re-
quired a tedious process of weight selection to obtain satisfactory results. Note that
the natural deformation of plant leaves is different from the deformation of humans’
or animals’ organs, we can constraint the movement of vertices in the venation skele-
ton. The most regular motions of plant leaves which can be seen by naked eyes are
curling and wilting. The major goal of our approach is to develop an approximately
kinematic model of plant leaf for simulating these motions.
(a) The movement locus of a skeletal joint (b) Child joints follow the parent’s movement
For our needs, we restrict the movement of each vertex in the venation skeleton by
rotating it around a fixed vector. Fig.2 illustrates how a venation skeleton works. For
convenience, the venation skeleton includes a vein, which consists of four segments.
The light black vertex serves as the root node. Note that the movement of a leaf blade
is always downward in the course of wilting, as such each joint in the skeleton seg-
ment does spherical movement. Take vertex Vi for example, it will align gradually
vector Vi-1VN in the course of wilting, in which Vi-1VN is a downward vector reverse to
Z axis, while vector Vi-1VM is perpendicular to the plane contains vector Vi-1Vi and Vi-
1VN. The movement of vertex Vi can be looked as Vi rotating round vector Vi-1VM.
To obtain a motion sequence of a vertex in a skeleton, a simple method is to rotate
the vertex with a fixed angle, such as θ in Fig. 2(a), and the angle is commonly given
Model and Animate Plant Leaf Wilting 731
by the users. We have mentioned before that a vertex in the skeleton will do spherical
rotation during the wilting of its corresponding leaf surface. As Fig. 2(a) illustrates,
we can calculate the new position of vertex Vi by the following parametric equation:
V( t ) = k( t ) * ( Vi + t * ( VN - Vi )) (1)
Where k( t ) = |Vi| / |Vi + t * ( VN - Vi )| ( 0≤ t ≤1). Further, the motion sequence of Vi
can be obtained by increasing parameter t. This may simplify the rotation operation.
As we know, all child segment of a vertex in the skeleton will follow the move-
ment of the vertex, this can be done by passing a displacement and rotation angle to
its child vertices when the vertex is rotated. Fig. 2(b) gives results after moving Vi
to V( t ).
We have constructed the leaf skeleton with two boundary curves and a mid-vein curve
as shows in Fig.1 (b). To mesh the void within these curves, we employ Delaunay
triangulation scheme. The reason for using Delaunay triangulation is to deal with the
problem of existing concave region in the leaf skeleton. When using Delaunay trian-
gulation, we can use directly the feature points in the mid-vein and silhouette, or ex-
tract a series of points from the mid-vein curve and silhouette curves with a fixed
interval. The mesh showed in Fig.3 (a) is generated from Fig.1 (b) by using Delaunay
triangulation scheme.
The initial mesh of a leaf surface generated by using Delaunay triangulation is gen-
erally irregular and rough. It is necessary to refine the mesh for later deformation.
Currently, we use a simple method to subdivide the initial mesh, which provides sev-
eral parameters to satisfy requirements for users’ interaction. Fig.3 (b) illustrates a
mesh subdividing the initial mesh shows in Fig.3 (a).
We have detailed the mechanism for controlling movement of the venation skeleton
in our approach and methods for constructing the leaf mesh. The last step is to deform
732 S. Lu et al.
the leaf surface basing on the deformed venation skeleton. This process can be illus-
trated as Fig. 4.
Firstly all the vertices in the subdivision mesh of the leaf are banded to the initial
venation skeleton. The banding is based on the distance of each vertex to the venation
skeleton. Then the initial venation skeleton is deformed by using the method de-
scribed in Section 4.1. For example, we can generate the shape of the venation skele-
ton showed in Fig. 4(b) from the initial skeleton showed in Fig.1(c) (with different
number of joints in each vein). Lastly the position of each vertex in the mesh is recal-
culated according to the new coordinate of its banded vertex in the venation skeleton.
Fig. 4c displays the resulted mesh. Fig. 4(d) demonstrates the rendering result. The
texture mapping is calculated before the deformation.
It needs to be noted that the number of joints in each vein in the venation skeleton
will influence greatly the effects of deformation. The larger the number of joints is,
the smoother the deformed surface will be. And large deformation needs large number
of joints. But larger number of joints also means more computation and more difficult
controlling. Users can obtain a satisfactory result by interactive experiment.
Constraints and collision detection are usually the common issues in surface deforma-
tion. For constraint, we have stated that each vertex in the venation skeleton can rotate
around a fixed vector. Additionally, the rotation needs to satisfy some extra con-
straints. In simulating the effect of wilting leaf surface, a vertex in the leaf mesh could
not be rotated after it had reached the maximum drooped distance. When simulating
curling of a leaf, it needs to avoid overlap of the leaf surface. This can be done by
keeping the included angle of two adjacent line sections on each vein being larger
than a pre-defined angle.
For collision detection, currently we just consider a collision detection to avoid self-
Intersect in the deformation of a leaf mesh. During deforming a leaf mesh, each han-
dling currently vertex needs to be checked if its movement will pierce some triangle in
the mesh. If no piercing occurs, no response is calculated. If there is an intersection,
then we calculate a maximum displacement from the pre-calculated displacement for
Model and Animate Plant Leaf Wilting 733
the vertex to move while avoid intersection, and correct the displacement of corre-
sponding vertex in the venation skeleton.
The application examples above demonstrate that the proposed venation skeleton-
driven approach for simulating wilting of leaf surface is effective and flexible, it can
generates realistic effects of wilted leaves similar to natural shape. Currently generat-
ing the venation skeleton is manual and interactive in our framework, and controlling
the motions of leaves in the scale of a plant is still simple. In fact, wilting of leaves
are natural response for plant to adapt themselves to the environment basing on their
inner state. An attractive area for future work might involve combining our dynamic
modeling technique with physiological model of the leaf. In addition, we just consider
a plant or a leaf in our framework. It is desirable to simulate the motions of plant
leaves in an ecosystem scale.
5 Conclusion
We have presented a model for modeling wilted leaf surface. This model deforms a
leaf surface by driving a venation skeleton which is embedded into the geometric
mesh of a leaf. The venation skeleton can be created from any polygonal mesh of leaf
surface, whereas the polygonal mesh can be captured from real leaves, which makes it
easy to create highly realistic leaf appearance models. Furthermore, our model pro-
vides an approximately kinematic model of plant leaves for simulating subtle motions
of plants.
It needs to be noted that motions of plant leaves would result from a series of com-
plex reasons, as such the mechanism of motions of leaves is not easy to model. The
leaf deformation model presented in this paper is an example of a model that provides
Model and Animate Plant Leaf Wilting 735
intuitive control for the simulating of some motions of plant leaves. An exciting area
for future work is the development of a framework for virtual agronomic experiment
for broader classes of plants.
Acknowledgments. This work is supported by National High Tech R&D Program of
China under grant No. 2007AA10Z226, Beijing Natural Science Foundation of China
under grant No. 4081001, and the National 11th Five-year Plan for Science & Tech-
nology of China under grant No. 2006BAD10A07.
References
1. Prusinkiewicz, P., Mündermann, L., Karwowski, R., Lane, B.: The Use of Positional In-
formation in the Modeling of Plants. In: Proceedings of SIGGRAPH, Miami Beach, FL,
USA, pp. 289–300 (2001)
2. Mündermann, L., Macmurchy, P., Pivovarov, J., Prusinkiewicz, P.: Modeling Lobed
Leaves. In: Proceedings of Computer Graphics International, CGI (ed.) Tokyo, Japan, pp.
60–65. CGI (2003)
3. Sung, M.H., Simpson, B., Gladimir, V.G.B.: Interactive Venation-Based Leaf Shape Mod-
eling. Computer Animation and Virtual Worlds 16(3-4), 415–427 (2005)
4. Nath, U., Crawford, B.C.W., Carpenter, R., Coen, E.: Genetic Control of Surface Curva-
ture. Science 299, 1404–1407 (2003)
5. Sharon, E., Roman, B., Harry, L.: Geometrically Driven Wrinkling Observed in Free Plas-
tic Sheets and Leaves. Physical Review E 75, 1–7 (2007)
6. Beaudoin, J., Keyser, J.: Simulation Levels of Detail for Plant Motion. In: ACM SIG-
GRAPH/Eurographics, Symposium on Computer Animation, pp. 297–304 (2004)
7. Jirasek, C., Prusinkiewicz, P.: A Biomechanical Model of Branch Shape in Plants. In:
Lantin, M. (ed.) Proceedings of the western computer graphics symposium, Whistler, Can-
ada, pp. 23–26 (1998)
8. Hart, J., Baker, B., Michaelraj, J.: Structural Simulation of Tree Growth and Response.
The Visual Computer 19, 151–163 (2003)
9. Wang, I., Wan, J., Baranoski, G.: Physically-Based Simulation of Plant Leaf Growth.
Computer Animation and Virtual Worlds 15, 237–244 (2004)
10. Amenta, N., Bern, M., Kamvysselis, M.: A New Voronoi-Based Surface Reconstruction
Algorithm. In: Proceedings of ACM SIGGRAPH, Orlando, Florida, USA, pp. 415–421
(1998)
The Technical Research and System Realization of 3D
Garment Fitting System Based on Improved
Collision-Check Algorithm
Introduction
3D garment fitting system is the result of the science’s development. The way to buy
clothes is changing while the science technology and the internet are developing. At
first, buyers needed to select clothes by fitting it on body in shops, then 2D level’s
garment fitting systems appeared which used picture-paste fashion to simulate the
effect of fitting. Now, various 3D garment fitting systems have appeared at home and
abroad. All of these represent the significance of technology in human life.
3D garment fitting system includes the human body modeling, the 3D garment
modeling and the simulation function. 3D garment modeling includes two phases: first,
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 736–744, 2008.
© Springer-Verlag Berlin Heidelberg 2008
The Technical Research and System Realization of 3D Garment Fitting System 737
Whether we can Real-timely and efficiently control the internet garment fitting system
lies on the rate of the internet. It means that the increasing of the internet rate is
important. If the internet delay always happens, the Real-time effect of the fitting
would not be acquired and we would waste a lot of time in the waiting of updating 3D
pictures. As a result, we use the C/S mode for the system. Users only need to download
Client and garment patches and run them at local computers. In this way, it can solve
the problems generated by slow internet rate.
2 Main Processes
There are two main ways to establish human body models: building models based on
measure-rebuild method and building models based on imported the 3DS ‘s mannequin.
Model-building based on measure-rebuild method: the best strongpoint of it is the
accuracy and the controllability of the human body size. But there also exists shortages.
It’s ok to use human platform instead of human body models in the design of the
garment structure. But human platform can’t replace human body models in some areas
because of the singular problem of motion and pose and the lacking problem of face
and extremities generated by human platform modeling. Such as the stage-exhibition of
738 Q. Chen et al.
the garments in virtual reality and the dynamic garment fitting function. In this paper,
we design an interface faced to 3DS files to import human body models in 3DS
format[2][3][4][5]. Imported the 3DS‘s mannequin translates the human body models
established by Poser and saves them as 3DS files, then it analyzes their structure,
imports them, and builds the human body models by rebuilding the 3D human bodies
according to the information of the 3D coordinates.
(I) We can design two classes to read 3DS files according to the analyzing of the
“block” structures of 3DS files[7]. The two classes can be described as Fig. 2.
(II) After importing the 3DS files, we can build human body models by using lots of
triangle planes according to the vertexes and the normal vectors read.
The mapping process from 2D patches to 3D garments is a complex and flexible de-
formation process. It needs to meet the following conditions[8]: first, it should keep the
area of the patches during the mapping process; second, patches between each other
should satisfy the correct relations; third, no collision happens during the mapping
process. The frequently used modeling method based on physics. It contains
Flexible-Distortion model, Particle-System model, Finite Elements method,
Spring-Mass model and so on. After analyzing we found that the Spring-Mass model is
simpler and it can acquire more real simulation effect and has higher simulation speed.
The fabric deformation model we built is based on the Spring-Mass model built by X
Provot. 2D patches and 3D garments are all dispersed and are expressed as
Spring-Mass system composed by regular triangle grids. The vertexes of the grids are
particles. The sides of the grids are springs. Each particle connects with particles
around it by springs. The relationship between particles is the stretching effect.
According to the mechanism capability of the fabric, the springs can be divided into
The Technical Research and System Realization of 3D Garment Fitting System 739
three types: structural spring, shearing spring and flexion spring[9]. They can be
described as Fig.3. The patch could be constructed by these 3 types of springs as de-
scribed by Fig.4.
2.4 Collision-Check
The collision contains the fabric collision and the collision between parts of the clothes.
It means that self-collision-check has two parts[10]. The usual 3D models (contain the
740 Q. Chen et al.
fabric and the objects around it) are expressed as triangle grids. In this way, the process
of the collision-check is just the way to check if any penetration happens between
particles and triangles and if it happens between sides of triangles. But, if we check
each particle-triangle pair or we check each pair of sides, the capacity of the
computation is too bulky. In order to reduce the bulkiness, we use a method based on
AABB hierarchical bounding volumes[11] to exclude the pairs of particle-triangle and
the pairs of sides that would not intersect with each other. We built a AABB tree for
human body models(as Fig.4 describes) and patches and then we just check the pairs of
particle-triangle and pairs of sides that intersect in hierarchical bounding volumes.
(4) If the sum of ∠AMC , ∠CMB, ∠BMA is 360o , then the cross point M is surely
uuuuur
inside the triangle ABC. We could make conclusion that the vector pop1 is colliding
with the triangle ABC.
742 Q. Chen et al.
3 Conclusion
The purpose of the 3D garment fitting system is to realize the simulation for the 3D
human body models and the 3D garments. The main processes of it are 3D human body
modeling, changing the plane structure from 2D patches to the solid structure of 3D
garments and the simulation for the garment fitting according to the material capability
of the fabric. This paper discusses the key technology of the fitting system’s
constructing in detail. It contains 3D human body modeling and the importing, the
mapping model of 2D patches and 3D garments, the improved collision-checking
technology and responding process. The 3D garment fitting system described in this
paper has the characteristics of rapid updating rate and real simulation effect. It can
realize more complex garment style. We trust that this system could excite the potential
of sell effectively after it is perfected and is extensively used.
Acknowledgement
Supported by Program for New Century Excellent Talents in Fujian Province Univer-
sity and the 2006 Foundation of Department of Science & Technology in Fujian Pro-
vincial (2006H0035).
744 Q. Chen et al.
References
[1] McCarney, J., Hinds, B.K., Seow, B.I., Gong, D.: Dedicated 3D CAD for garment mod-
eling. Journal of Materials Processing Technology 107, 31–36 (2000)
[2] Dongmei, Y., Shengyuan, Z., Weicheng, L.: The integration of OpenGL and 3D Studio
MAX to realize 3D simulation. Application Technology (2), 33–35 (2004)
[3] Bin, F.: The application of 3DMAX models in OpenGL. Journal of GuiZhou Industrial
University (6), 45–49 (1999)
[4] Wenguang, Z., Zhongxue, L., Cuiping, L.: The Visualization simulation of VC++ faced to
engineering,The integrative technology of OpenGL and 3DS. Journal of BeiJing Tech-
nology University (6), 53–565 (2001)
[5] Zhen, Z.: Poser4 3D Cartoon Design-God of War, vol. 1. National defense industry press,
Beijing (2004)
[6] Chen, S.F., Hu, J.L., Teng, J.G.: A finite-volume method for contact drape simulation of
woven fabrics and garments. Finite Elements in Analysis and Design 37, 513–531 (2001)
[7] Peng, W.: Studies on Constructing Mannequin and Fashion Pattern Design Technology in
Virtual Reality. Master’s degree thesis of JiangNan University, 35
[8] Fan, J., Zhou, J., Wang, Q.F., Yuan, M.H.: 2D/3D isometric transformation using
spring-mass system. Journal of Software[J] (in Chinese with English Abstract) 10(2),
140–148 (1999)
[9] Hui, L., Chun, C., Bole, S.: Simulation of 3D Garment Based on Improved Spring-Mass
Model. Journal of Software[J] (14), 620–621 (2003)
[10] Changfeng, L., Yi, X.: Cloth 3D Real-Time Simulation. Journal of Computer-Aided design
& Computer graphics[C] (18), 1375 (2006)
[11] Chengying, G., Ning, L., Xiaonan, L.: The establishment of fabric model and the dealing
with conllision-check in virtual fitting. Computer Applications (5), 34–37 (2002)
[12] Provot, X.: Collision and self-collision handling in cloth model dedicated to design gar-
ments[C]. In: Proceedings of Graphics Interface 1997, Kelowna, pp. 177–189 (1997)
[13] Louchet, J., Provot, X., et al.: Evolutionary identification of cloth animation models[A]. In:
Terzopoulos, D., Thalmann, D. (eds.) Proceedings of the Computer Animation and
Simulationc 1995[C], pp. 44–54. Springer, New York (1995)
[14] Hongyan, S., Jun, L.: The developing trend of garment internet shopping. Chemical Fi-
ber&Textile Technology (3) (2005)
[15] Kang, Y.M., Choi, J.H., Cho, H.G.: Fast and stable animation of cloth with an approxi-
mated implicit method. In: Thalmann, N.M. (ed.) Proceedings of the Computer Graphics
International 2000, pp. 247–255. IEEE Computer Society Press, Los Alamitos (2000)
[16] Breen, D.E.: Computer graphics in textiles and apparel modeling. IEEE Computer
Graphics and Applications 16(5), 26–27 (1996)
[17] Ng, H.N., Grimsdale, R.L.: Computer graphics techniques for modeling cloth. IEEE
Computer Graphics and Applications 16(5), 28–41 (1996)
[18] Hadap, S., Bangerter, E., Volino, P., Thalmann, N.M.: Animating wrinkles on clothes. In:
Ebert, D.S., Gross, M., Hamann, B. (eds.) Proceedings of the IEEE Visualization 1999, pp.
175–182. IEEE Computer Society Press, Los Alamitos (1999)
[19] Volino, P., Thalmann, N.M.: Implementing fast cloth simulation with collision response.
In: Thalmann, N.M. (ed.) Proceedings of the Computer Graphics International, pp.
257–266. IEEE Computer Society Press, Los Alamitos (2000)
[20] Okabe, H., Imaoka, H., Tomiha, T., Niwaya, H.: Three dimensional apparel CAD system.
Computer Graphics 26(2), 105–110 (1992)
Reconstruction of Tree Crown Shape from Scanned Data
1 Introduction
With the current development of virtual environment establishment, product design,
digital entertainment, antique protection, and city programming 3D geometry model
construction and processing is now an active development area. 3D geometry modeling
is regarded as the fourth digital multimedia in addition to digital audio, digital image,
and digital video. 3D geometry models are normally used to represent object surface to
identify extendedly shape and appearance attributes.
With the advancement of 3D scanning technology, more and more 3D digital
scanners are popularly used for different applications. Rich details of the object shape
can be acquired from scanned data with dense sampling points (point cloud), where no
topological connection relations are included. It becomes important to develop new
processing methods to represent, to process, to reconstruct and to render these highly
complex geometric bodies. Reconstruction of geometry model is one of the important
research topics in modern virtual reality.
Trees are typical objects in virtual reality, so it is very important to reconstruct and to
represent the real trees. Tree reconstruction can be used in many applications, including
digitization of vegetation scenes, design of a new scene, digital entertainment, and so on.
*
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 745–756, 2008.
© Springer-Verlag Berlin Heidelberg 2008
746 C. Zhu et al.
Reconstruction of the tree crown shape is useful to model a real tree and in various
research fields interested by growth simulation of virtual trees for light interception,
biomass evaluation, and so on. Many researches have been carried on surface recon-
structions, but the shape of a tree crown is more difficult than that of a usual solid object
in its heavy occlusions and its high complexity in geometry and topology.
Reconstruction of a real tree crown from scattered scanned points is a new challenge
in virtual reality. The points are unstructured, unevenly distributed, and sampled from
non-manifold shapes; it is very difficult to define typical boundary points in a tree
crown in accordance with the visual perception. Rich concavity is another feature of the
crown shape of a real tree. Classical surface reconstruction techniques do not work for
such tree crown data. Other difficulties are that the data have no topological informa-
tion among points, their 3D distribution is not even at all, and they are not complete; so
it becomes rather difficult to reconstruct it with sufficient details.
Alpha shape is a new technique [1] in the classification of all the simplexes from 3D
Delaunay triangulation of a 3D point set, and the result of this classification is three
categories: the internal simplexes of the shape, the regular one and the external one.
With a proper heuristic alpha value specified by the user [1], a concave silhouette shape
of a point set sampled from a regular manifold surface can be constructed.
We will improve this approach to the point cloud data acquired from the scan on a
real tree. Because of the limitations of alpha shape technique and complexity of
scanned tree data, it is not possible to have all details of the tree crown reconstructed to
a regular mesh through a direct application of the alpha shape technique [1]. In this
paper we solve this problem using a range of alpha values and testing the close property
of the constructed mesh, so that the mesh model is a concave closure of the point data.
The structure of this paper is as follows. Related work in shape information analysis
and plant modeling is introduced in section 2. Fundamental knowledge of our method is
described in section 3. Technical details of this new approach are described in section 4.
Experiments of this technique to reconstruct tree crown shapes are shown in section 5.
Conclusions about this technique and further investigation are described in section 6.
2 Related Work
In the past decades, many methods have been developed on point shape processing and
shape modeling of complex objects including plants, but with unequal results.
Jarvis [2] was the first to consider the problem of computing the shape as a generali-
zation of the convex hull of a planar point set. A mathematically definition of the shape
was developed by Edelsbrunner in [3]. For 3D points, Boissonnat [4] suggested to use
Delaunay triangulation to “sculpture” a single connected shape of a point set.
In the frame of projects such as the digital Michelangelo project [5] at Stanford
Computer Graphics Lab in the 2000’s, and with the improvement of computer hard-
ware, a numerous number of research papers have been published on point cloud
processing and rendering. Point geometry processing and analysis became an active
research topic.
Reconstruction of Tree Crown Shape from Scanned Data 747
The different approaches of 3D tree model construction can be roughly classified into
three categories: botanical models, geometrical models, and digitized models from real
plants.
There are a numerous methods to simulate real plant appearance. Many early
methods were based on rule iterations (botanical, physical, geometrical, mathematical),
or simply based on strong user control with advanced dedicated patterns. In the 1980s,
modeling by botanical rules appeared, and produced nice findings, researchers tried to
simulate the growth of natural plants, plants could be constructed by some botanical
rules or grammars. AMAP [6] modeling method is based on bud life cycles of botanical
knowledge with real measurement data (on plant topology). This modeling method
clear reflects the growth mechanism of plants, including space occupation and the
location of leaves, fruits, and flowers. L-systems presented by a Lindenmayer and
Prusinkiewicz were broadly applied to describe the growth process of plant organs,
which were based on fractal pattern [7, 12].
GreenLab [13] modeling approach is put forward as a mathematical model, which
simulates interactions of plant structure, leaves, trunk, branch and function. This model
can exactly engender the dynamics of plant, architecture and geometry of woody plants,
because of internal competition for resources, leaves sizes are different, and growth of
pruning can also be simulated.
These methods, used mainly in biology research fields are not dedicated to control
the 3D plant shape, and cannot easily do it, but aim to understand plant shape as the
result of a dynamic. It as be cited that this kind of model is not suitable to construct a 3D
models of real tree by using botanical methods [10].
Geometrically interactive modeling is another way to model virtual plants. Although
this method does not strictly follow the botanical rules, but visually realistic trees can
be produced [14]. In general, given 3D skeleton points of real plant, 3D model of each
branch can be generated with generalized circular cylinders [15]. Prism model is a
simplified application of this method. This approach is widely applied in some plant
software such as Xfrog, if combining rule-based method with traditional geometric
modeling approach. Nice 3D plant model could be produced, such as flowers, bushes,
and trees [8, 14].
To summarize, these rule based or pattern based methods used to build the real plant
faithful to botanic knowledge or appearance, can produce visually very realistic plants,
although they could not be used to model a specific existing real plant.
New modeling methods have been developed to digitalize real plants in very recent
years [9-11]. These methods can be used to reconstruct the trunk, the branches, and the
leaves, but the realism of the reconstructed model is still different from the real shape
due to the lack of crown silhouette shape information.
Plant digitization aims to reconstruct the shape of real plants from the information
digital instruments. The most popular techniques are the use of 3D laser scanner [9] or
the use of digital photos [10].
748 C. Zhu et al.
When scanning a real plant from a single viewpoint many occlusions occur. In par-
ticular, many leaves do usually hide branches and other organs from the view. One way
to make reconstruction efficient is to work both on plant branching structure recon-
struction and on plant crown reconstruction. The idea of the proposed approach is to
consider that when processing the branch reconstruction of a real plant, we must con-
strain the silhouette of branches from the crown shape.
By scanning a real tree, we have a point cloud data set, from which we could re-
construct the shape of the real tree crown by combining existing methods.
Considering the branch structure, we may underline several interesting works.
Cheng [16] reconstructs a real tree from a range image, using generalized circular
cylinders to fit incomplete data and compute the skeleton based on axis direction.
Pfeifer [17] introduces a set of algorithms for automatically fitting and tracking cyl-
inders along branches and reconstructing the entire tree.
With the appearance of advanced precise digital camera and laser scanner, the de-
velopment of digital plant is accelerated. Image-based and laser-scanning based
methods have come up to produce 3D model of real trees in nature. Shlyakhter [18]
builds a 3D model of tree from a set of photographs. His method constructs the visual
hull of tree first, then a plausible skeleton is built up from medial axis of visual hull, and
finally L-system is applied to construct branches and leaves. Teng [19] reconstruct 3D
trunk of plant only from two images, this method only estimates skeleton and radius of
branches roughly. Quan [10] also models a plant from digital image. Their work focus
on reconstruction of big leaves, branches are reconstructed by interaction.
These image-based approaches can build 3D plant from images of different view-
points, but because of inevitable noise of images and error of camera parameters, the
accuracy of those methods is limited.
The approaches of Xu [9, 11] are based on some prior knowledge. A skeleton is first
constructed by connection of the centroids of points, which have an analogous length of
the shortest path to a root point. Then the corresponding radius of skeleton nodes could
be computed by the allometric theory. Leaves are constructed in the end, so that the
reconstructed tree is visually impressive.
However during the reconstruction, the methods of imaged-based or 3D laser
scanning data based first construct the skeleton, and then construct leaves, but because
of much occlusion, the reconstructed skeleton is incomplete. 3D laser scanner could not
scan the thin branches because of its limited precision.
But we have to reconstruct these thin branches for the architecture shape of real
plants in botany and digital forestry and for high visual impression in virtual reality.
We must thus construct the shape of tree crown to constrain the reconstruction of
thin branches.
A set P of points can be used to construct a complex if the points do not lay in a plane.
Delaunay triangulation is a natural choice to do it. In literature, different Delaunay
triangulation techniques are proposed [20-22], where Lawson flip method is a typical
one. in Lawson’s method, the tetrahedron bounding the point set P is constructed at
first, and the other points are inserted into the triangulation one by one then. Each time,
the triangulation is optimized to satisfy the Delaunay property: the circumsphere of
every tetrahedron does not contain any other points. Those tetrahedrons, which do not
satisfy a local Delaunay property, are flipped.
The flip process in 3D can be described as follows. The triangulation in 3D is a set of
tetrahedrons constructing a simplical complex. We will explain the case of two tetra-
hedrons incident to a triangle ace (Figure 1). If the circumsphere of tetrahedron
aecd does not contain b and circumsphere of tetrahedron aecb does not contain d , it
can be said that (the triangle) aec is local Delaunay. Otherwise, this situation can be
modified inserting a new edge bd inserted. Therefore the complex is a Delaunay tri-
angulation.
The result of Delaunay triangulation of the point set is its convex hull composing
several tetrahedrons.
The concept of alpha-shapes formalizes the intuitive notion of shape for spatial point
sets on user’s selection. Alpha-shape is a mathematically well-defined generalization of
the convex hull. Its result is a series of subgraphs of the Delaunay triangulation, de-
pending on different alpha values. Given a finite point set, a family of simplexes can be
computed from the Delaunay triangulation of the point set, a real parameter alpha
controls the desired level of detail. All real alpha values lead to a whole family of
shapes. The alpha-shape of a point set is made up of the set of points, edges, triangles
and tetrahedrons, which satisfy the constraint condition: the alpha test [1]. This test
applies for each a triangle t of the triangulation. If t is not on the boundary of the
convex hull, there must be two tetrahedrons p, q , which are incident to t . Tetrahedrons
p and q are tested to be in the circumsphere of t or not. If they both are not in that
circumsphere, and the radius of the circumsphere is less than the alpha value, t is said
to satisfy alpha test, and it is regarded as one member of the alpha shape. So al-
pha-shape is a subset of the triangulation.
If we let alpha be large enough, the shape is the convex hull of the points set. If alpha
approaches 0, no tetrahedral, triangles and edges could pass the alpha test, so the alpha
750 C. Zhu et al.
shape is the points set. With the adjustment of the alpha values, this subset can follow
the topology of the points set. So, if we choose a proper value for alpha, we will find a
reasonable surface for a tree crown.
The alpha shape is a sub-complex of the Delaunay triangulation of the points set P .
This can be explained in the following. There is a ball eraser with alpha as its radius,
and it could move to all possible positions in the 3D space and with no point of P in-
cluded. This eraser will delete all simplexes whose size is bigger than alpha and it can
pass through. So the remaining simplexes construct the alpha shape.
It is an ordinary technique to sample the surface of real object using 3D laser scanner,
and then to reconstruct the shape from the sample data with limited precision. This
point cloud data describe the geometry and the appearance attribute of objects surface.
The normal point cloud is densely sampled from continuous or smooth surface, al-
though the data is unorganized and irregular. A number of successful methods have
been presented to deal with these data and to reconstruct appearance of object.
Plants, such as trees, have too many organs and its structure is too complex. A tree is
made up of trunk, branches, and a huge number leaves. The point cloud data of tree is
not sampled from a manifold surface, so it is more irregular than those from other data
from the manifold surface. The points from leaves are even more irregular. The density
variation of point cloud from leaves may be very large. Thus, traditional technique does
not work for these objects. Special methods should be developed to reconstruct real
plants. In order to keep the shape of plants, branch skeleton extraction and construction
of plant crown should be included. One difficulty of this work is that the points from
Reconstruction of Tree Crown Shape from Scanned Data 751
branches and those from leaves are mixed together, so that it is hard to initialize the
work of shape analysis.
From the above analysis and the range image data acquired from a single scan in Figure
3(a) and Figure 4(a), we can recognize the dense region and the sparse region of the
data by observation, but this recognition process is very difficulty to be performed in a
computer. The points from the tree side facing the scanner and the region with dense
leaves (the side of a tree facing the sun, for example) are denser. There may be some
interstices among dense leaves. When we scan a tree, laser lights will pass the interstice
and meet the branch or leaves at another side of tree, or pass through the tree. So there
should be holes in the data. Although we can distinguish the dense region, the sparse
region, the convex region and concave region, the algorithms processing very dense
point set will make mistakes in topological reconstruction. Therefore, we must con-
struct topological structure of points at first, where Delaunay triangulation is an ideal
choice.
Our algorithm contains four steps:
The first step is to triangulate point set P = { pi } with Delaunay triangulation, so that
a set of connected tetrahedrons T = {T j } are obtained. Flipping method in [23] is
adopted to correct irregular triangulation in P = { pi } . All tetrahedrons T = {T j } will
constitute a convex solid, the shell of this solid is a convex hull.
The second step is to compute all radii, R(T j ) , of circumsphere of every tetrahedron
after triangulation. This value will be one attribute of a tetrahedron T j . The radii, r ( Fk ) ,
of the circumcircle of each face of a tetrahedron arecomputed also, and they are thought
of as an attribute of each face.
The third step is to classify tetrahedrons {T j } and their all faces. The rule to classify
all T j is the size of R(T j ) . This classification is performed by the relation of R(T j ) with
threshold α and, where α is specified by users. The scope of α should be proper.
Then all tetrahedrons are classified into two categories according to a real value α :
interior tetrahedrons and exterior tetrahedrons. If R(T j ) > α , T j is classified as an ex-
terior tetrahedron. Otherwise, it is classified as an interior tetrahedron. All faces {Fk }
from each tetrahedron T j are classified into three categories also: interior faces, exterior
faces and boundary faces. The classification role is as follows. If a face on the convex
hull belongs to an exterior tetrahedron, it is an exterior faces; otherwise, if it belongs to
an interior tetrahedron, it is a boundary faces. For each face not on the hull, if it is an
intersection face of two exterior tetrahedrons, it is an exterior face. If it is an intersec-
tion face of two interior tetrahedrons, it is an interior face. If it is an intersection face of
one interior tetrahedron and one exterior tetrahedron, it is a boundary face. All
boundary faces will construct a mesh, and this mesh M will be an concave approxi-
mation of the crown.
Let rmax be the largest radius of all R(T j ) and all r ( Fk ) , and Let rmin be the
smallest radius of all R(T j ) and all r ( Fk ) . We acquire an interval [ A, B] , where
A = λ r min , B = μ r max , λ = 0.9 , and μ = 1.1 . The α value should be confined to the
752 C. Zhu et al.
interval [ A, B] ; otherwise, if α > B , the mesh M will be a convex hull, and if α < A , the
mesh M will not be a solid.
The fourth step is to test the validity of specific alpha values, so that the mesh
M builds a boundary surface of a manifold. If the alpha value is set larger than B ,
boundary points are on the convex hull, so the mesh cannot be concave. If the alpha
value is set smaller than A , some sample points are isolated from in the solid, so the
reconstructed shape is not complete. Those both extreme cases are not interesting for
tree crowns. Therefore, α must lay in interval [ A, B] . Finding the proper α value is an
iterative process. We initialize α as the average value of A and B . In each iteration
step, we check if the boundary triangles constitute a manifold surface; if so, the alpha
value can be reduced, if not, it is increased.
Figure 2 shows the pipeline of this approach.
Fig. 4. Extraction of the crown shape of a Candlenut tree; (a) is the source point cloud data
displayed with a cube for each point; (b) is a comparison of the extracted boundary mesh with the
source point cloud date; (c) is a close view of (b)
We reconstruct the shape of tree crowns with two data sets of two trees. The first
one is a single scan of a 20-meter high maple tree with leaves. Figure 3 (a) shows the
original point model of the maple tree of 114997 points displayed with a cube for each
point. When the alpha value is set as 4.2354, we acquire 2810 points on the boundary
(Figure 3 (b)). Figure 3 (c) shows the reconstructed tree crown mesh model.
The second example is a candlenut tree without leaves shown in Figure 4. The
original data of the candlenut tree has 86675 points (Figure 4 (a), and when the alpha
value is set as 0. 41399, 4291 points are left on the boundary. The implementation of
our algorithm is shown in Table 1, where the last column is the time spent from data
input, to Delaunay triangulation, and to the list of all triangular faces on the boundary.
To show the properness of this approach, the original point model of the candlenut
tree is combined to its reconstructed crown mesh model. Figure 4 (b) shows this
comparison, and Figure 4 (c) shows a close view of Figure 4 (b). It can be seen in
Figure 4 (b) and Figure 4 (c) that the reconstructed crown mesh model includes the
original point model well.
These two examples show that the shape concavity is well reconstructed. The ap-
proach is illustrated here on both dense crown and spare (unfoliaged) one.
6 Conclusion
Current 3D acquisition systems lead to model more and more 3D shapes of real life
objects. However, nowadays reconstruction approaches classically fail on high com-
plexity objects, such as trees. Even if nice progresses have been noticed on the main
branch structure on un-foliaged trees, the overall reconstruction is not satisfactory,
especially on small structures and leaves.
We proposed hereby a method to reconstruct in 3D the scanned tree crown, in order
to constrain the definition of the branch structures, especially the thinner ones, and
contribute to define local geometrical constraints for leaf area reconstruction.
The principle of our approach is based on the use of the alpha-shape on the range
point data set, a generalization of the convex hull and subgraph of the Delaunay tri-
angulation. In the Delaunay triangulation process, we choose the triangle candidates on
the boundary according to the alpha value, and constrain the surface mesh to stay a
manifold. Therefore, our constructed boundary mesh builds in fact the silhouette of the
crown. This shape of the tree crown is much more convincing than the convex hull of
the tree crown in keeping the major concave features of the crown. This shape can be
used to constrain faithfully the reconstruction of branches and foliage.
The proposed approach was successfully implemented and tested on two data sets.
Of course, the reconstructed crown shape mesh is rough, thus fast to render, and thus
not strongly concave, so that higher branching structures are not recreated. In future,
progress can be achieved by dividing the data into several subsets according to point
density, with different alpha values applied to each subset. Concave silhouette surfaces
can then be reconstructed independently, and then merged to a more detailed shape.
It is also interesting to note that such crown shapes do find applications in various
domains. Such tree crown can contribute to define intermediate LOD plant models,
from real plants or simulated ones. It contributes to define low weighted geometrical
models. Of course, appropriate color and transparency value computations can increase
the appearance while rendering such shapes.
Finally, the proposed technique may be of interest on a wide range of complex object,
showing high topological complexity, where simplified representation, based on in-
ternal complex structure is useful. Such could be the case of human organs represen-
tation build from their internal vessels.
Acknowledgments
The authors would like to show thanks to Dr. Baoquan CHEN and Ms. Wei Ma for
providing scanned data of trees. CGAL library was used for Delaunay triangulation
[24]. This work is supported in part by National Natural Science Foundation of China
with projects No. 60073007, 60672148, 60473110; in part by the National High
Reconstruction of Tree Crown Shape from Scanned Data 755
References
[1] Edelsbrunner, H., Mucke, E.P.: Three dimensional alpha shapes. ACM Trans. Graph. 13,
43–72 (1994)
[2] Jarvis, R.A.: Computing the shape hull of points in the plane. In: Proceedings of IEEE
Computer Society Conference on Pattern Recognition and Image Processing, pp. 231–241.
IEEE, New York (1977)
[3] Edelsbrunner, H., Kirkpatrick, D.G., Seridel, R.: On the shape of a set of points in the
plane. IEEE Trans. Inform. Theory 29(4), 551–559 (1983)
[4] Boissonnat, J.D.: Geometry structure for three-dimensional shape representation. ACM
Trans. Graph 3, 266–286 (1984)
[5] Levoy, M., Pulli, K., Curless, B., Rusinkiewicz, S., Koller, D., Pereira, L., Ginzton, M.,
Anderson, S., Davis, J., Ginsberg, J., Shade, J., Fulk, D.: The Digital Michelangelo Project:
3D Scanning of Large Statues. In: Proceedings of ACM SIGGRAPH 2000, pp. 131–144
(2000)
[6] de Reffye, P., Edelin, C., Françon, J., Jaeger, M., Puech, C.: Plant models faithful to bo-
tanical structure and development. In: SIGGRAPH 1988 Proceedings of the 15th annual
conference on Computer graphics and interactive techniques, pp. 151–158. ACM Press,
New York (1988)
[7] Prusinkiewicz, P., Lindenmayer, A.: The algorithmic beauty of plants. Springer, New York
(1990)
[8] Deussen, O., Lintermann, B.: A modeling method and user interface for creating plants. In:
Conference on Graphics interface 1997, Toronto, Ont., Canada, pp. 189–197 (1997)
[9] Xu, H., Gossett, N., Chen, B.Q.: Knowledge-based modeling of laser scanned trees. In:
SIGGRAPH 2005: ACM SIGGRAPH 2005 Sketches, p. 124. ACM Press, New York
(2005)
[10] Quan, L., Tan, P., Zeng, G., Yuan, L., Wang, J.D., Kang, S.B.: Image-based plant mod-
eling. ACM Trans. Graph. 25(3), 599–604 (2006)
[11] Xu, H., Gossett, N., Chen, B.Q.: Knowledge and heuristic-based modeling of laser-scanned
trees. ACM Trans. Graph. 26(4), 19 (2007)
[12] Prusinkiewicz, P., James, M., Mech, R.: Synthetic topiary. In: SIGGRAPH 1994: Pro-
ceedings of the 21st annual conference on Computer graphics and interactive techniques,
pp. 351–358. ACM Press, New York (1994)
[13] Counède, P.H., Kang, M.Z., Mathieu, A., Yan, H.P., Hu, B.G., de Reffye, P.: Structural
factorization of plants to compute their functional and architectural growth, Simulation.
Transactions of the Society for Modeling and Simulation International 82(7), 427–438
(2006)
[14] Lintermann, B., Deussen, O.: Interactive modeling of plants. IEEE Comput. Graph.
Appl. 19(1), 56–65 (1999)
[15] Bloomenthal, J.: Modeling the mighty maple. In: SIGGRAPH 1985: Proceedings of the
12th annual conference on Computer graphics and interactive techniques, pp. 305–311.
ACM Press, New York (1985)
[16] Cheng, Z.L., Zhang, X.P., Chen, B.Q.: Simple reconstruction of tree branches from a single
range image. J. Comput. Sci. Technol. 22(6), 846–858 (2007)
756 C. Zhu et al.
[17] Pfeifer, N., Gorte, B., Winterhalder, D.: Automatic reconstruction of single trees from
terrestrial laser scanner data. In: Proceedings of 20th ISPRS Congress, pp. 114–119 (2004)
[18] Shlyakhter, I., Rozenoer, M., Dorsey, J., Teller, S.: Reconstructing 3d tree models from
instrumented photographs. IEEE Comput. Graph. Appl. 21(3), 53–61 (2001)
[19] Teng, C.H., Chen, Y.S., Hsu, W.H.: Constructing a 3D trunk model from two images.
Graph. Models 69(1), 33–56 (2007)
[20] Edelsbrunner, H.: Algorithms in Combinatorial Geometry. EATCS Monographs on
Theoretical Computer Science, vol. 10. Springer, Heidelberg (1987)
[21] Preparata, F.P., Shamos, M.I.: Computational Geometry: an Introduction. Springer, New
York (1985)
[22] Dey, T.K., Sugihara, K., Bajaj, C.L.: Delaunay triangulations in three dimensions with fi-
nite precision arithmetic. Comput. Aided Geom. Design 9(6), 457–470 (1992)
[23] Edelsbrunner, H., Shah, N.R.: Incremental toplogical flipping works for regular triangu-
lations. In: Proceedings of 8th Annual ACM Symposium on Computational. Geometry, pp.
43–52 (1992)
[24] CGAL - Computational Geometry Algorithms Library, http://www.cgal.org/
A Survey of Modeling and Rendering Trees
1 Introduction
Reconstruction of natural scenes has always been the main purpose of computer
graphics. Especially, many visualizations applications, such as virtual environment
and computer game, occur in natural scenes with trees. In recent years, coupled with
the rapid development of Computer Science and technology, it becomes probable for
us to create extremely complex outdoor scenes routinely. Use of hardware accelera-
tion also propels the development of rendering trees in realtime. However, trees
which are quit different from general regular objects because of their particular prop-
erties, are often hard to be constructed with a convincing effect.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 757–764, 2008.
© Springer-Verlag Berlin Heidelberg 2008
758 Q.-L. Zhang and M.-Y. Pang
3 Classification
Those methods used for representations of trees can be categorized according to dif-
ferent dimensions as follows:
(1) Off-line or realtime rendering based on rendering time. (2)static representations
or dynamic representations based on adaptive complexity. (3)polygon, image, point
set and volume based on render primitive.
There is no single criterion to survey all methods used in modeling and rendering
of trees. Based on several criteria above, we will review and discuss various typical
methods used in modeling and rendering of trees. Coupled with the distance change
between observers and objects in many visualizations applications, appearance of
trees may change a lot. The farther the distance is, the more details trees lose. In close
view, trees appear with a distinct trunk, main branch and leaves. In far view, trees
appear only with a trunk and outline, even a pixel point on screen. Thus, our classifi-
cation is first related to the level of details (LOD) of trees representations. Usually,
static representations mean a fixed LOD which is only appropriate for single view
(close or far view), while dynamic representations mean multiresolution or adaptable
LOD which is often appropriate for a certain range of views. Obviously, they need
different methods to build respectively.
Commonly, static representations adressing a fixed LOD may use various primitives,
such as polygon, image, volume and point set, etc.
speed up rendering. Since the foliage represents the majority of a tree’s geometric
complexity, groups of leaves or small branches can be approximated as a single, tex-
ture mapped polygon [1].
calibrated pictures. This opacity value is associated to a set of textures extracted from
photographs. For the rendering of the whole tree, the algorithm traverse the cells
back-to-front, and render the billboard which is textured by the blend of the two tex-
tures corresponding to the two closest views for each cell. Nevertheless, the method
just works for the tree with sparse leaves, rather than dense forest, because single
pixels contain the blended projection of numerous leaves/branches and background.
Another limitation is that capturing the top of the trees is difficult, which is important
in landscape rendering with flying movement.
initial subdivision mesh for arbitrarily complex branching structures. The author in
[13] proposes another improved approach for mesh refinement and growth. And RFL
recently apply it to modeling leaves growth [14].
combine two slicing methods to render volumetric textures efficiently: a simple one
used for most locations, and a more complex one used at silhouettes.
According to the complexity and relative size of objects, the space of object is di-
vided into several cubic units [27], [28], [7]. Model in the same cubic unit is replaced
with special primitive. The advantage of this method is converting any model input
without considering topolopy information. Nevertheless, the model have a poor visual
quality in view and difficulty in error control. As preservation of topological relation-
ship can not be guaranteed.
4 Conclusion
Various methods are used to represent realistic trees. Static representations are com-
mon addressing several trees representations with fixed level of complexity. Polygon
based methods propose interesting techniques, but rely on geometry modification
which can produce unadapted results at highest simplification rates. Point based
methods offer interesting support to achieve realtime rendering but require hardware
support. IBR methods which are usually used for realtime rendering seem to be lack
of reality. With the purpose of reducing complexity of natural scenes with trees, Dy-
namic representations try to improve static representations by defining adaptive
model. However, the definition of such progressive simplification scheme is complex
on sparse model like trees.
In sum, none of these techniques perfectly answers all. Nevertheless, interesting
compromise between realism and efficiency has been proposed and rendering of large
landscape is possible. In fact, some points such as animation or trees diversity stay an
issue. Moreover, large number of data lead to memory issue. And the massive use of
instantiation result in a lack of diversity. As a result, the idea of on the fly data gen-
eration begin to be exploited. Point based methods that take full advantage of hard-
ware acceleration seem to be a promiseful method.
Acknowledgments. This work was partly supported by the Natural Science Founda-
tion of the Jiangsu Higher Education Institutions of China (Grant No. 07KJD460108),
the Outstanding High-end Talent Foundation of Nanjing Normal University (Grant
No. 2007013XGQ0150).
References
1. Interactive Data Visualization, Inc. Speedtree product homepage. Web page (2002),
http://www.idvinc.com/speedtree/
2. Siggraph Course Notes CD-ROM. Advanced Graphics Programming Techniques Using
OpenGL. Addison-Wesley (1998)
3. Pulli, K., Cohen, M., Duchamp, T., Hoppe, H., Shapiro, L., Stuetzle, W.: Viewbased ren-
dering: Visualizing real objects from scanned range and color data. In: Eurographics Ren-
dering Workshop (1997)
4. Meyer, A., Neyret, F., Poulin, P.: Interactive rendering of trees with shading and shadows.
In: Eurographics Workshop on Rendering (2001)
A Survey of Modeling and Rendering Trees 763
5. Jakulin, A.: Interactive vegetation rendering with slicing and blending. In: de Sousa, A.,
Torres, J.C. (eds.) Proceedings of Eurographics (2000)
6. Qin, X., Nakamae, E., Tadamura, K., Nagai, Y.: Fast photorealistic rendering of trees in
daylight. In: Computer Graphics Forum Proceedings of Eurographics (2003)
7. Reche, A., Martin, I., Drettakis, G.: Volumetric reconstruction and interactive rendering of
trees from photographs. In: ACM Transactions on Graphics, SIGGRAPH 2004 Conference
Proceedings (2004)
8. Reeves, W.T., Blau, R.: Approximate and probabilistic algorithms for shading and render-
ing structured particle systems. In: ACM Transactions on Graphics (SIGGRAPH 1985
Conference Proceedings), pp. 313–322 (1985)
9. Remolar, I., Chover, M., Belmonte, O., Ribelles, J., Rebollo, C.: Geometric simplification
of foliage. In: Eurographics 2002 Short Presentation Proceedings, Saarbrcken, Germany,
pp. 397–404 (2002)
10. Remolar, I., Chover, M., Ribelles, J., Belmonte, O.: Viewdependent multiresolution model
for foliage. Journal of WSCG (WSCG 2003 Proceedings) (2003)
11. Zhang, X.P., Blaise, F.: Progressive polygon foliage simplification. In: Hu, B., Jaeger, M.
(eds.) Plant Growth Modeling and Applications (Proceedings of PMA 2003), Beijing,
China (2003)
12. Tobler, R.F., Maierhofer, S., Wilkie, A.: Mesh-based parametrized l-systems and general-
ized subdivision for generating complex geometry. International Journal of Shape Model-
ing (2002)
13. Smith, C., Prusinkiewicz, P., Samavati, F.: Relational specification of subdivision algo-
rithms. In: Pfaltz, J.L., Nagl, M., Böhlen, B. (eds.) AGTIVE 2003. LNCS, vol. 3062, pp.
313–327. Springer, Heidelberg (2004)
14. Runions, A., Fuhrer, M., Lane, B., Federl, P., Rolland-Lagan, A., Prusinkiewicz, P.: Mod-
eling and visualization of leaf venation patterns. In: ACM Transactions on Graphics (SIG-
GRAPH 2005 Conference Proceedings), vol. 24(3), pp. 702–711 (2005)
15. Weber, J., Penn, J.: Creation and rendering of realistic trees. In: ACM Transactions on
Graphics (SIGGRAPH 1995 Conference Proceedings), pp. 119–128 (1995)
16. Deussen, O., Colditz, C., Stamminger, M., Drettakis, G.: Interactive visualization of com-
plex plant ecosystems. In: Proceedings of the IEEE Visualization Conference, vol. 19,
IEEE, October RESEARCH REPORT RR-LIRIS-2006-003, (2002)
17. Lluch, J., Camahort, E., Vivo, R.: Procedural multiresolution for plant and tree rendering.
In: Proceedings of the 2nd international conference on Computer graphics, virtual Reality,
visualisation and interaction in Africa, pp. 31–38. ACM Press, New York (2003)
18. Di Giacomo, T., Capo, S., Faure, F.: An interactive forest. In: Cani, M.-P., Magnenat-
Thalmann, N., Thalmann, D. (eds.) Eurographics Workshop on Computer Animation and
Simulation, pp. 65–74. Springer, Manchester (2001)
19. Behrendt, S., Colditz, C., Franzke, O., Kopf, J., Deussen, O.: Realistic real-time rendering
of landscapes using billboard clouds. In: Eurographics (2005)
20. Colditz, C., Coconu, L., Deussen, O., Hege, H.: Real-time rendering of complex photoreal-
istic landscapes using hybrid levelof-detail approaches. In: Real-time visualization and
participation, 6th International Conference for Information Technologies in Landscape Ar-
chitecture (2005)
21. Interactive Data Visualization. Speedtree product homepage. web page (2002),
http://www.idvinc.com/speedtree/
22. Software Blueberry 3D, http://www.blueberry3d.com
23. Max, N., Deussen, O., Keating, B.: Hierarchical image-based rendering using texture map-
ping hardware. In: Eurographics Workshop on Rendering (1999)
764 Q.-L. Zhang and M.-Y. Pang
24. Decaudin, P., Neyret, F.: Rendering forest scenes in real-time. In: Keller, A., Jensen, H.W.
(eds.) Eurographics Symposium on Rendering (2004)
25. Marshall, D., Fussell, D., Campbell III, A.T.: Multiresolution rendering of complex bo-
tanical scenes. In: Davis, W.A., Mantei, M., Klassen, R.V. (eds.) Graphics Interface 1997,
pp. 97–104. Canadian Human-Computer Communications Society (1997)
26. Kajiya, J.T., Kay, T.L.: Rendering fur with three dimensional textures. ACM Transactions
on Graphics (SIGGRAPH 1989 Conference Proceedings) 23(3), 271–280 (1989)
27. Neyret, F.: Modeling, animating, and rendering complex scenes using volumetric textures.
IEEE Transactions on Visualization and Computer Graphics 4(1), 55–70 (1998)
28. Meyer, A., Neyret, F.: Interactive volumetric textures. In: Eurographics Workshop on
Rendering, pp. 157–168 (1998)
Creating Boundary Curves of Point-Set Models
in Interactive Environment
1 Introduction
In past decade, with wide use of 3D data acquisition devices, lots of point-
set models [1] arises in application fields, such as digital geometry processing,
reverse engineering and virtual reality, etc. At the same time, a variety of tech-
niques for processing point-set models have been proposed by computer graphics
scientists, e.g., the algorithms for triangulation of point clouds [2] and surface re-
construction from point-sampled models [3][4], etc. Usually, the scattered points
of a point-set model were sampled from surface of certain physical model by 3D
scanner. Thus, the point-set model is a discrete representation of the continuous
solid surface. With the enhance of computer processing ability, especially for the
use of high-performance graphics acceleration hardware (e.g., GPU), performing
various operations on the scatted point-set models directly rather than triangu-
lating them into meshes or piecewise linear surface firstly is becoming more and
more popular [2][5]. These operations include texture-mapping, shape-editing,
animation controlling, and so on. As the base of the operations, feature extrac-
tion, mesh-less parameterization and segmenting point-set models [6] have been
becoming important topics of digital geometry processing.
Corresponding author.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 765–772, 2008.
c Springer-Verlag Berlin Heidelberg 2008
766 P. Xiao and M.-Y. Pang
2 Algorithm
The goal of this paper is to offer user an interactive operating mechanism, such
that user can deeply participate in extracting boundary curve of a point-sampled
model. They can use interactive devices, typically, mouse or keyboard to select
the boundary vertices from models interactively, and generate and manage the
boundary automatically. Obviously, the boundary can be taken as a kerf, which
divides a genus-0 model into two divided parts.
where each node in the list represents one boundary vertex. For the convenience
of discussion, we simply name the list data structure above boundary or boundary
loop in this paper.
In interactive environment, with using mouse or other devices, user can pick up
some points from a point-sampled model to define a boundary. When gradually
adding new point/vertex to the current boundary, one can increase more details
of the boundary. If a vertex has been in the boundary, clicking it again by
mouse may means removing the vertex from the boundary. Thus, the boundary
creating progress involves two basic operations: a) add new vertex to the current
boundary; and b) remove a selected vertex from the boundary.
Before adding a new vertex to the boundary, the algorithm needs to perform
an important step to determine appropriate position in current boundary loop,
where the new vertex will be inserted to. This task can be solved by calculating
the distance between the new vertex and polyline of the current boundary. In
this paper, the distance is defined as the shortest one of the distances from the
new vertex to all line segments in the boundary curves (the line segments will
be called boundary line segments or BLS in next sections). Our algorithm will
minimize the length of the boundary polyline as optimizing strategy of creating
boundary.
As a basic calculating step of our algorithm, we now discuss how to calculate
the distance from an arbitrary space point to a BLS of the boundary. Here, let
us suppose a and b are the two adjacent vertices in the current boundary and s
is the space point in 3D space (see Fig. 1).
p = a + vt (1)
According to (1), the distance between p and s can further be described such
as
−b (a − s) · (b − a)
t= = (4)
2a (b − a)2
In this case, the function D(t) get its minimum. Substitute (4) to (2), the
distance between s and L can be evaluated on the fly. However, D(t) must not
be the distance between s and ab in deed. Three special cases must be further
analyzed as follows (see Fig.1(b) to 1(d)):
Insert new vertex to boundary loop. Once the distances between s and
BLS’s in current boundary loop are calculated, we can find the shortest one. If
the distance exactly corresponds to one BLS, we denote the BLS by vi vi+1 , the
work of adding the new vertex becomes to simply insert the vertex between vi
and vi+1 . Otherwise, there must be several BLS’s corresponding to the shortest
distance, see Fig. 2(b) for a reference. In this case, inserting the new vertex to
the boundary loop is needed to be farther treated carefully:
As showed in Fig. 2(a), assume vi−1 , vi and vi+1 are neighboring vertices in
the current boundary loop. There are two planes passing vi and perpendicular to
vi−1 vi and vi vi+1 respectively, which bound a space domain illustrated by the
Creating Boundary Curves of Point-Set Models in Interactive Environment 769
shadowed area in the figure. If s is in the domain C, the distance between s and
vi−1 vi is equal to the distance between s and vi vi+1 . In this situation, according
to Equ.(4), we know ti−1 > 1 for vi−1 vi and ti < 0 for vi vi+1 . This result
can inversely be used to decide if s is in the shadowed area or not. Moreover,
we denote by π the plane, which equably bisect the angle ∠vi−1 vi vi+1 (see
Fig.2(a)). If s and vi−1 are in the same side of π, insert s into vi−1 vi , otherwise
insert s into vi vi+1 . This approach can avoid serious distortion of boundary
curve in local area. If s is exactly in the plane π, it is needed first to compare the
distances from s to vi−1 and vi+1 , if ||s − vi−1 || ≤ ||s − vi+1 ||, insert s between
vi−1 and vi , otherwise, between vi and vi+1 . By this way, we can make the
boundary curve as short as possible.
In the process above, evaluating the equation of plane π is needed. Since
vi is a point of π, the only thing for evaluating the equation is to decide the
normal n of π. In the plane defined by vi−1 vi vi+1 , the direction of the bisector
of vi−1 vi vi+1 can be expressed as:
vi−1 − vi vi+1 − vi
m1 = +
||vi−1 − vi || ||vi+1 − vi ||
On the other hand, the normal of plane vi−1 vi vi+1 is:
vi−1 − vi vi+1 − vi
m2 = ×
||vi−1 − vi || ||vi+1 − vi ||
Obviously, n//(m1×m2 ), that is to say, the equation of plane π can be written
as:
vi−1 −vi vi+1 −vi vi−1 −vi vi+1 −vi
+ × × · (v−vi ) = 0
||vi−1 −vi || ||vi+1 −vi || ||vi−1 −vi || ||vi+1 −vi ||
An extreme case is illustrated in Fig.2(b), in which the shortest distance
corresponds to a set of BLS’s. Here, let Ω is the set consisted of the BLS’s. The
first step we do is to find the BLS’s that satisfy t ∈ [0, 1] in Equ.(4) from Ω.
For example, v1 v2 and v5 v6 in the figure. And then, in these BLS’s, to search
the BLS, which satisfies that the distance sum from s (v in the figure) to its
two ends is minimum. Without loss of generality, we denote the found BLS by
vk vk+1 , we now can insert the new vertex between vk and vk+1 to keep the
total length of the boundary optimized.
If there is no BLS satisfied t ∈ [0, 1] in Ω, we can assert that each BLS in Ω
belongs to the pattern illustrated in Fig.2(a), i.e., each BLS has one end apart
from s the shortest distance. In this situation, evaluate distances from s to the
other end of each BLS, then take the shortest one as the candidate BLS and
insert the new vertex between its ends.
mouse, we can delete it from the boundary. In this case, we only need to remove
the corresponding node from the list.
skills. The data structure and related interactive operation are encapsulated
in separate classes, so the functions of our implementation is independence to
particular window system. On different GUI platform, one can manage the point-
set boundary curve through calling the interface functions of the classes of the
algorithm.
In our experiments, we take GLUT library as window system to visualize the
experimental results. Through the link between the algorithm and the callbacks
of display window, keyboard and mouse, our tester can carry selecting and delet-
ing of the boundary vertex. At the same time, we use OpenGL to accomplish
3D transformation of point-set models [8]. Picking and selecting functions are
implemented by OpenGL API.
Fig.3 shows some results of creating boundary curves interactively based on
our algorithm. Fig.3(a) shows the boundary curve generating result of the apple
model which is composed by 867 sample points, the boundary divided the apple
model into two parts. Fig.3(b) use the algorithm creating manhead model section
curve, this curve can be used as the boundary of plane parameter domain during
point-set model parameterizing. Fig.3(c) selects the nose region of the manhead,
the related boundary curve can be used in local editing of the model. Fig.3(d)
shows the situation of cross-section of the simplified horse point-set model, the
relevant curve is used to calculating the parameterization on sphere surface of
the model.
4 Conclusion
In this paper, an algorithm that can interactively generates boundary curves for
point-set models was proposed. According to a candidate boundary-point set, in
which points are selected by user gradually, calculate the distance from each new
candidate point in the set to current boundary and proximately optimizes the
length of the boundary curve. Cooperate with OpenGL and windows system, we
implement the algorithm, and apply it successfully in parameterization, feature
designate, point-set segmentation.
Acknowledgement
This work was partly supported by the Natural Science Foundation of the
Jiangsu Higher Education Institutions of China (Grant No. 07KJD460108), the
Outstanding High-end Talent Foundation of Nanjing Normal University (Grant
No. 2007013XGQ0150).
References
1. Egenhofer, M.J., Franzosa, R.D.: Point-Set Topological Spatial Relations. Interna-
tional Journal for Geographical Information Systems 5(2), 161–174 (1991)
2. Hormann, K., Labsik, U., Greiner, G.: Remeshing triangulated surfaces with optimal
parameterizations. Computer-Aided Design 33(11), 779–788 (2001)
772 P. Xiao and M.-Y. Pang
1
Dep. of Math., Xiangnan Univ., Chenzhou 423000, Hunan, China
2
Sch. of Info. Sci. & Tech., Hunan Agricultural Univ., Changsha 410128, Hunan, China
3
Sch. of Math. & Computer, Hunan Normal Univ., Changsha 410081, Hunan, China
4
Dep. of Math. & Physics, Hefei Univ., Hefei 230601, Anhui, China
dengsq66@163.com
1 Introduction
The construction method of the curve and surface and the mathematical description of
them is a key issue in computer aided geometric design. There are many ways to tackle
this problem [1-13], for example, the polynomial spline method, the Non-Uniform
Rational B-Spline(NURBS)method, and the Bézier method. These methods are effec-
tive and applied widely in shape design of industrial products, such as aircraft and
ships. Generally speaking, most of the polynomial spline methods are interpolating
methods, which means that the curves or surfaces constructed by those methods pass
through the interpolant points. To construct the polynomial spline, the derivative values
are usually needed , as well as the function values as interpolating data. Unfortunately,
in many practical problems, such as the description of the rainfall in some rainy region
and some geometric shapes, the derivative values are difficult to get. On the other hand,
one of the disadvantages of the polynomial spline method is its global property; it is not
possible for local modification under the condition that the given data are not changed.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 773–780, 2008.
© Springer-Verlag Berlin Heidelberg 2008
774 S. Deng et al.
The NURBS and Bézier methods are the so-called “no-interpolating type” methods.
This means that the constructed curve and surface do not pass through the given data,
and the given points play the role of the control points. Thus, constructing the inter-
polating function, which satisfies the following conditions, will be necessary in CAGD:
there are only function values as the interpolating data; the interpolating functions have
simple and explicit representations, so they may be convenient to use both in practical
applications and in theoretical study; the constructed curves and surfaces can be
modified under the conditions that the given data are not changed.
In recent years, univariate rational spline interpolations with parameters have been
constructed[14-15]. Motivated by the univariate rational spline interpolation, the
bivariate rational bicubic interpolation with parameters, based only on the values of the
function being interpolated, and some properties of the interpolation have been studied
in [16]. The bivariate rational bicubic interpolation based on function values and partial
derivatives with parameters have been constructed, and some properties of the inter-
polation have been studied in [17]. A bivariate rational biquartic interpolation have
been constructed [20]. This paper will deal with the new bivariate rational biquartic
interpolation based on function values of the function being interpolated.
The paper is arranged as follows. In Section 2, the new bivariate rational biquartic
spline based on function values with two parameters is constructed , and this spline is
with biquartic numerator and bilinear denominator. Section 3 deals with the smooth-
ness of the interpolating surfaces, when one of the parameters satisfies a simple con-
dition: the interpolating function is C 1 in the interpolating region. In Section 4, the
basis of the interpolation are derived. Section 5 deals with bounded property and ap-
proximation of the interpolation. For the given interpolation data, it is proved that the
values of the interpolating function in the interpolating region are bounded no matter
what the parameters might be; this is called the bounded property of the interpolation.
Also, the approximation expressions of the interpolation are derived.
2 Interpolation
Pi *,j ( x) =
pi*, j ( x) , i = 1,2,L, n − 1 , (1)
qi*, j ( x)
where
pi*, j ( x) = α i , j f i , j (1 − θ ) 4 + U i*, jθ (1 − θ ) 3 + Vi*,jθ 2 (1 − θ ) 2 + Wi *,jθ 3 (1 − θ ) + f i +1, jθ 4 ,
qi*, j ( x) = α i , j (1 − θ ) + θ ,
Rational Biquartic Interpolating Surface Based on Function Values 775
and
U i*, j = (2α i , j + 1) f i , j + α i , j f i +1, j , Vi*,j = 3 f i , j + 3α i , j f i +1, j ,
Wi *,j = (α i , j + 3) f i +1, j − hi Δ*i +1, j .
with α i , j > 0 , and Δ*i , j = ( f i +1, j − f i , j ) / hi . This interpolation is called the rational quartic
interpolation based on function values which satisfies
Pi , j ( x, y ) =
p i , j ( x, y )
, i = 1,2,L , n − 1 ; j = 1,2,L, m − 1, (2)
qi , j ( y )
where
p i , j ( x, y ) = β i , j Pi*,j ( x)(1 − η ) 4 + U i , jη (1 − η ) 3 +
Vi , jη 2 (1 − η ) 2 + Wi , jη 3 (1 − η ) + Pi*, j +1 ( x )η 4 ,
q i , j ( y ) = β i , j (1 − η ) + η ,
and
U i , j = (2β i , j + 1) Pi*,j ( x) + β i , j Pi*, j +1 ( x) ,V = 3Pi*,j ( x) + 3β i , j Pi *, j +1 ( x) ,
( x) ,
i, j
Wi , j = (β i , j + 3) Pi*, j +1 ( x) − l j Δ i , j +1
with β i , j > 0 , and Δ i , j ( x) = ( Pi*, j +1 ( x) − Pi *,j ( x)) / l j . Therefore, Pi , j ( x, y ) is called the
bivariate rational biquartic interpolating function based on function values which sat-
isfies
Pi , j ( x r , y s ) = f ( x r , y s ) , r = i, i + 1 , s = j , j + 1 .
It is easy to understand that interpolating function Pi , j ( x, y ) on [ xi , xi +1 ; y j , y j +1 ] is
existent for the given data ( xr , y s , f ( xr , y s )) , r = i, i + 1, i + 2 , s = j, j + 1, j + 2 and
parameters α i, j , β i, j .
1
3 Condition for C Interpolatory Surface
The rational interpolant function Pi*,j ( x) defined by (1) has continuous first-order de-
rivative when x ∈ [ x1 , xn ] , so ti is easy to see that the bivariable interpolant function
776 S. Deng et al.
∂P ( x, y ) and
Pi , j ( x, y ) defined by (2) has continuous first-order partial derivative i , j
∂Pi , j ( x, y ) ∂y
in the interpolating region [ x1 , x n ; y1, y m ] except for every y ∈ [ y j , y j +1 ] ,
∂x
j = 1,2,L, m − 1 at the points ( x i , y ) , i = 2,3,L , n − 1 , so it is sufficient for
Pi , j ( x, y ) ∈ C in the whole interpolating region [ x1 , x n −1 ; y1, y m −1 ] is to find the con-
1
∂Pi , j ( xi +, y) ∂Pi , j ( xi −, y )
dition that = holds. This leads to the following theorem.
∂x ∂x
Theorem 1. The sufficient condition for the interpolating function Pi , j ( x, y ) ,
i = 1,2,L , n ; j = 1,2,L , m to be C in the whole interpolating region [ x1 , x n ; y1, y m ] is
1
the parameters β i , j = cons tan t , for each j ∈ {1,2,L, m − 1} and all i = 1,2, L , n − 1 .
Proof. Without loss of generality, for any pare real numbers of (i, j ),1 ≤ i ≤ n − 1 ,
∂Pi , j ( xi +, y)
1 ≤ j ≤ m − 1 and y ∈ [ y j , y j +1 ] , it is sufficient to prove that = ∂Pi , j ( xi −, y ) .
∂x ∂x
Since
dWi , j dPi *, j +1 ( x)
+ η 3 (1 − η ) + η4],
dx dx
and
Pi *,r′ ( xi + ) = Δ*i ,r , r = j , j + 1, j + 2 ,
so
lj
Wi′, j ( xi + ) = ( β i , j + 3)Δ*i , j +1 − (Δ*i , j + 2 − Δ*i , j +1 ) .
l j +1
and thus
∂Pi , j ( x, y ) Q1 (η ) , (3)
=
∂x x = xi +
β i , j (1 − η ) + η
where
+ (( β + 3) Δ* lj
i, j i , j +1 − (Δ*i , j + 2 − Δ*i , j +1 ))η 3 (1 − η ) + Δ*i , j +1η 4 .
l j +1
Rational Biquartic Interpolating Surface Based on Function Values 777
where
Q2 (η ) = β i −1, j Δ*i , j (1 − η ) 4 + ((2 β i −1, j + 1)Δ*i , j + β i −1, j Δ*i , j +1 )η (1 − η ) 3 + (3Δ*i , j + 3β i −1, j Δ*i , j +1 )η 2 (1 − η ) 2
+ (( β
lj
i −1, j + 3)Δ*i , j +1 − (Δ*i , j + 2 − Δ*i , j +1 ))η 3 (1 − η ) + Δ*i , j +1η 4 .
l j +1
∂Pi , j ( xi +, y ) ∂Pi , j ( xi −, y )
= .
∂x ∂x
In what follows in this paper, consider the equally spaced knots case, namely,for all
i = 1,2,L, n and j = 1, 2, L, m , hi = h j ,denote it by h , and li = l j , denote it by l .
From Theorem 1, the C 1 -continuous interpolation defined by (1) must satisfy
β i , j = cons tan t , for each j ∈ {1,2,L, m − 1} and all i = 1,2, L , n − 1 . Denote this value
by β j . In the following the case that α i , j = cons tan t , for each i ∈ {1,2,L, n − 1} and all
j = 1,2, L , m − 1 is considered, and denote it by α i . Under the conditions above, Pi*,j ( x) de-
fined by (1) could be rewritten as
where
α i (1 − θ ) 3 (1 + θ ) + θ (1 − θ ) 2 (1 + 2θ ) ,
ω 0 (θ ,α i ) =
α i (1 − θ ) + θ
α iθ (1 − θ )(1 + θ (1 − θ )) + 4θ 3 − 3θ 4 , − θ 3 (1 − θ ) .
ω1 (θ , α i ) = ω 2 (θ , α i ) =
α i (1 − θ ) + θ α i (1 − θ ) + θ
and
2
∑ω
r =0
r (θ , α i ) ≡ 1 .
where
ω rs (θ , α i ;η , β j ) = ω r (θ , α i )ω s (η , β j ) (7)
and
2 2
∑∑ ω
r = 0 s =0
rs (θ , α i ;η , β j ) ≡ 1 . (8)
For the given data, the values of the bivariate interpolating function defined by (2) are
bounded in the interpolation interval as described by the following Theorem 2.
Theorem 2. Let P i , j ( x, y ) be the bivariate interpolating function defined by (2) in
[ xi , xi +1 ; y j , y j +1 ] , and denote
r = 2, s = 2
M = max f i + r , j + s .
r = 0, s = 0
Whatever the positive values of the parameters α i and β j might be, the values of
P i , j ( x, y ) in [ xi , xi +1 ; y j , y j +1 ] satisfy
1225 .
Pi , j ( x, y ) ≤ M
729
Proof. From (6) and (7),
2 2 2 2
Pi , j ( x, y ) ≤ ∑ ∑ ωrs (θ , α i ; η , β j ) fi + r , j + s ≤ M ∑ ωr (θ , α i ) ∑ ω s (η , β j ) . (9)
r =0 s = 0 r =0 s =0
show that
2
2θ 3 (1 − θ ) .
∑ω
r =0
r (θ , α i ) = 1 +
α i (1 − θ ) + θ
(10)
Similarly,
2
2η 3 (1 − η ) .
∑ω
s =0
s (η , β j ) = 1 +
β j (1 − η ) + η
(11)
Denote
2θ 3 (1 − θ ) ,
g (θ ) = 1 +
α i (1 − θ ) + θ
2 2
35 and 35
∑ω
r =0
r (θ , α i ) ≤
27
∑ω
s =0
s (η , β j ) ≤
27
(12)
so
35 2 1225
Pi , j ( x, y ) ≤ ( ) M = M .
27 729
Consider the approximation of the interpolation. Denote
∂f
= max
∂f ( x, y ) , ∂f = max
∂f ( x, y ) .
∂x x∈[ x1 , xn +1 ] ∂x ∂y y∈[ y1 , y m +1 ] ∂y
Theorem 3. Let P i , j ( x, y ) be the bivariate interpolating function defined by (2) in
[ xi , xi +1 ; y j , y j +1 ] , Whatever the positive values of the parameters α i and β j might be,
the error of the interpolation satisfies
2450 ∂f ∂f .
f ( x, y ) − Pi , j ( x, y ) ≤ (h +l )
729 ∂x ∂y
6 Conclusion
For bivariate interpolation in general, finding the bounds of the values of the interpo-
lating function expressed by the interpolation data is hard work, and deriving the error
estimate formula of the bivariate interpolation function is much more difficult. In this
paper, two of them are worked out in Theorem 2 and Theorem 3, respectively. This is
because of the convenient basis of the interpolation. The basis are very useful both in
practical design and in theoretical study.
Acknowledgements. The supports of the National Nature Science Foundation of
China, the Natural Science Foundation of Hunan Province of China and the Research
Project of Department of Education of Hunan Province of China are gratefully ac-
knowledged.
References
1. Farin, G.: Curves and surfaces for computer aided geometric design: A practical guide.
Academic Press, London (1988)
2. Chui, C.K.: Multivariate spline. SIAM, Philadelphia (1988)
3. Bézier, P.E.: The mathematical basis of the UNISURF CAD system. Butterworth, London
(1986)
4. Dierck, P., Tytgat, B.: Generating the Bézier points of β-spline curve. Computer Aided
Geometric Design 6, 279–291 (1989)
5. Piegl, L.: On NURBS: A survey. IEEE Computer Graphics and Application 11, 55–71
(1991)
6. Mielson, N.G.: CAGD’s Top Ten: What to watch. IEEE Computer Graphics and Automa-
tion, 35–37 (1993)
780 S. Deng et al.
7. Konno, K., Chiyokura, H.: An approach of designing and controlling free-form surfaces by
using NURBS boundary Gregory patches. Computer Aided Geometric Design 13, 825–849
(1996)
8. Wilcox, L.M.: First and second contributions to surface interpolation. Vision Research 39,
2335–2347 (1999)
9. Lin, R.S.: Real-time surface interpolator for 3-D parametric surface machining on 3-axis
machine tools. Machine tools and Manufacture 40, 1513–1526 (2000)
10. Jiang, D., Liu, H., Wang, W.: Test a modified surface wind interpolation scheme for com-
plex terrain in a stable atmosphere. Atmospheric Environment 35, 4877–4885 (2001)
11. Comninos, P.: An interpolating piecewise bicubic surface with shape parameters. Computer
and Graphics 25, 463–481 (2001)
12. Müller, R.: Universal parametrization and interpolation on cubic surfaces. Computer Aided
Geometric Design 19, 479–502 (2002)
13. Wang, R.H.: Multivariats Spline Functions and Their Application. Science press/Kluwer
Academic, Beijing (2001)
14. Duan, Q., Djidjeli, K., Price, W.G., et al.: A rational cubic spline based on function values.
Computer and Graphics 22, 479–486 (1998)
15. Duan, Q., Djidjeli, K., Price, W.G., et al.: The approximation properties of some rational
cubic splines. International J. of Computer Mathematics 72, 155–166 (1999)
16. Duan, Q., Wang, L., Twizell, E.H.: A new bivariate rational interpolation based on function
values. Information Sciences 166, 181–191 (2004)
17. Duan, Q., Zhang, Y., Twizell, E.H.: A bivariate rational interpolation and the properties.
Applied Mathematics and Computation 179, 190–199 (2006)
18. Duan, Q., Zhang, H., Zhang, Y., et al.: Bounded property and point control of a bivariate
rational interpolating surface. Computers and Mathematics with Applications 52, 975–984
(2006)
19. Wang, Q., Tan, J.: Rational quartic spline involving shape parameters. Journal of Informa-
tion and Computational Science 1, 127–130 (2004)
20. Wang, Q., Tan, J.: Shape preserving piecewise rational biquartic surface. Journal of Infor-
mation and Computational Science 3, 295–302 (2006)
3D Modelling for Metamorphosis for Animation
1 Introduction
3D metamorphosis is a smooth transformation from a source 3D object to a target 3D
object. Unlike 2D image morphing, 3D metamorphosis is independent of viewing or
lighting parameters. Therefore, it is a powerful technique for the entertainment industry.
The primary task in 3D metamorphosis is to automatically establish surface corre-
spondence between the source and target objects. By mapping each point on the source
object to a point on the target object, a smooth transition can be generated by interpo-
lating the source to the target object.
A common approach to establishing correspondence between two objects is to
generate a common connectivity, which is generally accomplished by decomposing
each object into several patches, embedding the patches into a 2D parametric domain,
and finally merging the corresponding embeddings to form a common mesh [4,6 8,9].
This approach has three drawbacks. First, since objects are represented as dense
polygon meshes, the large data set is difficult to manipulate. Second, the difference in
size, scale, and topology between the source and target objects pose more problems. To
handle this situation, most morphing techniques involve the use of a sparse set of user
selected feature pairs to guide decomposition and to establish an initial coarse corre-
spondence, upon which a dense correspondence can be based [7]. However, manually
marking correspondence feature points on 3D objects is a difficult and tedious task.
Third, the common connectivity generated is object-dependent. If either the source or
target object is altered, the whole process of establishing correspondences must be
repeated. Therefore, the motivation of our research is to automatically construct a
compact 3D representation using B-Splines and establish a dense correspondence
between objects automatically, without any user intervention. Although there has been
considerable work on fitting B-Spline surfaces to 3D point clouds, the object corre-
spondence is seldom addressed.
Z. Pan et al. (Eds.): Edutainment 2008, LNCS 5093, pp. 781–788, 2008.
© Springer-Verlag Berlin Heidelberg 2008
782 L. Bai, Y. Song, and Y. Wang
In the works described in [2,5], complex surfaces are often reconstructed using a
network of surface patches. Due to the uncertainty in the automatic division of surface
patches, it is difficult to establish correspondences between objects. Although single
B-Spline patch fitting [3] can avoid the problem of division uncertainty and eliminate
the nontrivial consideration on surface continuity, previous approaches are limited to
gird data representing simple topological relationships, e.g. a deformed quadrilateral
region or a deformed cylinder. Our reconstruction method is not limited to grid data,
and doe not demand reconstructed patches to be square shaped, so a complex object can
be reconstructed on a common parameter space using much fewer patches, compared to
previous approach. Detailed algorithms are presented in next section. Section 3 shows
the potential applications, and conclusion is given in Section 4.
2 B-Spline Modelling
2.1 Methodology
n
f (t ) = ∑ Bik (t )Ci (1)
i =0
⎧1 ui ≤ t ≤ ui +1
Bi0 (t ) = ⎨
⎩0 otherwise
t − ui u − t k −1 (2)
Bik (t ) = Bik −1 (t ) + i + k +1 Bi +1 (t )
ui + k − ui ui + k +1 − ui +1
Similarly, a B-Spline surface is defined over a pair of knot vector U = {u0 , u1 ,L, un }
and V = {v0 , v1 ,L vm } by:
m n
Γ ( s , t ) = ∑∑ N ik ( s ) B rj (t )ci , j (3)
i=0 j =0
Therefore, the task of B-Spline surface fitting is to find a set of control points, which
defines a surface Γ giving the best approximation to a given dataset D . In other words,
each point of the given data set D will be approximated by a point on the reconstructed
B-Spline, as follows:
Γ ( s, t ) = D ( s, t ) = ∑∑ NBc = ∑ N (∑ Bc) (4)
Thus, the reconstruction of a B-Spline surface, i.e. the computation of a set of
control points C , can actually be seen as M × N B-Spline curve-fitting processes.
3D Modelling for Metamorphosis for Animation 783
However, for non-grid datasets, each curve-fitting process will be defined on a different
knot vector. We propose a knot vector standardization algorithm.
Suppose F and L are two curves fitting the original non-grid dataset independ-
ently. F is defined on knot vector X = [ X 0 , X 1 ,L, X nx + g +1 ] by nx + 1 control points f :
[ f1 , f 2 ,L , f nx ] , L is defined on knot vector Y = [ y0 , y1 ,L, yny + g +1 ] by nY + 1 control
points l : [l1 , l2 ,Lln y ] .
nx
F ( x) = ∑ Bi , g ( x) f i (5)
i=0
x − xi xi + g +1 − x
Bi , g ( x) = Bi , g −1 ( x) + Bi +1, g −1 ( x ) (6)
xi + g − xi xi + g +1 − xi +1
ny
L( y ) = ∑ N j , g ( y )li (7)
j =0
y − yj y j + g +1 − y
N j , g ( y) = N j , g −1 ( y ) + N j +1, g −1 ( y ) (8)
h j + g − yi y j + g +1 − y j +1
The aim is to standardise X and Y to new knot vectors X ' and Y ' so that
X = Y ' = U . Instead of simply merging all knot vectors together [10], i.e.
'
X ' = Y ' = X ∪ Y = U , resulting a large knot vector, our approach is to standardise all
knot vectors to a predefined knot vector U = {u0 , u1 ,L, un + g +1} . The method works as
follows: for each element in U and X , if xi ∈ U , then do nothing; If ∃k .
(u k ∈ U) ∩ (u k ∉ X) , insert u k into X ; The control points f are re-calculated as
f ' = [ f '1 , f '2 ,L f 'n ]T and the basis function becomes:
x − uk u −x '
Bk' , g ( x) = Bk' , g −1 ( x) + k + g +1 Bk +1, g −1 ( x) (9)
uk + g − u k uk + g +1 − uk +1
The original curve is thus:
n
f ' ( x ) = ∑ Bk' , g ( x) f k' (10)
k =0
Similarly, control points l are re-calculated as l ' = [l '1 , l ' 2 ,L, l 'n ]T , and the basis func-
tion is re-defined on U :
y − uk uk + g +1 − y
N k' , g ( y ) = N k' , g −1 ( y ) + N k' +1, g −1 ( y ) (11)
uk + g − uk uk + g +1 − uk +1
n
L' ( y ) = ∑ Nk' , g ( y )lk' (12)
k =0
784 L. Bai, Y. Song, and Y. Wang
From Equation 9 and 11, it can be seen that the basis functions B ' and N ' are
identical, which can be generalised as:
s − uk u k + g +1 − s
Qk , g ( s ) = Qk , g −1 ( s ) + Qk +1, g −1 ( s ) (13)
u k + g − uk u k + g +1 − u k +1
Consequently, Equation 10 and 12 can be rewritten as:
n
F ' ( s ) = ∑ Qk , g ( s ) f ' k = A( s ) ⋅ f ' (14)
k =0
n
L' ( s ) = ∑ Qk , g ( s )l ' k = A( s ) ⋅ l ' (15)
k =0
where A( s ) = [Q0, g ( s ), Q1, g ( s ),L, Qn, g ( s )] . f ' and l ' are shape descriptors of curve F and
L , respectively. Shape descriptors have several important properties, including:
• One-to-one mapping from the parameter domain to the object space. For each
pair of parameter value ( s, t ) , there is a unique corresponding B-Spline surface
point in the object space.
• Compact representation for 3D objects. Over 90% compression rate is achieved
and similar rendering result to that using original polygon representation.
Fig. 1. Four patches B-Spline surface reconstruction example. The four patches are front and
back body, left and right arms, respectively.
Fig. 2. Textured-rendering results. Left: Polygon model (75,232 points). Right: single patch
B-Spline surface model (the surface is presented by 616 shape descriptors).
(a) (b)
Fig. 3. Difference analysis. (a) Comparison between our reconstructed single B-Spline patch and
the original data. The standard deviation is 0.002mm. (b) Comparison between the multiple
B-Spline representation (by previous approach [5]) and the original data. The standard deviation
is 0.000005mm.
Table 1. Performance comparisons between our approach and the previous B-Spline surface
modelling method
2.2 Correspondence
(s,t) ⇒ Γ k ( s, t ) (17)
Γ k ( s, t ) ⇒ Γk+1 ( s, t ) (19)
By sampling the parameter domain, e.g. uniform sampling, we obtain a set of cor-
responding B-Spline surface points on each object surface.
3 Applications
Unlike some common techniques for data compression such as principal components
analysis (PCA) describing an object as a weighted sum of principal components which
often bear little resemblance to the underlying interdependent structure of biological
forms [1], shape descriptors contain geometrical information about the objects.
Therefore, apart from recognition purpose [11] [12], the reconstructed 3D models can
be employed for freeform deformation and animation, see Fig. 4.
(a) (b)
3.2 3D Morphing
Γ 1 ( s, t ) = A( s, t ) ⋅ C 1 (20)
Γ 2 ( s, t ) = A( s, t ) ⋅ C 2 (21)
3D Modelling for Metamorphosis for Animation 787
where, the value of A( s, t ) is only related with the sampling scheme on the common
parameter domain [0, 1]×[0, 1]. Γ1 ( s, t ) and Γ 2 ( s, t ) are corresponding surface points.
Since A( s, t ) is the same for all objects when the same sampling scheme is adopted,
it can be computed once in advance.
To create a series of intermediate objects between Γ1 and Γ 2 , we simply apply
linear interpolation between corresponding points. Supposing n intermediate objects
Γ i (1 ≤ i ≤ n) are required, they can be generated by
i
Γ i ( s, t ) = Γ1 ( s, t ) + (Γ 2 ( s, t ) − Γ1 ( s, t )) (22)
n +1
Smooth morphing from one face to another is shown in Fig. 5. Four intermediate
faces are displayed between the source face (left end) and the target face (right end).
Suppose the target object is changed from Γ 2 to Γ 3 and m intermediate objects are
required while source object is still Γ1 , then the intermediate objects Γ j (1 ≤ j ≤ m) are
computed as
j
Γ j ( s, t ) = Γ1 ( s, t ) + (Γ 3 ( s, t ) − Γ1 ( s, t )) (23)
m +1
By simply changing the sampling scheme, the morphing sequence can be rendered
in different resolutions. Similarly, A( s ' , t ' ) can be computed one time for all objects.
4 Conclusions
A novel 3D modelling technique for to 3D metamorphosis is presented. In contrast to
previous works using vertices and polygons, our approach uses shape descriptors.
Despite high compression rate, rendering result using shape descriptors is still similar
to that using the original polygon representation. Moreover, one-to-one mapping from
the object space to a common parameter space can be established and therefore cor-
respondence between any pair of objects.
Though our approach provides a smooth and compact B-Spline representation for
surfaces, it may not be the best for preserving fine geometric details of objects. To
overcome this problem, a displacement map may be computed for each pixel in which
an offset from a point on a surface to a point in the original raw data is recorded.
Texture mapping is another economical solution to make up for the loss of fine surface
788 L. Bai, Y. Song, and Y. Wang
details. Since the sharp edges are rather easy to detect, they can be used to guide the
object division into several patches. The single patch B-Spline modelling technique can
be applied to each patch independently. In this case, although multiple B-Spline
patches are needed, the amount of the patches can be significantly reduced compared to
previous approaches. The nontrivial problem of enforcing G1 continuity between ad-
jacent patches is also avoided in this case. Similar solution can be applied to objects
with very complex surfaces.
References
[1] Blanz, V., Vetter, T.: A Morphable Model for the Synthesis of 3D Faces. In: Proceeding of
ACM SIGGRAPH, LA, pp. 187–194 (1999)
[2] Eck, M., Hoppe, H.: Automatic Reconstruction of B-Spine Surfaces of Arbitrary Topo-
logical Type. In: Proc. 23rd Int. Conf. on Computer Graphics and Interactive Techniques
SIGGRAPH 1996, pp. 325–334. ACM, New York (1996)
[3] Forsey, D., Bartels, R.: Surface Fitting with Hierarchical splines. ACM Transactions on
Graphics 14(2), 134–161 (1995)
[4] Hutton, T.: Dense Surface Models of the Human Face. PhD thesis (2004)
[5] Krishnamurthy, V., Levoy, M.: Fitting Smooth Surfaces to Dense Polygon Meshes.
ACM-0-89791-746-4/96/008 (1996)
[6] Kanai, T., Suzuki, H., Kimura, F.: Metamorphosis of Arbitrary Triangular Meshes. IEEE
Computer Graphics and Applications, 62–75 (2000)
[7] Lee, A., Dobkin, D., Sweldens, W., Schroder, P.: Multiresolution Mesh Morphing. In:
Proceedings of SIGGRAPH 1999, pp. 343–350 (1999)
[8] Lee, T., Huang, P.: Fast and Intuitive Metamorphosis of 3D Polyhedral Models Using
SMCC Mesh Merging Scheme. IEEE Transactions on Visualization and Computer
Graphics 9(1) (2003)
[9] Lin, C., Lee, T.: Metamorphosis of 3D Polyhedral Models Using Progressive Connectivity
Transformations. IEEE Transactions on Visualization and Computer Graphics 10(6)
(2004)
[10] Watt, A., Watt, M.: Advanced Animation and Rendering Techniques: Theory and Practise.
Addison-Wesley, Reading (1992)
[11] Bai, L., Song, Y.: Combining Graphics and Vision for 3D Face Recognition. In: Second
International Conference on Vision, Video and Graphics, Edinburgh (July 2005)
[12] Song, Y., Bai, L.: Single B-Spline Patch 3D Modelling for Facial Analysis. In: The 6th
International Conference on Recent Advances in Soft Computing, Kent, Canterbury, UK
(July 2006)
Author Index