Educational uses for Rev (was Re: Plea to sell Dan's book widely)

Michael J. Lew michaell at unimelb.edu.au
Wed Aug 11 19:02:06 EDT 2004


At 4:31 PM -0400 11/8/04, use-revolution-request at lists.runrev.com wrote:
>Perhaps another educational use of Rev-based products would be exploratory
>learning... then assessed, perhaps, by the dreaded m/c questions
>
>Judy

Exactly! While assessment can drive learning, there is more to 
teaching and learning than tests ;-)

I use simulations that I've written with Rev to first allow 
(university) students to conduct experiments and learn from the 
results, but then to design and conduct their own experiments to 
answer questions. The learning objectives of the two stages are 
different, but obviously synergistic. At the moment most of the use 
of the simulations is in supervised conditions, but I am planning a 
kit of simulations that students will use as part of self-directed 
projects. At the moment I toying with the idea that students will be 
required to make reports that can be distributed to the class as 
learning resources; having to teach something is a terrific incentive 
for learning it first!

>On Wed, 11 Aug 2004, Richard Gaskin wrote:
>
>>  Marian Petrides wrote:
>>
>>  > Not only in teaching programming but in designing custom educational
>>  > courseware. Who wants the student to have ONLY simple multiple-guess
>>  > questions to work with?
>>  >
>>  > Life doesn't come with a series of four exclusive-or questions tattooed
>>  > across it, so why give student this unrealistic view of the real world,
>>  > when a little work in Rev will permit far more challenging interactivity?
>>
>>  Agreed wholeheartedly.  Education-related work was the largest single
>>  set of tasks folks did with HyperCard, and for all the tools that have
>>  come out since there remains an unaddressed gap which may be an ideal
>>  focus for DreamCard.
>>
>>  But moving beyond simple questions models like multiple choice is
>>  difficult.  The AICC courseware interoperability standard describes
>  > almost a dozen question models, but most are variants of "choose one",
>>  "choose many", "closest match", etc., sometimes enlived by using
>>  drag-and-drop as the mechanism for applying the answer but not
>>  substantially different from what gets tested with a simple multiple
>>  choice in terms of truer assessment of what's been learned.
>>
>>  The challenge is to find more open-ended question models which can still
>>  be assessed by the computer.  For example, the most open-ended question
>>  is an essay, but I sure don't want to write the routine that scores
>>  essays. :)
>>
>>  What sorts of enhanced question models do you think would be ideal for
>>  computer-based learning?
>>
>>  --
>>    Richard Gaskin
>  >   Fourth World Media Corporation

-- 
Michael J. Lew

Senior Lecturer
Department of Pharmacology
The University of Melbourne
Parkville 3010
Victoria
Australia

Phone +613 8344 8304

**
New email address: michaell at unimelb.edu.au
**


More information about the use-livecode mailing list