In article <[EMAIL PROTECTED]>,
  [EMAIL PROTECTED] (Donald Burrill) wrote:
> On Sat, 28 Oct 2000, Martin Boulger (impersonating haytham siala)
wrote:
>
> > I have to mark scripts based on a marking scheme thus:
>
> How much of the ensuing paragraph comprises requirements externally
> imposed upon you and not subject to your control, what conditions are
> debatable, and which ones are your own choice?
>
> > 10 questions of equal weighting, grade each answer from A to F.
>
> Is this data as presented, or is part of your task the grading of
each
> question?  If the latter, do you get to mark the questions _de_novo_,
or
> do you receive a collection of numerical values to be classified into
> the categories A to F?  Do the categories A to F comprise 6
categories
> (as the sequel seems to imply), or are they modifiable by (e.g.) +
and -
> (and ++ and -- ?) suffixes?
>
> Frankly, I always prefer to work with the original raw scores (from
which
> I presume the initial letter-grades for the ten questions were
derived,
> and in the process coarsened).  But my responses below are based on
an
> assumption that the only information available to you are the letter
> grades for each question.
>
> > So each question, obviously, counts for 10%.
>
> In one sense, yes:  "equal weighting", in the usual sense, would
imply
> this if you understand the process to entail a weighted sum of the 10
> question-grades.  It is, however, possible (if perhaps not usual) to
> devise grading schemata that give equal importance to each question
but
> do not entail _adding_ the separate grades.
>
> > Given a random set of grades how do I work out a final grade?
>
> Interesting question!  Why, I wonder, would anyone be given a
_random_
> set of grades?  Usually one would have all the papers for a
particular
> course (or section of a course), or all the marks previously assigned
to
> those enrolled, or some systematically chosen subset thereof, I
should
> think.
>
> > I considered allocating for each question 10% for an "A", 8% for
an"B"
> > through to 0% for an F.  I can add up the grades for each question
to
> > get a grade out of 100%.
>
> Yes, you can do this.  I would not use the symbol "%", though.  You
can
> by this means produce a total score whose maximum is 100 points, but
it
> is not all clear that this score would properly represent a per
centum of
> anything at all, let alone anything interesting.
>
> > However, only 70% or above = an "A", 60 - 69 a "B", 50 - 59 a "C",
> > 40 - 49 a "D" and 30 - 39 = "E" below this is an F.
>
> Why?  Is this scheme imposed on you, or is this just your arbitrary
> choice of cutting points for recoding the total score?
>
> > So my idea doesn't work (and wouldn't be very accurate anyway).
>
> In what sense(s) does it not work?  (Apart from some obvious
> dysfunctions, like 10 B's on the original questions magically turning
> into an A for the overall grade, which some folks might not consider
> reasonable.)  As for accuracy, what kind of accuracy do you want to
> attain, and how do you wish to define "accuracy"?
>
> > Any clues to a statistical technique appropriate for this problem?
>
> Not without summat more in the way of detail, including what I
suspect to
> be a whole bunch of hidden assumptions.  (What, for instance, do the
> original 10 questions represent?  They might be different domains on
a
> single final examination;  or a chronological series of midterm tests
in
> some course of study;  or some peculiar combination of these;  or
> something entirely else.  Depending, I'd have different proposals for
> how to handle the aggregation process.)
>
> You might begin by considering how you want the results to turn out.
> One way of operationalizing "equal weight" is to report each case's
> marks in lexicographical order, e.g.   AAAAB BBBCD  for any case that
> has four As, four Bs, one C and one D, ignoring which particular
> question earned each of these marks.  Now consider the table of
possible
> marks (or, if you prefer, restrict your attention to the actual
> distribution of marks you have to deal with;  but better, I think, to
> consider the population first, then the sample).  I report some of
them, in
> groups of five for readability:
>
> AAAAA AAAAA
> AAAAA AAAAB
> AAAAA AAAAC
> AAAAA AAAAD
> AAAAA AAAAE
> AAAAA AAAAF
> AAAAA AAABB
>  . . .
>
> EFFFF FFFFF
> FFFFF FFFFF
>
> For which of these patterns ought the output to be "A"?  ... "B"?
Etc.?
>
> This is perhaps easier to contemplate in the abstract if we consider
the
> task for four marks to be aggregated (instead of ten).  The 126
possible
> patterns are these:
>
> AAAA  AACF    ABCE    ACDF    BBBB    BBFF    BDEE    CCEE    DDDE
> AAAB  AADD    ABCF    ACEE    BBBC    BCCC    BDEF    CCEF    DDDF
> AAAC  AADE    ABDD    ACEF    BBBD    BCCD    BDFF    CCFF    DDEE
> AAAD  AADF    ABDE    ACFF    BBBE    BCCE    BEEE    CDDD    DDEF
> AAAE  AAEE    ABDF    ADDD    BBBF    BCCF    BEEF    CDDE    DDFF
> AAAF  AAEF    ABEE    ADDE    BBCC    BCDD    BEFF    CDDF    DEEE
> AABB  AAFF    ABEF    ADDF    BBCD    BCDE    BFFF    CDEE    DEEF
> AABC  ABBB    ABFF    ADEE    BBCE    BCDF    CCCC    CDEF    DEFF
> AABD  ABBC    ACCC    ADEF    BBCF    BCEE    CCCD    CDFF    DFFF
> AABE  ABBD    ACCD    ADFF    BBDD    BCEF    CCCE    CEEE    EEEE
> AABF  ABBE    ACCE    AEEE    BBDE    BCFF    CCCF    CEEF    EEEF
> AACC  ABBF    ACCF    AEEF    BBDF    BDDD    CCDD    CEFF    EEFF
> AACD  ABCC    ACDD    AEFF    BBEE    BDDE    CCDE    CFFF    EFFF
> AACE  ABCD    ACDE    AFFF    BBEF    BDDF    CCDF    DDDD    FFFF
>
> Of these, there are some that would generally be agreed to
represent "A"
> work overall (AAAA and AAAB surely, probably AAAC and AABB, maybe
AAAD?)
> (and similarly for "B", "C", etc.);  and some that are less easy to
> categorize into a set of six ranks (e.g., AAAF in the above list).
>  Once one has wrapped one's mind around the underlying task, it may
be
> more straightforward (if no less tedious!) to devise an algorithm
that
> approximates that categorization acceptably.  Simple arithmetic is
> unlikely to be subtle enough, though.
>       (And of course, even for simple arithmetic it is unclear
whether
> the grading scheme should be treated as an equal-interval scale.  I
would
> argue that it shouldn't.)
>
> I don't suppose this has made what you thought you were about any
> easier;  sorry about that!  On the other hand, I didn't read you as
> asking (at least, not explicitly) for "easier".
>                                               -- DFB.
>  ---------------------------------------------------------------------
-
>  Donald F. Burrill
[EMAIL PROTECTED]
>  348 Hyde Hall, Plymouth State College,
[EMAIL PROTECTED]
>  MSC #29, Plymouth, NH 03264                             (603) 535-
2597
>  Department of Mathematics, Boston University
[EMAIL PROTECTED]
>  111 Cummington Street, room 261, Boston, MA 02215       (617) 353-
5288
>  184 Nashua Road, Bedford, NH 03110                      (603) 471-
7128
>
> =================================================================
> Instructions for joining and leaving this list and remarks about
> the problem of INAPPROPRIATE MESSAGES are available at
>                   http://jse.stat.ncsu.edu/
> =================================================================
>


Sent via Deja.com http://www.deja.com/
Before you buy.


=================================================================
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
                  http://jse.stat.ncsu.edu/
=================================================================

Reply via email to