Computer Chess Club Archives


Search

Terms

Messages

Subject: Re: challenge to Vincent et al

Author: Vincent Diepeveen

Date: 03:14:36 12/07/99

Go up one level in this thread


On December 06, 1999 at 21:08:15, Peter Kappler wrote:

>On December 06, 1999 at 13:23:32, Christophe Theron wrote:
>
>>On December 05, 1999 at 19:54:33, Vincent Diepeveen wrote:
>>
>>>On December 05, 1999 at 17:02:52, Christophe Theron wrote:
>>>
>>>>On December 03, 1999 at 20:48:17, Vincent Diepeveen wrote:
>>>>
>>>>>On December 03, 1999 at 16:42:31, Will Singleton wrote:
>>>>>
>>>>>>On December 03, 1999 at 08:32:39, Vincent Diepeveen wrote:
>>>>>>
>>>>>>>Current DIEP version, so no tricks, no extra extensions turned on
>>>>>>>as the ones that are currently turned on (all are out except check
>>>>>>>and a passed pawn extension), limited at
>>>>>>>6 ply against amateur at 6 ply.
>>>>>>>
>>>>>>>So limiting *all moves* it to 6 ply, including endgame.
>>>>>>>then DIEP at 10 ply against amateur at 10 ply.
>>>>>>>
>>>>>>>No pondering as we play with near to infinite time.
>>>>>>>If you want to i can also run everything on a single processor, so
>>>>>>>that transpositiontable luck because of parallellism
>>>>>>>at 10 ply is not gonna get blamed for the higher % afterwards by some
>>>>>>>scientific dudes that agree with you.
>>>>>>>
>>>>>>>Just searching n ply against n ply, without tricks as suggested some
>>>>>>>years ago like modifying eval to be an eval that did a 4 ply search.
>>>>>>>
>>>>>>>Just a 2 minute change that turns off time-check in program and forces
>>>>>>>it to
>>>>>>>use the depth=n setting.
>>>>>>>
>>>>>>>What i predict is a much closer to 50% score for the 6 against 6 ply
>>>>>>>match,
>>>>>>>and a complete annihilation of amateur when both searching at 10 ply.
>>>>>>>
>>>>>>>I assume here already of course that DIEP's eval is better than
>>>>>>>Amateurs.
>>>>>>>If i would consider the evaluation of amateur a lot better then i would
>>>>>>>obviously predict the opposite outcome (amateur annihilating diep at
>>>>>>>10 ply and nearer to 50% score at 6 ply).
>>>>>>>
>>>>>>>Of course things like learning and such must be turned off.
>>>>>>>Preferably both versions playing the same openings. However by just
>>>>>>>playing a large quantity of games we can measure statistically accurate
>>>>>>>what has happened.
>>>>>>>
>>>>>>>To make the experiment even better we should log the evaluations too
>>>>>>>and see where things went wrong and calculate whether that had
>>>>>>>statistical influence on the outcome.
>>>>>>>
>>>>>>>Also we could start with the same book if you want to.
>>>>>>>
>>>>>>
>>>>>>Vincent,
>>>>>>
>>>>>>An interesting offer, sounds like fun.  However, it would be worthwhile only if
>>>>>>
>>>>>>1) Our programs were bugfree.  Mine is not.
>>>>>>2) Extensions are identical.
>>>>>>3) Search and Qsearch are identical.
>>>>>>4) HT and null-move are identical.
>>>>>>5) Books identical.
>>>>>>6) Independent verification of all this.
>>>>>
>>>>>Except for the books this is not needed.
>>>>>
>>>>>>What you are trying to measure is the proposition that, given unequal evals,
>>>>>>games between deep searchers will enable the better eval to win more often than
>>>>>>games between shallow searchers, due to the vagaries of tactics (which you
>>>>>>contend are pretty random at 6 ply, but not at 10 ply).  I contend that tactics
>>>>>>would continue to be significant at 10 ply, though *perhaps* not to as great a
>>>>>>degree as 6 ply.
>>>>>>
>>>>>>So I think it would be a worthwhile experiment, but not using unequal programs.
>>>>>>Use unequal evals, with the same program.
>>>>>>
>>>>>>I will make my eval suck even more than it does now, and play a 100 game match
>>>>>>at 6-ply and then at 9-ply levels (mine won't reach 10 too often, besides
>>>>>>there's some bug that causes an overflow in something when I get too deep).
>>>>>>
>>>>>>If you or others wish to do the same experiment, then we might be able to get
>>>>>>some usefuls results.
>>>>>
>>>>>This experiment is not good, as we don't talk about an eval that's
>>>>>tuned to play well with little info, but just a sucking eval.
>>>>>
>>>>>You need to tune your sucking eval first like nimzo and junior
>>>>>and chesstiger are tuned before starting the experiment.
>>>>
>>>>
>>>>Or degrade your sucking eval to the level of Diep.
>>>>
>>>>If my evaluation sucks, Vincent, what should I say about yours?
>>>
>>>Well obviously yours sucks. it's blowing position after position
>>>after being won out of book.
>>
>>
>>So you are implying my program wins only because of his book?
>>
>>Or is it the opposite?
>>
>>Both assertions are ridiculous anyway.
>>
>>If you want a match between Diep and Chess Tiger, just ask. We'll find an
>>independent tester, send him our programs, and let him tell us the result.
>>
>>We'll see which one sucks.
>>
>>    Christophe
>
>
>Vincent has been claiming for years that Diep's eval is second to none.
>
>I don't doubt that he has written a good eval, as he is a decent chessplayer,
>but in doing so it seems that he has made major concessions in search speed.
>
>I have little doubt that Tiger would win a match of reasonable length, but maybe
>Vincent will try to prove me wrong.
>
>As for his comment that Tiger "blows position after position" out of book, what
>he probably means is that he saw one or two such games recently on ICC.  He
>tends to overstate these things, IMHO...

You seem to be forgetting 2 things
  a) tiger has played a lot of games recently and seemingly no
     one has analyzed them yet except me, Jeroen Noomen, Alexander Kure.

  b) you can play diep any time at the icc server.

>--Peter



This page took 0 seconds to execute

Last modified: Thu, 15 Apr 21 08:11:13 -0700

Current Computer Chess Club Forums at Talkchess. This site by Sean Mintz.