POV-Ray : Newsgroups : povray.beta-test : Experimental v3.8.0-x.tokenizer.9960461 Server Time: 17 Jan 2019 02:45:31 GMT
  Experimental v3.8.0-x.tokenizer.9960461 (Message 1 to 6 of 6)  
From: clipka
Subject: Experimental v3.8.0-x.tokenizer.9960461
Date: 9 Dec 2018 02:26:32
Message: <5c0c7d58$1@news.povray.org>
Hi folks,

another iteration of the new parser needs thorough testing:

https://github.com/POV-Ray/povray/releases/tag/v3.8.0-x.tokenizer.9960461

Please let me know of any SDL constructs that work in the latest alpha 
but cause this version to throw a tantrum.


Post a reply to this message

From: jr
Subject: Re: Experimental v3.8.0-x.tokenizer.9960461
Date: 9 Dec 2018 15:10:00
Message: <web.5c0d2f9298d8491ae43b71790@news.povray.org>
hi,

clipka <ano### [at] anonymousorg> wrote:
> Hi folks,
> another iteration of the new parser needs thorough testing:
> https://github.com/POV-Ray/povray/releases/tag/v3.8.0-x.tokenizer.9960461
> Please let me know of any SDL constructs that work in the latest alpha
> but cause this version to throw a tantrum.

it's quite a bit slower than the previous x.tokenizer version (from 5576 to 6095
seconds for the first test), also shows a slight increase in "K tokens"
processed for same scene.


regards, jr.


Post a reply to this message

From: clipka
Subject: Re: Experimental v3.8.0-x.tokenizer.9960461
Date: 10 Dec 2018 01:54:51
Message: <5c0dc76b$1@news.povray.org>
Am 09.12.2018 um 16:09 schrieb jr:
> hi,
> 
> clipka <ano### [at] anonymousorg> wrote:
>> Hi folks,
>> another iteration of the new parser needs thorough testing:
>> https://github.com/POV-Ray/povray/releases/tag/v3.8.0-x.tokenizer.9960461
>> Please let me know of any SDL constructs that work in the latest alpha
>> but cause this version to throw a tantrum.
> 
> it's quite a bit slower than the previous x.tokenizer version (from 5576 to 6095
> seconds for the first test), also shows a slight increase in "K tokens"
> processed for same scene.

Performance is the least of my "tactical" priorities (despite being a 
strategical goal of the whole thing).

The token count thing worries me a bit. I couldn't care less if I 
inadvertently introduced a slight change to the rules for how the number 
of tokens is counted as a side effect of some other sensible change, but 
I can't think of any recent change that might have such an effect; are 
you sure it's not simply an artifact of the way you observe this number? 
And are you sure we are talking about the same reference version 
(x-tokenizer.9945666)? If so, can you narrow down the scene language 
construct for which the values differ?

That said, the main focus of this version is on furthering my 
understanding of what's left of the legacy parser code. To that end, 
I've peppered the code with checks to verify some assumptions that may 
or may not hold true, and what I'm really interested in right now is 
reports of cases where they don't. Those should manifest as parse errors.


Post a reply to this message

From: jr
Subject: Re: Experimental v3.8.0-x.tokenizer.9960461
Date: 10 Dec 2018 13:15:00
Message: <web.5c0e65cc98d8491ae43b71790@news.povray.org>
hi,

clipka <ano### [at] anonymousorg> wrote:
> Am 09.12.2018 um 16:09 schrieb jr:
> > it's quite a bit slower than the previous x.tokenizer version (from 5576 to 6095
> > seconds for the first test), also shows a slight increase in "K tokens"
> > processed for same scene.
> The token count thing worries me a bit. I couldn't care less if I
> inadvertently introduced a slight change to the rules for how the number
> of tokens is counted as a side effect of some other sensible change, but
> I can't think of any recent change that might have such an effect; are
> you sure it's not simply an artifact of the way you observe this number?

there's more than one way to read it? :-)

> And are you sure we are talking about the same reference version
> (x-tokenizer.9945666)? If so, can you narrow down the scene language
> construct for which the values differ?

yes, and not really.  I used 9945627 as the "baseline", then ran the same scene
with 9945666 and 9960461.  all three were run "remote" on an otherwise idling
machine.  as for the language construct, I can not even guess, I have never
really looked at POV-Ray sources.  essentially, nested loops calling a macro
that uses inside() to trace, ie a scan of a volume.

curiously, the alpha reports the "no objects in scene" warning one line earlier
than the x.tokenizer versions.

> That said, the main focus of this version is on furthering my
> understanding of what's left of the legacy parser code. To that end,
> I've peppered the code with checks to verify some assumptions that may
> or may not hold true, and what I'm really interested in right now is
> reports of cases where they don't. Those should manifest as parse errors.

ah, I might not be much help here, I have no code requiring earlier than 3.6,
still, I'll keep using the latest x.tokenizer as the day-to-day version.


regards ,jr.


Post a reply to this message

From: clipka
Subject: Re: Experimental v3.8.0-x.tokenizer.9960461
Date: 10 Dec 2018 15:39:21
Message: <5c0e88a9$1@news.povray.org>
Am 10.12.2018 um 14:10 schrieb jr:
> hi,
> 
> clipka <ano### [at] anonymousorg> wrote:
>> Am 09.12.2018 um 16:09 schrieb jr:
>>> it's quite a bit slower than the previous x.tokenizer version (from 5576 to 6095
>>> seconds for the first test), also shows a slight increase in "K tokens"
>>> processed for same scene.
>> The token count thing worries me a bit. I couldn't care less if I
>> inadvertently introduced a slight change to the rules for how the number
>> of tokens is counted as a side effect of some other sensible change, but
>> I can't think of any recent change that might have such an effect; are
>> you sure it's not simply an artifact of the way you observe this number?
> 
> there's more than one way to read it? :-)

As far as I'm aware, POV-Ray for Windows only reports the token count in 
the status line, so the count disappears as soon as the render proper 
starts.

I take it you're not running Windows then. (Sorry, I tend to not keep 
track of who runs which OS.)

> curiously, the alpha reports the "no objects in scene" warning one line earlier
> than the x.tokenizer versions.

Not much of a surprise there. Line number counting has been rewritten 
from scratch, and may differ in how it behaves at the end of the file.

>> That said, the main focus of this version is on furthering my
>> understanding of what's left of the legacy parser code. To that end,
>> I've peppered the code with checks to verify some assumptions that may
>> or may not hold true, and what I'm really interested in right now is
>> reports of cases where they don't. Those should manifest as parse errors.
> 
> ah, I might not be much help here, I have no code requiring earlier than 3.6,
> still, I'll keep using the latest x.tokenizer as the day-to-day version.

Sorry, misunderstanding I guess. The term "legacy" above wasn't 
referring to language constructs (as in e.g. v3.7 vs v3.0 scene 
language), but rather the parser itself (as in alpha vs. x.tokenizer 
parser code).


Post a reply to this message

From: jr
Subject: Re: Experimental v3.8.0-x.tokenizer.9960461
Date: 10 Dec 2018 16:50:06
Message: <web.5c0e984398d8491ae43b71790@news.povray.org>
hi,

clipka <ano### [at] anonymousorg> wrote:
> >> ... sure it's not simply an artifact of the way you observe this number?
> >
> > there's more than one way to read it? :-)
>
> As far as I'm aware, POV-Ray for Windows only reports the token count in
> the status line, so the count disappears as soon as the render proper
> starts.

same for the *nix version, afaict.

when needed, I capture terminal io with:

  $ script mylog.txt
  $ povray ...
  $ exit

and have complete "transcript" for later.

> I take it you're not running Windows then. (Sorry, I tend to not keep
> track of who runs which OS.)

not anymore.  my household is 100% Microsoft-free.  (well, I kept the fonts :-))

> > curiously, the alpha reports the "no objects in scene" warning one line earlier
> > than the x.tokenizer versions.
>
> Not much of a surprise there. Line number counting has been rewritten
> from scratch, and may differ in how it behaves at the end of the file.

ok.

> >> That said, the main focus of this version is on furthering my
> >> understanding of what's left of the legacy parser code. ...
> > ah, I might not be much help here, I have no code requiring earlier than 3.6,
> > still, I'll keep using the latest x.tokenizer as the day-to-day version.
>
> Sorry, misunderstanding I guess. The term "legacy" above wasn't
> referring to language constructs (as in e.g. v3.7 vs v3.0 scene
> language), but rather the parser itself (as in alpha vs. x.tokenizer
> parser code).

ah yes, I (now) see you wrote legacy parser.  so, so far then, six scenes
tested, all run slower in the alpha (all versions self-compiled etc).


regards, jr.


Post a reply to this message

Copyright 2003-2008 Persistence of Vision Raytracer Pty. Ltd.