[links-list] bug? (scanning NUL characters is VERY slow)

José Luis González González jlg80 at mi.madritel.es
Fri Nov 29 09:18:22 PST 2002


I noticed that both Links and ELinks take lots of time parsing files
with NUL characters.  This time seems to be proportional to the number
of NUL characters, to say the least.

It's easy to reproduce:

$ cat testfile.html
<p>This file includes NUL characters</p>
$ dd if=/dev/zero bs=1k count=70 >>testfile.html
$ echo '</body></html>' >>testfile.html
$ time links -dump testfile.html >/dev/null # This will be very slow

Since some of you may think a HTML document should never contain them,
take a look at http://www.joelonsoftware.com/navLinks/fog0000000247.html

NUL characters should be ignored when scanning, so where does the
overhead come from?  Are they actually not ignored?
Unsubscribe: send email to links-list-request at linuxfromscratch.org
and put unsubscribe in the subject header of the message

More information about the links-list mailing list