--- test/html-webhacc/cc.cgi 2007/06/27 12:35:24 1.2 +++ test/html-webhacc/cc.cgi 2007/09/02 07:59:01 1.17 @@ -2,12 +2,10 @@ use strict; use lib qw[/home/httpd/html/www/markup/html/whatpm - /home/wakaba/work/manakai/lib - /home/wakaba/public_html/-temp/wiki/lib]; + /home/wakaba/work/manakai2/lib]; use CGI::Carp qw[fatalsToBrowser]; use Scalar::Util qw[refaddr]; - -use SuikaWiki::Input::HTTP; ## TODO: Use some better CGI module +use Time::HiRes qw/time/; sub htescape ($) { my $s = $_[0]; @@ -15,147 +13,179 @@ $s =~ s//>/g; $s =~ s/"/"/g; - $s =~ s!([\x00-\x09\x0B-\x1F\x7F-\x80])!sprintf 'U+%04X', ord $1!ge; + $s =~ s{([\x00-\x09\x0B-\x1F\x7F-\xA0\x{FEFF}\x{FFFC}-\x{FFFF}])}{ + sprintf 'U+%04X', ord $1; + }ge; return $s; } # htescape -my $http = SuikaWiki::Input::HTTP->new; - -## TODO: _charset_ - - my $input_format = $http->parameter ('i') || 'text/html'; - my $inner_html_element = $http->parameter ('e'); - my $input_uri = 'thismessage:/'; + use Message::CGI::HTTP; + my $http = Message::CGI::HTTP->new; - my $s = $http->parameter ('s'); - if (length $s > 1000_000) { - print STDOUT "Status: 400 Document Too Long\nContent-Type: text/plain; charset=us-ascii\n\nToo long"; + if ($http->get_meta_variable ('PATH_INFO') ne '/') { + print STDOUT "Status: 404 Not Found\nContent-Type: text/plain; charset=us-ascii\n\n400"; exit; } + binmode STDOUT, ':utf8'; + $| = 1; + + require Message::DOM::DOMImplementation; + my $dom = Message::DOM::DOMImplementation->new; + + load_text_catalog ('en'); ## TODO: conneg + + my @nav; print STDOUT qq[Content-Type: text/html; charset=utf-8 Web Document Conformance Checker (BETA) - - + -

Web Document Conformance Checker (beta)

+

Web Document Conformance Checker +(beta)

+]; + $| = 0; + my $input = get_input_document ($http, $dom); + my $inner_html_element = $http->get_parameter ('e'); + my $char_length = 0; + my %time; + my $time1; + my $time2; + + print qq[ +
+
Request URI
+
<@{[htescape $input->{request_uri}]}>
Document URI
-
<@{[htescape $input_uri]}>
-
Internet Media Type
-
@{[htescape $input_format]}
+
<@{[htescape $input->{uri}]}>
]; # no
yet + push @nav, ['#document-info' => 'Information']; + +if (defined $input->{s}) { + $char_length = length $input->{s}; + + print STDOUT qq[ +
Base URI
+
<@{[htescape $input->{base_uri}]}>
+
Internet Media Type
+
@{[htescape $input->{media_type}]} + @{[$input->{media_type_overridden} ? '(overridden)' : '']}
+
Character Encoding
+
@{[defined $input->{charset} ? ''.htescape ($input->{charset}).'' : '(none)']} + @{[$input->{charset_overridden} ? '(overridden)' : '']}
+
Length
+
$char_length byte@{[$char_length == 1 ? '' : 's']}
+ +
+]; + + print_http_header_section ($input); - require Message::DOM::DOMImplementation; - my $dom = Message::DOM::DOMImplementation->____new; my $doc; my $el; - if ($input_format eq 'text/html') { + if ($input->{media_type} eq 'text/html') { require Encode; require Whatpm::HTML; - - $s = Encode::decode ('utf-8', $s); - print STDOUT qq[ -
Character Encoding
-
(none)
- + $input->{charset} ||= 'ISO-8859-1'; ## TODO: for now. -
-]; - print_source_string (\$s); - print STDOUT qq[ -
+ $time1 = time; + my $t = Encode::decode ($input->{charset}, $input->{s}); + $time2 = time; + $time{decode} = $time2 - $time1; + print STDOUT qq[

Parse Errors

- + print STDOUT qq[
]; - } elsif ($input_format eq 'application/xhtml+xml') { - require Message::DOM::XMLParserTemp; - require Encode; - - my $t = Encode::decode ('utf-8', $s); - print STDOUT qq[ -
Character Encoding
-
(none)
- + print_source_string_section (\($input->{s}), $input->{charset}); + } elsif ({ + 'text/xml' => 1, + 'application/atom+xml' => 1, + 'application/rss+xml' => 1, + 'application/svg+xml' => 1, + 'application/xhtml+xml' => 1, + 'application/xml' => 1, + }->{$input->{media_type}}) { + require Message::DOM::XMLParserTemp; -
-]; - print_source_string (\$t); print STDOUT qq[ -
-

Parse Errors

- + print STDOUT qq[
+ ]; + print_source_string_section (\($input->{s}), $doc->input_encoding); } else { + ## TODO: Change HTTP status code?? print STDOUT qq[ - - -

Media type @{[htescape $input_format]} is not supported!

+
+

Media type @{[htescape $input->{media_type}]} is not supported!

+
]; + push @nav, ['#result-summary' => 'Result']; } @@ -164,6 +194,7 @@

Document Tree

]; + push @nav, ['#document-tree' => 'Tree']; print_document_tree ($el || $doc); @@ -173,50 +204,235 @@

Document Errors

- + print STDOUT qq[
]; + + if (@{$elements->{table}}) { + require JSON; + + push @nav, ['#tables' => 'Tables']; + print STDOUT qq[ +
+

Tables

+ + + + +]; + + my $i = 0; + for my $table_el (@{$elements->{table}}) { + $i++; + print STDOUT qq[

] . + get_node_link ($table_el) . q[

]; + + ## TODO: Make |ContentChecker| return |form_table| result + ## so that this script don't have to run the algorithm twice. + my $table = Whatpm::HTMLTable->form_table ($table_el); + + for (@{$table->{column_group}}, @{$table->{column}}, $table->{caption}) { + next unless $_; + delete $_->{element}; + } + + for (@{$table->{row_group}}) { + next unless $_; + next unless $_->{element}; + $_->{type} = $_->{element}->manakai_local_name; + delete $_->{element}; + } + + for (@{$table->{cell}}) { + next unless $_; + for (@{$_}) { + next unless $_; + for (@$_) { + $_->{id} = refaddr $_->{element} if defined $_->{element}; + delete $_->{element}; + $_->{is_header} = $_->{is_header} ? 1 : 0; + } + } + } + + print STDOUT '
]; + } + + print STDOUT qq[
]; + } + + if (keys %{$elements->{id}}) { + push @nav, ['#identifiers' => 'IDs']; + print STDOUT qq[ +
+

Identifiers

+ +
+]; + for my $id (sort {$a cmp $b} keys %{$elements->{id}}) { + print STDOUT qq[
@{[htescape $id]}
]; + for (@{$elements->{id}->{$id}}) { + print STDOUT qq[
].get_node_link ($_).qq[
]; + } + } + print STDOUT qq[
]; + } + + if (keys %{$elements->{term}}) { + push @nav, ['#terms' => 'Terms']; + print STDOUT qq[ +
+

Terms

+ +
+]; + for my $term (sort {$a cmp $b} keys %{$elements->{term}}) { + print STDOUT qq[
@{[htescape $term]}
]; + for (@{$elements->{term}->{$term}}) { + print STDOUT qq[
].get_node_link ($_).qq[
]; + } + } + print STDOUT qq[
]; + } + + if (keys %{$elements->{class}}) { + push @nav, ['#classes' => 'Classes']; + print STDOUT qq[ +
+

Classes

+ +
+]; + for my $class (sort {$a cmp $b} keys %{$elements->{class}}) { + print STDOUT qq[
@{[htescape $class]}
]; + for (@{$elements->{class}->{$class}}) { + print STDOUT qq[
].get_node_link ($_).qq[
]; + } + } + print STDOUT qq[
]; + } } ## TODO: Show result +} else { + print STDOUT qq[ + +
+ +
+

Input Error: @{[htescape ($input->{error_status_text})]}

+
+]; + push @nav, ['#result-summary' => 'Result']; + +} + + print STDOUT qq[ + ]; + for (qw/decode parse parse_xml check/) { + next unless defined $time{$_}; + open my $file, '>>', ".cc-$_.txt" or die ".cc-$_.txt: $!"; + print $file $char_length, "\t", $time{$_}, "\n"; + } + exit; -sub print_source_string ($) { - my $s = $_[0]; - my $i = 1; - print STDOUT qq[
    \n]; - while ($$s =~ /\G([^\x0A]*?)\x0D?\x0A/gc) { - print STDOUT qq[
  1. ], htescape $1, "
  2. \n"; - $i++; +sub print_http_header_section ($) { + my $input = shift; + return unless defined $input->{header_status_code} or + defined $input->{header_status_text} or + @{$input->{header_field}}; + + push @nav, ['#source-header' => 'HTTP Header']; + print STDOUT qq[
    +

    HTTP Header

    + +

    Note: Due to the limitation of the +network library in use, the content of this section might +not be the real header.

    + + +]; + + if (defined $input->{header_status_code}) { + print STDOUT qq[]; + print STDOUT qq[]; + } + if (defined $input->{header_status_text}) { + print STDOUT qq[]; + print STDOUT qq[]; } - if ($$s =~ /\G([^\x0A]+)/gc) { - print STDOUT qq[
  3. ], htescape $1, "
  4. \n"; + + for (@{$input->{header_field}}) { + print STDOUT qq[]; + print STDOUT qq[]; + } + + print STDOUT qq[
    Status code@{[htescape ($input->{header_status_code})]}
    Status text@{[htescape ($input->{header_status_text})]}
    @{[htescape ($_->[0])]}@{[htescape ($_->[1])]}
    ]; +} # print_http_header_section + +sub print_source_string_section ($$) { + require Encode; + my $enc = Encode::find_encoding ($_[1]); ## TODO: charset name -> Perl name + return unless $enc; + + my $s = \($enc->decode (${$_[0]})); + my $i = 1; + push @nav, ['#source-string' => 'Source']; + print STDOUT qq[
    +

    Document Source

    +
      \n]; + if (length $$s) { + while ($$s =~ /\G([^\x0A]*?)\x0D?\x0A/gc) { + print STDOUT qq[
    1. ], htescape $1, "
    2. \n"; + $i++; + } + if ($$s =~ /\G([^\x0A]+)/gc) { + print STDOUT qq[
    3. ], htescape $1, "
    4. \n"; + } + } else { + print STDOUT q[
    5. ]; } - print STDOUT "
    "; -} # print_input_string + print STDOUT "
"; +} # print_input_string_section sub print_document_tree ($) { my $node = shift; @@ -233,46 +449,61 @@ my $node_id = 'node-'.refaddr $child; my $nt = $child->node_type; if ($nt == $child->ELEMENT_NODE) { - $r .= qq'
  • ' . htescape ($child->tag_name) . + my $child_nsuri = $child->namespace_uri; + $r .= qq[
  • ] . htescape ($child->tag_name) . ''; ## ISSUE: case if ($child->has_attributes) { $r .= ''; } - if ($node->has_child_nodes) { + if ($child->has_child_nodes) { $r .= '
      '; - unshift @node, @{$child->child_nodes}, '
    '; + unshift @node, @{$child->child_nodes}, '
  • '; + } else { + $r .= ''; } } elsif ($nt == $child->TEXT_NODE) { - $r .= qq'
  • ' . htescape ($child->data) . '
  • '; + $r .= qq'
  • ' . htescape ($child->data) . '
  • '; } elsif ($nt == $child->CDATA_SECTION_NODE) { - $r .= qq'
  • <[CDATA[' . htescape ($child->data) . ']]>
  • '; + $r .= qq'
  • <[CDATA[' . htescape ($child->data) . ']]>
  • '; } elsif ($nt == $child->COMMENT_NODE) { - $r .= qq'
  • <!--' . htescape ($child->data) . '-->
  • '; + $r .= qq'
  • <!--' . htescape ($child->data) . '-->
  • '; } elsif ($nt == $child->DOCUMENT_NODE) { - $r .= qq'
  • Document
  • '; + $r .= qq'
  • Document'; + $r .= qq[]; if ($child->has_child_nodes) { - $r .= '
      '; - unshift @node, @{$child->child_nodes}, '
    '; + $r .= '
      '; + unshift @node, @{$child->child_nodes}, '
  • '; } } elsif ($nt == $child->DOCUMENT_TYPE_NODE) { - $r .= qq'
  • <!DOCTYPE>