mirror of
https://github.com/lexborisov/Modest
synced 2024-11-22 13:51:39 +03:00
744 lines
33 KiB
C
744 lines
33 KiB
C
/*
|
|
Copyright (C) 2015-2017 Alexander Borisov
|
|
|
|
This library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
This library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with this library; if not, write to the Free Software
|
|
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
Author: lex.borisov@gmail.com (Alexander Borisov)
|
|
*/
|
|
|
|
#include "myhtml/tokenizer_end.h"
|
|
|
|
|
|
size_t myhtml_tokenizer_end_state_data(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_size + tree->global_offset), MyHTML_TOKEN_TYPE_DATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(token_node->raw_begin < (html_size + tree->global_offset)) {
|
|
if(token_node->raw_begin) {
|
|
token_node->raw_length = (html_offset + tree->global_offset) - token_node->raw_begin;
|
|
myhtml_check_tag_parser(tree, token_node, html, html_offset);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
else {
|
|
token_node->type ^= (token_node->type & MyHTML_TOKEN_TYPE_WHITESPACE);
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_size + tree->global_offset), MyHTML_TOKEN_TYPE_DATA);
|
|
}
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_tag_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_size + tree->global_offset), MyHTML_TOKEN_TYPE_DATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_end_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(token_node->raw_begin < (html_size + tree->global_offset))
|
|
{
|
|
token_node->raw_length = (html_offset + tree->global_offset) - token_node->raw_begin;
|
|
token_node->type ^= (token_node->type & MyHTML_TOKEN_TYPE_WHITESPACE);
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_size + tree->global_offset), MyHTML_TOKEN_TYPE_DATA);
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_self_closing_start_tag(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_size + tree->global_offset), MyHTML_TOKEN_TYPE_DATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_markup_declaration_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(token_node->raw_begin > 1) {
|
|
tree->incoming_buf->length = myhtml_tokenizer_state_bogus_comment(tree, token_node, html, token_node->raw_begin, html_size);
|
|
|
|
if(token_node != tree->current_token_node)
|
|
{
|
|
token_node = tree->current_token_node;
|
|
token_node->raw_length = (html_size + tree->global_offset) - token_node->raw_begin;
|
|
|
|
if(token_node->raw_length)
|
|
{
|
|
token_node->type ^= (token_node->type & MyHTML_TOKEN_TYPE_WHITESPACE);
|
|
token_node->tag_id = MyHTML_TAG__TEXT;
|
|
token_node->type |= MyHTML_TOKEN_TYPE_DATA;
|
|
|
|
token_node->raw_length = (html_size + tree->global_offset) - token_node->raw_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
}
|
|
else {
|
|
token_node->type ^= (token_node->type & MyHTML_TOKEN_TYPE_WHITESPACE);
|
|
token_node->tag_id = MyHTML_TAG__COMMENT;
|
|
token_node->type |= MyHTML_TOKEN_TYPE_COMMENT;
|
|
|
|
token_node->raw_length = (html_size + tree->global_offset) - token_node->raw_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_before_attribute_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_attribute_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->attr_current->raw_key_length = (html_offset + tree->global_offset) - tree->attr_current->raw_key_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_after_attribute_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_before_attribute_value(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
tree->attr_current = myhtml_token_attr_create(tree->token, tree->token->mcasync_attr_id);
|
|
if(tree->attr_current == NULL) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_attribute_value_double_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_attribute_value_single_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_attribute_value_unquoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->attr_current->raw_value_length = (html_offset + tree->global_offset) - tree->attr_current->raw_value_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
tree->attr_current = myhtml_token_attr_create(tree->token, tree->token->mcasync_attr_id);
|
|
if(tree->attr_current == NULL) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_after_attribute_value_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment_start(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment_start_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment_end(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(token_node->raw_length > 2) {
|
|
token_node->raw_length -= 2;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment_end_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_comment_end_bang(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_bogus_comment(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_cdata_section(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
token_node->raw_length = ((html_offset + tree->global_offset) - token_node->raw_begin);
|
|
|
|
if(token_node->raw_length) {
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rcdata(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(token_node->raw_begin < (html_size + tree->global_offset)) {
|
|
token_node->type |= MyHTML_TOKEN_TYPE_RCDATA;
|
|
token_node->tag_id = MyHTML_TAG__TEXT;
|
|
token_node->raw_length = (html_size + tree->global_offset) - token_node->raw_begin;
|
|
|
|
if(myhtml_queue_add(tree, 0, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rcdata_less_than_sign(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RCDATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rcdata_end_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RCDATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rcdata_end_tag_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RCDATA);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rawtext(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RAWTEXT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rawtext_less_than_sign(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RAWTEXT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rawtext_end_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RAWTEXT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_rawtext_end_tag_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RAWTEXT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_plaintext(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
// all we need inside myhtml_tokenizer_state_plaintext
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_RAWTEXT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_before_doctype_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->compat_mode = MyHTML_TREE_COMPAT_MODE_QUIRKS;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->attr_current->raw_key_length = (html_offset + tree->global_offset) - tree->attr_current->raw_key_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_after_doctype_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_custom_after_doctype_name_a_z(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_before_doctype_public_identifier(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->compat_mode = MyHTML_TREE_COMPAT_MODE_QUIRKS;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype_public_identifier_double_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->compat_mode = MyHTML_TREE_COMPAT_MODE_QUIRKS;
|
|
|
|
if(tree->attr_current->raw_key_begin && html_size) {
|
|
tree->attr_current->raw_key_length = (html_offset + tree->global_offset) - tree->attr_current->raw_key_begin;
|
|
}
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype_public_identifier_single_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_end_state_doctype_public_identifier_double_quoted(tree, token_node, html, (html_offset + tree->global_offset), html_size);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_after_doctype_public_identifier(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype_system_identifier_double_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
tree->compat_mode = MyHTML_TREE_COMPAT_MODE_QUIRKS;
|
|
|
|
if(tree->attr_current->raw_key_begin && html_size) {
|
|
tree->attr_current->raw_key_length = (html_offset + tree->global_offset) - tree->attr_current->raw_key_begin;
|
|
}
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_doctype_system_identifier_single_quoted(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_end_state_doctype_system_identifier_double_quoted(tree, token_node, html, (html_offset + tree->global_offset), html_size);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_after_doctype_system_identifier(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_bogus_doctype(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
if(myhtml_queue_add(tree, html_offset, token_node) != MyHTML_STATUS_OK) {
|
|
myhtml_tokenizer_state_set(tree) = MyHTML_TOKENIZER_STATE_PARSE_ERROR_STOP;
|
|
return 0;
|
|
}
|
|
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_less_than_sign(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_end_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_end_tag_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escape_start(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escape_start_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped_dash_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped_less_than_sign(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped_end_tag_open(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_escaped_end_tag_name(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escape_start(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escaped(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escaped_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escaped_dash_dash(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escaped_less_than_sign(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_script_data_double_escape_end(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
token_node->element_length = (tree->global_offset + html_size) - token_node->element_begin;
|
|
|
|
myhtml_tokenizer_queue_create_text_node_if_need(tree, token_node, html, (html_offset + tree->global_offset), MyHTML_TOKEN_TYPE_SCRIPT);
|
|
return html_offset;
|
|
}
|
|
|
|
size_t myhtml_tokenizer_end_state_parse_error_stop(myhtml_tree_t* tree, myhtml_token_node_t* token_node, const char* html, size_t html_offset, size_t html_size)
|
|
{
|
|
|
|
|
|
return html_size;
|
|
}
|