@@ -199,10 +199,8 @@ function text(node, state) {
199
199
200
200
resetTokenizer ( state , pointStart ( node ) )
201
201
// @ts -expect-error: private.
202
- // type-coverage:ignore-next-line
203
202
state . parser . currentToken = token
204
203
// @ts -expect-error: private.
205
- // type-coverage:ignore-next-line
206
204
state . parser . _processToken ( state . parser . currentToken )
207
205
}
208
206
@@ -229,10 +227,8 @@ function doctype(node, state) {
229
227
230
228
resetTokenizer ( state , pointStart ( node ) )
231
229
// @ts -expect-error: private.
232
- // type-coverage:ignore-next-line
233
230
state . parser . currentToken = token
234
231
// @ts -expect-error: private.
235
- // type-coverage:ignore-next-line
236
232
state . parser . _processToken ( state . parser . currentToken )
237
233
}
238
234
@@ -292,10 +288,8 @@ function comment(node, state) {
292
288
}
293
289
resetTokenizer ( state , pointStart ( node ) )
294
290
// @ts -expect-error: private.
295
- // type-coverage:ignore-next-line
296
291
state . parser . currentToken = token
297
292
// @ts -expect-error: private.
298
- // type-coverage:ignore-next-line
299
293
state . parser . _processToken ( state . parser . currentToken )
300
294
}
301
295
@@ -313,8 +307,6 @@ function handleRaw(node, state) {
313
307
// Reset preprocessor:
314
308
// See: <https://github.com/inikulin/parse5/blob/6f7ca60/packages/parse5/lib/tokenizer/preprocessor.ts#L18-L31>.
315
309
state . parser . tokenizer . preprocessor . html = ''
316
- // @ts -expect-error: private.
317
- // type-coverage:ignore-next-line
318
310
state . parser . tokenizer . preprocessor . pos = - 1
319
311
// @ts -expect-error: private.
320
312
// type-coverage:ignore-next-line
@@ -325,8 +317,6 @@ function handleRaw(node, state) {
325
317
// @ts -expect-error: private.
326
318
// type-coverage:ignore-next-line
327
319
state . parser . tokenizer . preprocessor . skipNextNewLine = false
328
- // @ts -expect-error: private.
329
- // type-coverage:ignore-next-line
330
320
state . parser . tokenizer . preprocessor . lastChunkWritten = false
331
321
state . parser . tokenizer . preprocessor . endOfChunkHit = false
332
322
// @ts -expect-error: private.
@@ -337,7 +327,6 @@ function handleRaw(node, state) {
337
327
setPoint ( state , pointStart ( node ) )
338
328
state . parser . tokenizer . write ( node . value , false )
339
329
// @ts -expect-error: private.
340
- // type-coverage:ignore-next-line
341
330
state . parser . tokenizer . _runParsingLoop ( )
342
331
343
332
// Character references hang, so if we ended there, we need to flush
@@ -350,19 +339,18 @@ function handleRaw(node, state) {
350
339
// See: <https://github.com/inikulin/parse5/blob/46cba43/packages/parse5/lib/tokenizer/index.ts#L58>
351
340
// Note: a change to `parse5`, which breaks this, was merged but not released.
352
341
// Investigate when it is.
342
+ // To do: remove next major.
343
+ /* c8 ignore next 12 -- removed in <https://github.com/inikulin/parse5/pull/897> */
353
344
if (
354
345
state . parser . tokenizer . state === 72 /* NAMED_CHARACTER_REFERENCE */ ||
346
+ // @ts -expect-error: removed.
355
347
state . parser . tokenizer . state === 78 /* NUMERIC_CHARACTER_REFERENCE_END */
356
348
) {
357
- // @ts -expect-error: private.
358
- // type-coverage:ignore-next-line
359
349
state . parser . tokenizer . preprocessor . lastChunkWritten = true
360
350
/** @type {number } */
361
351
// @ts -expect-error: private.
362
- // type-coverage:ignore-next-line
363
352
const cp = state . parser . tokenizer . _consume ( )
364
353
// @ts -expect-error: private.
365
- // type-coverage:ignore-next-line
366
354
state . parser . tokenizer . _callState ( cp )
367
355
}
368
356
}
@@ -413,18 +401,15 @@ function resetTokenizer(state, point) {
413
401
// Process final characters if they’re still there after hibernating.
414
402
/** @type {Token.CharacterToken } */
415
403
// @ts -expect-error: private.
416
- // type-coverage:ignore-next-line
417
404
const token = state . parser . tokenizer . currentCharacterToken
418
405
419
406
if ( token && token . location ) {
420
407
token . location . endLine = state . parser . tokenizer . preprocessor . line
421
408
token . location . endCol = state . parser . tokenizer . preprocessor . col + 1
422
409
token . location . endOffset = state . parser . tokenizer . preprocessor . offset + 1
423
410
// @ts -expect-error: private.
424
- // type-coverage:ignore-next-line
425
411
state . parser . currentToken = token
426
412
// @ts -expect-error: private.
427
- // type-coverage:ignore-next-line
428
413
state . parser . _processToken ( state . parser . currentToken )
429
414
}
430
415
@@ -436,35 +421,26 @@ function resetTokenizer(state, point) {
436
421
// But also if broken HTML is in `raw`, and then a correct element is given.
437
422
// See GH-11.
438
423
// @ts -expect-error: private.
439
- // type-coverage:ignore-next-line
440
424
state . parser . tokenizer . paused = false
441
425
// @ts -expect-error: private.
442
- // type-coverage:ignore-next-line
443
426
state . parser . tokenizer . inLoop = false
444
427
445
428
// Note: don’t reset `state`, `inForeignNode`, or `lastStartTagName`, we
446
429
// manually update those when needed.
447
430
state . parser . tokenizer . active = false
448
431
// @ts -expect-error: private.
449
- // type-coverage:ignore-next-line
450
432
state . parser . tokenizer . returnState = TokenizerMode . DATA
451
433
// @ts -expect-error: private.
452
- // type-coverage:ignore-next-line
453
434
state . parser . tokenizer . charRefCode = - 1
454
435
// @ts -expect-error: private.
455
- // type-coverage:ignore-next-line
456
436
state . parser . tokenizer . consumedAfterSnapshot = - 1
457
437
// @ts -expect-error: private.
458
- // type-coverage:ignore-next-line
459
438
state . parser . tokenizer . currentLocation = null
460
439
// @ts -expect-error: private.
461
- // type-coverage:ignore-next-line
462
440
state . parser . tokenizer . currentCharacterToken = null
463
441
// @ts -expect-error: private.
464
- // type-coverage:ignore-next-line
465
442
state . parser . tokenizer . currentToken = null
466
443
// @ts -expect-error: private.
467
- // type-coverage:ignore-next-line
468
444
state . parser . tokenizer . currentAttr = { name : '' , value : '' }
469
445
}
470
446
@@ -496,7 +472,6 @@ function setPoint(state, point) {
496
472
state . parser . tokenizer . preprocessor . droppedBufferSize = point . offset
497
473
state . parser . tokenizer . preprocessor . line = point . line
498
474
// @ts -expect-error: private.
499
- // type-coverage:ignore-next-line
500
475
state . parser . tokenizer . currentLocation = location
501
476
}
502
477
}
@@ -552,10 +527,8 @@ function startTag(node, state) {
552
527
// So we act *as if* the tokenizer emits tokens:
553
528
554
529
// @ts -expect-error: private.
555
- // type-coverage:ignore-next-line
556
530
state . parser . currentToken = tag
557
531
// @ts -expect-error: private.
558
- // type-coverage:ignore-next-line
559
532
state . parser . _processToken ( state . parser . currentToken )
560
533
561
534
// …but then we still need a bunch of work that the tokenizer would normally
@@ -609,10 +582,8 @@ function endTag(node, state) {
609
582
// So we act *as if* the tokenizer emits tokens:
610
583
611
584
// @ts -expect-error: private.
612
- // type-coverage:ignore-next-line
613
585
state . parser . currentToken = tag
614
586
// @ts -expect-error: private.
615
- // type-coverage:ignore-next-line
616
587
state . parser . _processToken ( state . parser . currentToken )
617
588
618
589
// …but then we still need a bunch of work that the tokenizer would normally
0 commit comments