summaryrefslogtreecommitdiff
path: root/packages/integrations/markdoc/src
diff options
context:
space:
mode:
authorGravatar Erika <3019731+Princesseuh@users.noreply.github.com> 2023-09-13 16:49:22 +0200
committerGravatar GitHub <noreply@github.com> 2023-09-13 16:49:22 +0200
commit4e395a6cab27d79a785ca386114b6229d8f6a7b9 (patch)
tree5b02e0c7cd6412cef2d4f6a3370f5363e2f0d34b /packages/integrations/markdoc/src
parentb85c8a78a116dbbddc901438bc0b7a1917dc0238 (diff)
downloadastro-4e395a6cab27d79a785ca386114b6229d8f6a7b9.tar.gz
astro-4e395a6cab27d79a785ca386114b6229d8f6a7b9.tar.zst
astro-4e395a6cab27d79a785ca386114b6229d8f6a7b9.zip
config: migrate us to moduleResolution: 'node16' (#8519)
Diffstat (limited to 'packages/integrations/markdoc/src')
-rw-r--r--packages/integrations/markdoc/src/html/index.ts4
-rw-r--r--packages/integrations/markdoc/src/html/transform/html-token-transform.ts1
2 files changed, 3 insertions, 2 deletions
diff --git a/packages/integrations/markdoc/src/html/index.ts b/packages/integrations/markdoc/src/html/index.ts
index 8798d3c9a..3f947736c 100644
--- a/packages/integrations/markdoc/src/html/index.ts
+++ b/packages/integrations/markdoc/src/html/index.ts
@@ -1,2 +1,2 @@
-export { htmlTag } from './tagdefs/html.tag';
-export { htmlTokenTransform } from './transform/html-token-transform';
+export { htmlTag } from './tagdefs/html.tag.js';
+export { htmlTokenTransform } from './transform/html-token-transform.js';
diff --git a/packages/integrations/markdoc/src/html/transform/html-token-transform.ts b/packages/integrations/markdoc/src/html/transform/html-token-transform.ts
index cfa511a9f..10796cdc0 100644
--- a/packages/integrations/markdoc/src/html/transform/html-token-transform.ts
+++ b/packages/integrations/markdoc/src/html/transform/html-token-transform.ts
@@ -1,5 +1,6 @@
import type { Tokenizer } from '@markdoc/markdoc';
import { Parser } from 'htmlparser2';
+// @ts-expect-error This type isn't exported
import type * as Token from 'markdown-it/lib/token';
export function htmlTokenTransform(tokenizer: Tokenizer, tokens: Token[]): Token[] {