Erster Docker-Stand
This commit is contained in:
165
_node_modules/@electric-sql/pglite-tools/CHANGELOG.md
generated
Normal file
165
_node_modules/@electric-sql/pglite-tools/CHANGELOG.md
generated
Normal file
@@ -0,0 +1,165 @@
|
||||
# @electric-sql/pglite-tools
|
||||
|
||||
## 0.2.20
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [45bff97]
|
||||
- Updated dependencies [5ec474f]
|
||||
- @electric-sql/pglite@0.3.15
|
||||
|
||||
## 0.2.19
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [8785034]
|
||||
- Updated dependencies [90cfee8]
|
||||
- @electric-sql/pglite@0.3.14
|
||||
|
||||
## 0.2.18
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- ad3d0d8: Updated pg_dump to use callback data exchange; built pg_dump with emscripten
|
||||
- Updated dependencies [ad3d0d8]
|
||||
- @electric-sql/pglite@0.3.13
|
||||
|
||||
## 0.2.17
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [ce0e74e]
|
||||
- @electric-sql/pglite@0.3.12
|
||||
|
||||
## 0.2.16
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [9a104b9]
|
||||
- @electric-sql/pglite@0.3.11
|
||||
|
||||
## 0.2.15
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [ad765ed]
|
||||
- @electric-sql/pglite@0.3.10
|
||||
|
||||
## 0.2.14
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- e40ccad: Upgrade emsdk
|
||||
- Updated dependencies [e40ccad]
|
||||
- @electric-sql/pglite@0.3.9
|
||||
|
||||
## 0.2.13
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- be677b4: fix pg_dump on Windows systems
|
||||
|
||||
When calling **pg_dump** on Windows system the function fails with an error as the one bellow.
|
||||
❗ Notice the double drive letter
|
||||
`Error: ENOENT: no such file or directory, open 'E:\C:\Users\<USERNAME>\AppData\Local\npm-cache\_npx\ba4f1959e38407b5\node_modules\@electric-sql\pglite-tools\dist\pg_dump.wasm'`
|
||||
|
||||
The problem is in execPgDump function at line
|
||||
`const blob = await fs.readFile(bin.toString().slice(7))`
|
||||
I think the intention here was to remove `file://` from the begging of the path. However this is not necesarry readFile can handle URL objects.
|
||||
Moreover this will fail on Windows becase the slice creates a path like '/C:/<USERNAME>...' and the readFile function will add the extra drive letter
|
||||
|
||||
- Updated dependencies [f12a582]
|
||||
- Updated dependencies [bd263aa]
|
||||
- @electric-sql/pglite@0.3.8
|
||||
|
||||
## 0.2.12
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [0936962]
|
||||
- @electric-sql/pglite@0.3.7
|
||||
|
||||
## 0.2.11
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [6898469]
|
||||
- Updated dependencies [469be18]
|
||||
- Updated dependencies [64e33c7]
|
||||
- @electric-sql/pglite@0.3.6
|
||||
|
||||
## 0.2.10
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 8172b72: new pg_dump wasm blob
|
||||
- Updated dependencies [6653899]
|
||||
- Updated dependencies [5f007fc]
|
||||
- @electric-sql/pglite@0.3.5
|
||||
|
||||
## 0.2.9
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 38a55d0: fix cjs/esm misconfigurations
|
||||
- Updated dependencies [1fcaa3e]
|
||||
- Updated dependencies [38a55d0]
|
||||
- Updated dependencies [aac7003]
|
||||
- Updated dependencies [8ca254d]
|
||||
- @electric-sql/pglite@0.3.4
|
||||
|
||||
## 0.2.8
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [ea2c7c7]
|
||||
- @electric-sql/pglite@0.3.3
|
||||
|
||||
## 0.2.7
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [e2c654b]
|
||||
- @electric-sql/pglite@0.3.2
|
||||
|
||||
## 0.2.6
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- Updated dependencies [713364e]
|
||||
- @electric-sql/pglite@0.3.1
|
||||
|
||||
## 0.2.5
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 317fd36: Specify a peer dependency range on @electric-sql/pglite
|
||||
- Updated dependencies [97e52f7]
|
||||
- Updated dependencies [4356024]
|
||||
- Updated dependencies [0033bc7]
|
||||
- @electric-sql/pglite@0.3.0
|
||||
|
||||
## 0.2.4
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- bbfa9f1: Restore SEARCH_PATH after pg_dump
|
||||
|
||||
## 0.2.3
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 8545760: pg_dump error messages set on the thrown Error
|
||||
- d26e658: Run a DEALLOCATE ALL after each pg_dump to cleanup the prepared statements.
|
||||
|
||||
## 0.2.2
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 17c9875: add node imports to the package.json browser excludes
|
||||
|
||||
## 0.2.1
|
||||
|
||||
### Patch Changes
|
||||
|
||||
- 6547374: Alpha version of pg_dump support in the browser and Node using a WASM build of pg_dump
|
||||
176
_node_modules/@electric-sql/pglite-tools/LICENSE
generated
Normal file
176
_node_modules/@electric-sql/pglite-tools/LICENSE
generated
Normal file
@@ -0,0 +1,176 @@
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. Definitions.
|
||||
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
||||
END OF TERMS AND CONDITIONS
|
||||
72
_node_modules/@electric-sql/pglite-tools/README.md
generated
Normal file
72
_node_modules/@electric-sql/pglite-tools/README.md
generated
Normal file
@@ -0,0 +1,72 @@
|
||||
# pglite-tools
|
||||
|
||||
A selection of tools for working with [PGlite](https://github.com/electric-sql/pglite) databases, including pg_dump.
|
||||
|
||||
Install with:
|
||||
|
||||
```bash
|
||||
npm install @electric-sql/pglite-tools
|
||||
```
|
||||
|
||||
## `pgDump`
|
||||
|
||||
pg_dump is a tool for dumping a PGlite database to a SQL file, this is a WASM build of pg_dump that can be used in a browser or other JavaScript environments. You can read more about pg_dump [in the Postgres docs](https://www.postgresql.org/docs/current/app-pgdump.html).
|
||||
|
||||
Note: pg_dump will execute `DEALLOCATE ALL;` after each dump. Since this is running on the same (single) connection, any prepared statements that you have made before running pg_dump will be affected.
|
||||
|
||||
### Options
|
||||
|
||||
- `pg`: A PGlite instance.
|
||||
- `args`: An array of arguments to pass to pg_dump - see [pg_dump docs](https://www.postgresql.org/docs/current/app-pgdump.html) for more details.
|
||||
- `fileName`: The name of the file to write the dump to, defaults to `dump.sql`.
|
||||
|
||||
There are a number of arguments that are automatically added to the end of the command, these are:
|
||||
|
||||
- `--inserts` - use inserts format for the output, this ensures that the dump can be restored by simply passing the output to `pg.exec()`.
|
||||
- `-j 1` - concurrency level, set to 1 as multithreading isn't supported.
|
||||
- `-f /tmp/out.sql` - the output file is always written to `/tmp/out.sql` in the virtual file system.
|
||||
- `-U postgres` - use the postgres user is hard coded.
|
||||
|
||||
### Returns
|
||||
|
||||
- A `File` object containing the dump.
|
||||
|
||||
### Caveats
|
||||
|
||||
- After restoring a dump, you might want to set the same search path as the initial db.
|
||||
|
||||
### Example
|
||||
|
||||
```typescript
|
||||
import { PGlite } from '@electric-sql/pglite'
|
||||
import { pgDump } from '@electric-sql/pglite-tools/pg_dump'
|
||||
|
||||
const pg = await PGlite.create()
|
||||
|
||||
// Create a table and insert some data
|
||||
await pg.exec(`
|
||||
CREATE TABLE test (
|
||||
id SERIAL PRIMARY KEY,
|
||||
name TEXT
|
||||
);
|
||||
`)
|
||||
await pg.exec(`
|
||||
INSERT INTO test (name) VALUES ('test');
|
||||
`)
|
||||
|
||||
// store the current search path so it can be used in the restored db
|
||||
const initialSearchPath = (await pg1.query<{ search_path: string }>('SHOW SEARCH_PATH;')).rows[0].search_path
|
||||
|
||||
// Dump the database to a file
|
||||
const dump = await pgDump({ pg })
|
||||
// Get the dump text - used for restore
|
||||
const dumpContent = await dump.text()
|
||||
|
||||
// Create a new database
|
||||
const restoredPG = await PGlite.create()
|
||||
// ... and restore it using the dump
|
||||
await restoredPG.exec(dumpContent)
|
||||
|
||||
// optional - after importing, set search path back to the initial one
|
||||
await restoredPG.exec(`SET search_path TO ${initialSearchPath};`);
|
||||
```
|
||||
4
_node_modules/@electric-sql/pglite-tools/dist/chunk-WAOUATYI.js
generated
vendored
Normal file
4
_node_modules/@electric-sql/pglite-tools/dist/chunk-WAOUATYI.js
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
1
_node_modules/@electric-sql/pglite-tools/dist/chunk-WAOUATYI.js.map
generated
vendored
Normal file
1
_node_modules/@electric-sql/pglite-tools/dist/chunk-WAOUATYI.js.map
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
4
_node_modules/@electric-sql/pglite-tools/dist/index.cjs
generated
vendored
Normal file
4
_node_modules/@electric-sql/pglite-tools/dist/index.cjs
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
1
_node_modules/@electric-sql/pglite-tools/dist/index.cjs.map
generated
vendored
Normal file
1
_node_modules/@electric-sql/pglite-tools/dist/index.cjs.map
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
2
_node_modules/@electric-sql/pglite-tools/dist/index.d.cts
generated
vendored
Normal file
2
_node_modules/@electric-sql/pglite-tools/dist/index.d.cts
generated
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
export { pgDump } from './pg_dump.cjs';
|
||||
import '@electric-sql/pglite';
|
||||
2
_node_modules/@electric-sql/pglite-tools/dist/index.d.ts
generated
vendored
Normal file
2
_node_modules/@electric-sql/pglite-tools/dist/index.d.ts
generated
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
export { pgDump } from './pg_dump.js';
|
||||
import '@electric-sql/pglite';
|
||||
2
_node_modules/@electric-sql/pglite-tools/dist/index.js
generated
vendored
Normal file
2
_node_modules/@electric-sql/pglite-tools/dist/index.js
generated
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
import{a as o}from"./chunk-WAOUATYI.js";export{o as pgDump};
|
||||
//# sourceMappingURL=index.js.map
|
||||
1
_node_modules/@electric-sql/pglite-tools/dist/index.js.map
generated
vendored
Normal file
1
_node_modules/@electric-sql/pglite-tools/dist/index.js.map
generated
vendored
Normal file
@@ -0,0 +1 @@
|
||||
{"version":3,"sources":[],"sourcesContent":[],"mappings":"","names":[]}
|
||||
4
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.cjs
generated
vendored
Normal file
4
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.cjs
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
1
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.cjs.map
generated
vendored
Normal file
1
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.cjs.map
generated
vendored
Normal file
File diff suppressed because one or more lines are too long
14
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.d.cts
generated
vendored
Normal file
14
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.d.cts
generated
vendored
Normal file
@@ -0,0 +1,14 @@
|
||||
import { PGlite } from '@electric-sql/pglite';
|
||||
|
||||
interface PgDumpOptions {
|
||||
pg: PGlite;
|
||||
args?: string[];
|
||||
fileName?: string;
|
||||
verbose?: boolean;
|
||||
}
|
||||
/**
|
||||
* Execute pg_dump
|
||||
*/
|
||||
declare function pgDump({ pg, args, fileName, }: PgDumpOptions): Promise<File>;
|
||||
|
||||
export { pgDump };
|
||||
14
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.d.ts
generated
vendored
Normal file
14
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.d.ts
generated
vendored
Normal file
@@ -0,0 +1,14 @@
|
||||
import { PGlite } from '@electric-sql/pglite';
|
||||
|
||||
interface PgDumpOptions {
|
||||
pg: PGlite;
|
||||
args?: string[];
|
||||
fileName?: string;
|
||||
verbose?: boolean;
|
||||
}
|
||||
/**
|
||||
* Execute pg_dump
|
||||
*/
|
||||
declare function pgDump({ pg, args, fileName, }: PgDumpOptions): Promise<File>;
|
||||
|
||||
export { pgDump };
|
||||
2
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.js
generated
vendored
Normal file
2
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.js
generated
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
import{a}from"./chunk-WAOUATYI.js";export{a as pgDump};
|
||||
//# sourceMappingURL=pg_dump.js.map
|
||||
1
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.js.map
generated
vendored
Normal file
1
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.js.map
generated
vendored
Normal file
@@ -0,0 +1 @@
|
||||
{"version":3,"sources":[],"sourcesContent":[],"mappings":"","names":[]}
|
||||
BIN
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.wasm
generated
vendored
Normal file
BIN
_node_modules/@electric-sql/pglite-tools/dist/pg_dump.wasm
generated
vendored
Normal file
Binary file not shown.
29
_node_modules/@electric-sql/pglite-tools/eslint.config.js
generated
Normal file
29
_node_modules/@electric-sql/pglite-tools/eslint.config.js
generated
Normal file
@@ -0,0 +1,29 @@
|
||||
import globals from 'globals'
|
||||
import rootConfig from '../../eslint.config.js'
|
||||
|
||||
export default [
|
||||
...rootConfig,
|
||||
{
|
||||
ignores: ['release/**/*', 'examples/**/*', 'dist/**/*'],
|
||||
},
|
||||
{
|
||||
languageOptions: {
|
||||
globals: {
|
||||
...globals.browser,
|
||||
...globals.node,
|
||||
},
|
||||
},
|
||||
rules: {
|
||||
...rootConfig.rules,
|
||||
'@typescript-eslint/no-explicit-any': 'off',
|
||||
},
|
||||
},
|
||||
{
|
||||
files: ['tests/targets/deno/**/*.js'],
|
||||
languageOptions: {
|
||||
globals: {
|
||||
Deno: false,
|
||||
},
|
||||
},
|
||||
},
|
||||
]
|
||||
76
_node_modules/@electric-sql/pglite-tools/package.json
generated
Normal file
76
_node_modules/@electric-sql/pglite-tools/package.json
generated
Normal file
@@ -0,0 +1,76 @@
|
||||
{
|
||||
"name": "@electric-sql/pglite-tools",
|
||||
"version": "0.2.20",
|
||||
"description": "Tools for working with PGlite databases",
|
||||
"author": "Electric DB Limited",
|
||||
"homepage": "https://pglite.dev",
|
||||
"license": "Apache-2.0",
|
||||
"repository": {
|
||||
"type": "git",
|
||||
"url": "git+https://github.com/electric-sql/pglite",
|
||||
"directory": "packages/pglite-tools"
|
||||
},
|
||||
"keywords": [
|
||||
"postgres",
|
||||
"sql",
|
||||
"database",
|
||||
"wasm",
|
||||
"pglite",
|
||||
"pg_dump",
|
||||
"pg_restore"
|
||||
],
|
||||
"private": false,
|
||||
"publishConfig": {
|
||||
"access": "public"
|
||||
},
|
||||
"type": "module",
|
||||
"main": "./dist/index.cjs",
|
||||
"module": "./dist/index.js",
|
||||
"types": "./dist/index.d.ts",
|
||||
"exports": {
|
||||
".": {
|
||||
"import": {
|
||||
"types": "./dist/index.d.ts",
|
||||
"default": "./dist/index.js"
|
||||
},
|
||||
"require": {
|
||||
"types": "./dist/index.d.cts",
|
||||
"default": "./dist/index.cjs"
|
||||
}
|
||||
},
|
||||
"./pg_dump": {
|
||||
"import": {
|
||||
"types": "./dist/pg_dump.d.ts",
|
||||
"default": "./dist/pg_dump.js"
|
||||
},
|
||||
"require": {
|
||||
"types": "./dist/pg_dump.d.cts",
|
||||
"default": "./dist/pg_dump.cjs"
|
||||
}
|
||||
}
|
||||
},
|
||||
"browser": {
|
||||
"fs": false,
|
||||
"fs/promises": false
|
||||
},
|
||||
"devDependencies": {
|
||||
"@arethetypeswrong/cli": "^0.18.1",
|
||||
"@types/emscripten": "^1.41.1",
|
||||
"@types/node": "^20.16.11",
|
||||
"tsx": "^4.19.2",
|
||||
"vitest": "^1.3.1",
|
||||
"@electric-sql/pglite": "0.3.15"
|
||||
},
|
||||
"peerDependencies": {
|
||||
"@electric-sql/pglite": "0.3.15"
|
||||
},
|
||||
"scripts": {
|
||||
"build": "tsup",
|
||||
"check:exports": "attw . --pack --profile node16",
|
||||
"lint": "eslint ./src ./tests --report-unused-disable-directives --max-warnings 0",
|
||||
"format": "prettier --write ./src ./tests",
|
||||
"typecheck": "tsc",
|
||||
"stylecheck": "pnpm lint && prettier --check ./src ./tests",
|
||||
"test": "vitest"
|
||||
}
|
||||
}
|
||||
1
_node_modules/@electric-sql/pglite-tools/src/index.ts
generated
Normal file
1
_node_modules/@electric-sql/pglite-tools/src/index.ts
generated
Normal file
@@ -0,0 +1 @@
|
||||
export * from './pg_dump'
|
||||
41
_node_modules/@electric-sql/pglite-tools/src/pgDumpModFactory.ts
generated
Normal file
41
_node_modules/@electric-sql/pglite-tools/src/pgDumpModFactory.ts
generated
Normal file
@@ -0,0 +1,41 @@
|
||||
import PgDumpModFactory from '../release/pg_dump'
|
||||
|
||||
type IDBFS = Emscripten.FileSystemType & {
|
||||
quit: () => void
|
||||
dbs: Record<string, IDBDatabase>
|
||||
}
|
||||
|
||||
export type FS = typeof FS & {
|
||||
filesystems: {
|
||||
MEMFS: Emscripten.FileSystemType
|
||||
NODEFS: Emscripten.FileSystemType
|
||||
IDBFS: IDBFS
|
||||
}
|
||||
quit: () => void
|
||||
}
|
||||
|
||||
export interface PgDumpMod
|
||||
extends Omit<EmscriptenModule, 'preInit' | 'preRun' | 'postRun'> {
|
||||
preInit: Array<{ (mod: PgDumpMod): void }>
|
||||
preRun: Array<{ (mod: PgDumpMod): void }>
|
||||
postRun: Array<{ (mod: PgDumpMod): void }>
|
||||
FS: FS
|
||||
WASM_PREFIX: string
|
||||
INITIAL_MEMORY: number
|
||||
_set_read_write_cbs: (read_cb: number, write_cb: number) => void
|
||||
addFunction: (
|
||||
cb: (ptr: any, length: number) => void,
|
||||
signature: string,
|
||||
) => number
|
||||
removeFunction: (f: number) => void
|
||||
_main: (args: string[]) => number
|
||||
onExit: (status: number) => void
|
||||
print: (test: string) => void
|
||||
printErr: (text: string) => void
|
||||
}
|
||||
|
||||
type PgDumpFactory<T extends PgDumpMod = PgDumpMod> = (
|
||||
moduleOverrides?: Partial<T>,
|
||||
) => Promise<T>
|
||||
|
||||
export default PgDumpModFactory as PgDumpFactory<PgDumpMod>
|
||||
157
_node_modules/@electric-sql/pglite-tools/src/pg_dump.ts
generated
Normal file
157
_node_modules/@electric-sql/pglite-tools/src/pg_dump.ts
generated
Normal file
@@ -0,0 +1,157 @@
|
||||
import { PGlite } from '@electric-sql/pglite'
|
||||
import PgDumpModFactory, { PgDumpMod } from './pgDumpModFactory'
|
||||
|
||||
const dumpFilePath = '/tmp/out.sql'
|
||||
|
||||
/**
|
||||
* Creates a new Uint8Array based on two different ArrayBuffers
|
||||
*
|
||||
* @private
|
||||
* @param {ArrayBuffers} buffer1 The first buffer.
|
||||
* @param {ArrayBuffers} buffer2 The second buffer.
|
||||
* @return {ArrayBuffers} The new ArrayBuffer created out of the two.
|
||||
*/
|
||||
function concat(buffer1: ArrayBuffer, buffer2: ArrayBuffer) {
|
||||
const tmp = new Uint8Array(buffer1.byteLength + buffer2.byteLength)
|
||||
tmp.set(new Uint8Array(buffer1), 0)
|
||||
tmp.set(new Uint8Array(buffer2), buffer1.byteLength)
|
||||
return tmp
|
||||
}
|
||||
|
||||
interface ExecResult {
|
||||
exitCode: number
|
||||
fileContents: string
|
||||
stderr: string
|
||||
stdout: string
|
||||
}
|
||||
|
||||
/**
|
||||
* Inner function to execute pg_dump
|
||||
*/
|
||||
async function execPgDump({
|
||||
pg,
|
||||
args,
|
||||
}: {
|
||||
pg: PGlite
|
||||
args: string[]
|
||||
}): Promise<ExecResult> {
|
||||
let pgdump_write, pgdump_read
|
||||
let exitStatus = 0
|
||||
let stderrOutput: string = ''
|
||||
let stdoutOutput: string = ''
|
||||
const emscriptenOpts: Partial<PgDumpMod> = {
|
||||
arguments: args,
|
||||
noExitRuntime: false,
|
||||
print: (text) => {
|
||||
stdoutOutput += text
|
||||
},
|
||||
printErr: (text) => {
|
||||
stderrOutput += text
|
||||
},
|
||||
onExit: (status: number) => {
|
||||
exitStatus = status
|
||||
},
|
||||
preRun: [
|
||||
(mod: PgDumpMod) => {
|
||||
mod.onRuntimeInitialized = () => {
|
||||
let bufferedBytes: Uint8Array = new Uint8Array()
|
||||
|
||||
pgdump_write = mod.addFunction((ptr: any, length: number) => {
|
||||
let bytes
|
||||
try {
|
||||
bytes = mod.HEAPU8.subarray(ptr, ptr + length)
|
||||
} catch (e: any) {
|
||||
console.error('error', e)
|
||||
throw e
|
||||
}
|
||||
const currentResponse = pg.execProtocolRawSync(bytes)
|
||||
bufferedBytes = concat(bufferedBytes, currentResponse)
|
||||
return length
|
||||
}, 'iii')
|
||||
|
||||
pgdump_read = mod.addFunction((ptr: any, max_length: number) => {
|
||||
let length = bufferedBytes.length
|
||||
if (length > max_length) {
|
||||
length = max_length
|
||||
}
|
||||
try {
|
||||
mod.HEAP8.set(bufferedBytes.subarray(0, length), ptr)
|
||||
} catch (e) {
|
||||
console.error(e)
|
||||
}
|
||||
bufferedBytes = bufferedBytes.subarray(length, bufferedBytes.length)
|
||||
return length
|
||||
}, 'iii')
|
||||
|
||||
mod._set_read_write_cbs(pgdump_read, pgdump_write)
|
||||
// default $HOME in emscripten is /home/web_user
|
||||
mod.FS.chmod('/home/web_user/.pgpass', 0o0600) // https://www.postgresql.org/docs/current/libpq-pgpass.html
|
||||
}
|
||||
},
|
||||
],
|
||||
}
|
||||
|
||||
const mod = await PgDumpModFactory(emscriptenOpts)
|
||||
let fileContents = ''
|
||||
if (!exitStatus) {
|
||||
fileContents = mod.FS.readFile(dumpFilePath, { encoding: 'utf8' })
|
||||
}
|
||||
|
||||
return {
|
||||
exitCode: exitStatus,
|
||||
fileContents,
|
||||
stderr: stderrOutput,
|
||||
stdout: stdoutOutput,
|
||||
}
|
||||
}
|
||||
|
||||
interface PgDumpOptions {
|
||||
pg: PGlite
|
||||
args?: string[]
|
||||
fileName?: string
|
||||
verbose?: boolean
|
||||
}
|
||||
|
||||
/**
|
||||
* Execute pg_dump
|
||||
*/
|
||||
export async function pgDump({
|
||||
pg,
|
||||
args,
|
||||
fileName = 'dump.sql',
|
||||
}: PgDumpOptions) {
|
||||
const getSearchPath = await pg.query<{ search_path: string }>(
|
||||
'SHOW SEARCH_PATH;',
|
||||
)
|
||||
const search_path = getSearchPath.rows[0].search_path
|
||||
|
||||
const baseArgs = [
|
||||
'-U',
|
||||
'postgres',
|
||||
'--inserts',
|
||||
'-j',
|
||||
'1',
|
||||
'-f',
|
||||
dumpFilePath,
|
||||
'postgres',
|
||||
]
|
||||
|
||||
const execResult = await execPgDump({
|
||||
pg,
|
||||
args: [...(args ?? []), ...baseArgs],
|
||||
})
|
||||
|
||||
pg.exec(`DEALLOCATE ALL; SET SEARCH_PATH = ${search_path}`)
|
||||
|
||||
if (execResult.exitCode !== 0) {
|
||||
throw new Error(
|
||||
`pg_dump failed with exit code ${execResult.exitCode}. \nError message: ${execResult.stderr}`,
|
||||
)
|
||||
}
|
||||
|
||||
const file = new File([execResult.fileContents], fileName, {
|
||||
type: 'text/plain',
|
||||
})
|
||||
|
||||
return file
|
||||
}
|
||||
201
_node_modules/@electric-sql/pglite-tools/tests/pg_dump.test.ts
generated
Normal file
201
_node_modules/@electric-sql/pglite-tools/tests/pg_dump.test.ts
generated
Normal file
@@ -0,0 +1,201 @@
|
||||
import { describe, it, expect } from 'vitest'
|
||||
import { PGlite } from '@electric-sql/pglite'
|
||||
import { pgDump } from '../dist/pg_dump.js'
|
||||
import * as fs from 'fs/promises'
|
||||
|
||||
describe('pgDump', () => {
|
||||
it('should dump an empty database', async () => {
|
||||
const pg = await PGlite.create()
|
||||
const dump = await pgDump({ pg })
|
||||
|
||||
expect(dump).toBeInstanceOf(File)
|
||||
expect(dump.name).toBe('dump.sql')
|
||||
|
||||
const content = await dump.text()
|
||||
expect(content).toContain('PostgreSQL database dump')
|
||||
})
|
||||
|
||||
it('should dump an empty database multiple times', async () => {
|
||||
const pg = await PGlite.create()
|
||||
|
||||
for (let i = 0; i < 5; i++) {
|
||||
const fileName = `dump_${i}.sql`
|
||||
const dump = await pgDump({ pg, fileName })
|
||||
|
||||
expect(dump).toBeInstanceOf(File)
|
||||
expect(dump.name).toBe(fileName)
|
||||
|
||||
const content = await dump.text()
|
||||
expect(content).toContain('PostgreSQL database dump')
|
||||
}
|
||||
})
|
||||
|
||||
it('should dump a database with tables and data', async () => {
|
||||
const pg = await PGlite.create()
|
||||
|
||||
// Create test tables and insert data
|
||||
await pg.exec(`
|
||||
CREATE TABLE test1 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
name TEXT
|
||||
);
|
||||
INSERT INTO test1 (name) VALUES ('test1-row1');
|
||||
|
||||
CREATE TABLE test2 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
value INTEGER
|
||||
);
|
||||
INSERT INTO test2 (value) VALUES (42);
|
||||
`)
|
||||
|
||||
const dump = await pgDump({ pg })
|
||||
const content = await dump.text()
|
||||
|
||||
// Check for table creation
|
||||
expect(content).toContain('CREATE TABLE public.test1')
|
||||
expect(content).toContain('CREATE TABLE public.test2')
|
||||
|
||||
// Check for data inserts
|
||||
expect(content).toContain('INSERT INTO public.test1')
|
||||
expect(content).toContain("'test1-row1'")
|
||||
expect(content).toContain('INSERT INTO public.test2')
|
||||
expect(content).toContain('42')
|
||||
})
|
||||
|
||||
it('should respect custom filename', async () => {
|
||||
const pg = await PGlite.create()
|
||||
const dump = await pgDump({ pg, fileName: 'custom.sql' })
|
||||
|
||||
expect(dump.name).toBe('custom.sql')
|
||||
})
|
||||
|
||||
it('should handle custom pg_dump arguments', async () => {
|
||||
const pg = await PGlite.create()
|
||||
await pg.exec(`
|
||||
CREATE TABLE test (id SERIAL PRIMARY KEY, name TEXT);
|
||||
INSERT INTO test (name) VALUES ('row1');
|
||||
`)
|
||||
|
||||
// Use --schema-only to exclude data
|
||||
const dump = await pgDump({ pg, args: ['--schema-only'] })
|
||||
const content = await dump.text()
|
||||
|
||||
expect(content).toContain('CREATE TABLE public.test')
|
||||
expect(content).not.toContain('INSERT INTO public.test')
|
||||
})
|
||||
|
||||
it('should be able to restore dumped database', async () => {
|
||||
const pg1 = await PGlite.create()
|
||||
|
||||
// Create original database
|
||||
await pg1.exec(`
|
||||
CREATE TABLE test (id SERIAL PRIMARY KEY, name TEXT);
|
||||
INSERT INTO test (name) VALUES ('row1'), ('row2');
|
||||
`)
|
||||
|
||||
const initialSearchPath = (
|
||||
await pg1.query<{ search_path: string }>('SHOW SEARCH_PATH;')
|
||||
).rows[0].search_path
|
||||
|
||||
// Dump database
|
||||
const dump = await pgDump({ pg: pg1 })
|
||||
const dumpContent = await dump.text()
|
||||
|
||||
// Create new database and restore
|
||||
const pg2 = await PGlite.create()
|
||||
await pg2.exec(dumpContent)
|
||||
|
||||
// after importing, set search path back to the initial one
|
||||
await pg2.exec(`SET search_path TO ${initialSearchPath};`)
|
||||
|
||||
// Verify data
|
||||
const result = await pg2.query<{ name: string }>(
|
||||
'SELECT * FROM test ORDER BY id',
|
||||
)
|
||||
expect(result.rows).toHaveLength(2)
|
||||
expect(result.rows[0].name).toBe('row1')
|
||||
expect(result.rows[1].name).toBe('row2')
|
||||
})
|
||||
|
||||
it('pg_dump should not change SEARCH_PATH', async () => {
|
||||
const pg = await PGlite.create()
|
||||
|
||||
await pg.exec(`SET SEARCH_PATH = amigo;`)
|
||||
const initialSearchPath = await pg.query('SHOW SEARCH_PATH;')
|
||||
|
||||
const dump = await pgDump({ pg })
|
||||
await dump.text()
|
||||
|
||||
const finalSearchPath = await pg.query('SHOW SEARCH_PATH;')
|
||||
|
||||
expect(initialSearchPath).toEqual(finalSearchPath)
|
||||
})
|
||||
|
||||
it('specify datadir: should dump a database with tables and data', async () => {
|
||||
const dataDir = '/tmp/pg_dump_pglite_data_dir'
|
||||
await fs.rm(dataDir, { force: true, recursive: true })
|
||||
const pg = await PGlite.create({
|
||||
dataDir: dataDir,
|
||||
})
|
||||
|
||||
// Create test tables and insert data
|
||||
await pg.exec(`
|
||||
CREATE TABLE test1 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
name TEXT
|
||||
);
|
||||
INSERT INTO test1 (name) VALUES ('test1-row1');
|
||||
|
||||
CREATE TABLE test2 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
value INTEGER
|
||||
);
|
||||
INSERT INTO test2 (value) VALUES (42);
|
||||
`)
|
||||
|
||||
const dump = await pgDump({ pg })
|
||||
const content = await dump.text()
|
||||
|
||||
// Check for table creation
|
||||
expect(content).toContain('CREATE TABLE public.test1')
|
||||
expect(content).toContain('CREATE TABLE public.test2')
|
||||
|
||||
// Check for data inserts
|
||||
expect(content).toContain('INSERT INTO public.test1')
|
||||
expect(content).toContain("'test1-row1'")
|
||||
expect(content).toContain('INSERT INTO public.test2')
|
||||
expect(content).toContain('42')
|
||||
})
|
||||
|
||||
it('param --quote-all-identifiers should work', async () => {
|
||||
const pg = await PGlite.create()
|
||||
|
||||
// Create test tables and insert data
|
||||
await pg.exec(`
|
||||
CREATE TABLE test1 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
name TEXT
|
||||
);
|
||||
INSERT INTO test1 (name) VALUES ('test1-row1');
|
||||
|
||||
CREATE TABLE test2 (
|
||||
id SERIAL PRIMARY KEY,
|
||||
value INTEGER
|
||||
);
|
||||
INSERT INTO test2 (value) VALUES (42);
|
||||
`)
|
||||
|
||||
const dump = await pgDump({ pg, args: ['--quote-all-identifiers'] })
|
||||
const content = await dump.text()
|
||||
|
||||
// Check for table creation
|
||||
expect(content).toContain('CREATE TABLE "public"."test1"')
|
||||
expect(content).toContain('CREATE TABLE "public"."test2"')
|
||||
|
||||
// Check for data inserts
|
||||
expect(content).toContain('INSERT INTO "public"."test1"')
|
||||
expect(content).toContain("'test1-row1'")
|
||||
expect(content).toContain('INSERT INTO "public"."test2"')
|
||||
expect(content).toContain('42')
|
||||
})
|
||||
})
|
||||
15
_node_modules/@electric-sql/pglite-tools/tests/setup.ts
generated
Normal file
15
_node_modules/@electric-sql/pglite-tools/tests/setup.ts
generated
Normal file
@@ -0,0 +1,15 @@
|
||||
import { beforeAll } from 'vitest'
|
||||
import { execSync } from 'child_process'
|
||||
import { existsSync } from 'fs'
|
||||
import { join } from 'path'
|
||||
|
||||
beforeAll(() => {
|
||||
// Check if we need to build
|
||||
const distPath = join(__dirname, '../dist')
|
||||
const wasmPath = join(distPath, 'pg_dump.wasm')
|
||||
|
||||
if (!existsSync(wasmPath)) {
|
||||
console.log('Building project before running tests...')
|
||||
execSync('pnpm build', { stdio: 'inherit' })
|
||||
}
|
||||
})
|
||||
10
_node_modules/@electric-sql/pglite-tools/tsconfig.json
generated
Normal file
10
_node_modules/@electric-sql/pglite-tools/tsconfig.json
generated
Normal file
@@ -0,0 +1,10 @@
|
||||
{
|
||||
"extends": "../../tsconfig.json",
|
||||
"compilerOptions": {
|
||||
"types": [
|
||||
"@types/emscripten",
|
||||
"node"
|
||||
]
|
||||
},
|
||||
"include": ["src", "tsup.config.ts", "vitest.config.ts"]
|
||||
}
|
||||
28
_node_modules/@electric-sql/pglite-tools/tsup.config.ts
generated
Normal file
28
_node_modules/@electric-sql/pglite-tools/tsup.config.ts
generated
Normal file
@@ -0,0 +1,28 @@
|
||||
import { cpSync } from 'fs'
|
||||
import { resolve } from 'path'
|
||||
import { defineConfig } from 'tsup'
|
||||
|
||||
const entryPoints = [
|
||||
'src/index.ts',
|
||||
'src/pg_dump.ts',
|
||||
]
|
||||
|
||||
const minify = process.env.DEBUG === 'true' ? false : true
|
||||
|
||||
export default defineConfig([
|
||||
{
|
||||
entry: entryPoints,
|
||||
sourcemap: true,
|
||||
dts: {
|
||||
entry: entryPoints,
|
||||
resolve: true,
|
||||
},
|
||||
clean: true,
|
||||
minify: minify,
|
||||
shims: true,
|
||||
format: ['esm', 'cjs'],
|
||||
onSuccess: async () => {
|
||||
cpSync(resolve('release/pg_dump.wasm'), resolve('dist/pg_dump.wasm'))
|
||||
}
|
||||
},
|
||||
])
|
||||
10
_node_modules/@electric-sql/pglite-tools/vitest.config.ts
generated
Normal file
10
_node_modules/@electric-sql/pglite-tools/vitest.config.ts
generated
Normal file
@@ -0,0 +1,10 @@
|
||||
import { defineConfig } from 'vitest/config'
|
||||
|
||||
export default defineConfig({
|
||||
test: {
|
||||
globals: true,
|
||||
environment: 'node',
|
||||
testTimeout: 30000,
|
||||
setupFiles: ['./tests/setup.ts'],
|
||||
},
|
||||
})
|
||||
Reference in New Issue
Block a user