@diqi/pg-to-ts v5.0.2
pg-to-ts
pg-to-ts generates TypeScript types that match your Postgres database schema.
It works by querying the Postgres metadata schema (pg_catalog) and generating
equivalent TypeScript types, as well as some JavaScript values that can be
helpful for generating queries at runtime.
Usage:
npm install pg-to-ts
pg-to-ts generate -c postgresql://user:pass@host/db -o dbschema.tsThe resulting file looks like:
// Table product
export interface Product {
id: string;
name: string;
description: string;
created_at: Date;
}
export interface ProductInput {
id?: string;
name: string;
description: string;
created_at?: Date;
}
const product = {
tableName: 'product',
columns: ['id', 'name', 'description', 'created_at'],
requiredForInsert: ['name', 'description'],
}
export interface TableTypes {
product: {
select: Product;
input: ProductInput;
};
}
export const tables = {
product,
};This gives you most of the types you need for static analysis and runtime.
This is a fork of PYST/schemats, which is a fork of SweetIQ/schemats. Compared to those projects, this fork:
- Drops support for MySQL in favor of deeper support for Postgres.
- Significantly modernizes the infrastructure and dependencies.
- Adds a few new features (see below).
Schema Features
Comments
If you set a Postgres comment on a table or column:
COMMENT ON TABLE product IS 'Table containing products';
COMMENT ON COLUMN product.name IS 'Human-readable product name';Then these come out as JSDoc comments in the schema:
/** Table containing products */
export interface Product {
id: string;
/** Human-readable product name */
name: string;
description: string;
created_at: Date;
}The TypeScript language service will surface these when it's helpful.
Dates as strings
node-postgres returns timestamp columns as JavaScript Date objects. This makes a lot of sense, but it can lead to problems if you try to serialize them as JSON, which converts them to strings. This means that the serialized and de- serialized table types will be different.
By default pg-to-ts will put Date types in your schema file, but if you'd
prefer strings, pass --datesAsStrings. Note that you'll be responsible for
making sure that timestamps/dates really do come back as strings, not Date objects.
See https://github.com/brianc/node-pg-types for details.
JSON types
By default, Postgres json and jsonb columns will be typed as unknown.
This is safe but not very precise, and it can make them cumbersome to work with.
Oftentimes you know what the type should be.
To tell pg-to-ts to use a specific TypeScript type for a json column, use
a JSDoc @type annotation:
ALTER TABLE product ADD COLUMN metadata jsonb;
COMMENT ON COLUMN product.metadata is 'Additional information @type {ProductMetadata}';On its own, this simply acts as documentation. But if you also specify the
--jsonTypesFile flag, these annotations get incorporated into the schema:
pg-to-ts generate ... --jsonTypesFile './db-types' -o dbschema.tsThen your dbschema.ts will look like:
import {ProductMetadata} from './db-types';
interface Product {
id: string;
name: string;
description: string;
created_at: Date;
metadata: ProductMetadata | null;
}Presumably your db-types.ts file will either re-export this type from elsewhere:
export {ProductMetadata} from './path/to/this-type';or define it itself:
export interface ProductMetadata {
year?: number;
designer?: string;
starRating?: number;
}Note that, on its own, TypeScript cannot enforce a schema on your json
columns. For that, you'll want a tool like postgres-json-schema.
Prefix tableNames with there corresponding schemaName
--prefixWithSchemaNames
It will prefix all exports with the schema name. i.e schemaname_tablename. This allows you to easily namespace your exports.
If the schema name is: maxi, then the following exports will be generated for you when using the --prefixWithSchemaNames:
// Table product
export interface MaxiProduct {
id: string;
name: string;
description: string;
created_at: Date;
}
export interface MaxiProductInput {
id?: string;
name: string;
description: string;
created_at?: Date;
}
const maxi_product = {
tableName: 'maxi.product',
columns: ['id', 'name', 'description', 'created_at'],
requiredForInsert: ['name', 'description'],
}
export interface TableTypes {
maxi_product: {
select: MaxiProduct;
input: MaxiProductInput;
};
}
export const tables = {
maxi_product,
};Command Line Usage
There are a few ways to control pg-to-ts:
Command line flags
pg-to-ts generate -c postgresql://user:pass@host/db -o dbschema.tsJS / JSON file
pg-to-ts generate --config path/to/config.json
pg-to-ts generate --config # defaults to pg-to-ts.json
cat pg-to-ts.jsonThe JSON file has configuration options as top-level keys:
{
"conn": "postgres://user@localhost:5432/postgres",
"output": "/tmp/cli-pg-to-ts-json.ts"
}Environment variables
Flags may also be specified using environment variables prefixed with PG_TO_TS:
PG_TO_TS_CONN=postgres://user@localhost:5432/postgres
PG_TO_TS_OUTPUT=/tmp/cli-env.ts
pg-to-ts generateDevelopment Quickstart
You'll need a Postgres instance running to do most development work with pg-to-ts.
git clone https://github.com/danvk/pg-to-ts.git
cd pg-to-ts
yarn
yarn buildYou can iterate using your own DB schema. Or, to load the test schema, run:
psql postgres://user:pass@host/postgres -a -f test/fixture/pg-to-ts.sqlThen generate a dbschema.ts file by running:
node dist/cli.js generate -c postgresql://user:pass@host/db -o dbschema.tsYou can use yarn build --watch to run tsc in watch mode.
To run the unit tests:
yarn build
POSTGRES_URL=postgres://user@localhost:5432/postgres yarn testTo run ESLint:
yarn lintSee SweetIQ/schemats for the original README.