Coverage Status npm

A tiny wrapper around pg that makes PostgreSQL a lot of fun to use. Written in TypeScript.

import Pg from '@mojojs/pg';

// Use standard PostgreSQL connection URIs
const pg = new Pg('postgres://user:[email protected]:5432/database');

// Single query with safe placeholder
const results = await pg.query`SELECT ${'One'} AS one`;
for (const row of results) {

// Multiple queries on the same connection
const db = await pg.db();
const results = await db.query`SELECT 2`;
const results = await db.query`SELECT 3`;
await db.release();

Tagged template literals are used everywhere to protect from SQL injection attacks and to make syntax highlighting easy.


TypeScript is fully supported, just pass along a type with your query.

interface User {
  id: number;
  name: string;

const results = await pg.query<User>`SELECT * FROM users`;
for (const {id, name} of results) {
  console.log(`${id}: ${name}`);

SQL building

For easier SQL query building with partials, there are also pg.sql and db.sql tagged template literals. They can be used recursively to build complex queries securely.

// Build safe SQL query with placeholder and partial SQL query
const role = 'admin';
const partialQuery = pg.sql`AND role = ${role}`;
const name = 'root';
const results = await pg.query`SELECT * FROM users WHERE name = ${name} ${partialQuery}`;

But if you need a little more control over the generated SQL query, you can of course also bypass safety features with the tagged template literals pg.sqlUnsafe and db.sqlUnsafe.

const role = 'role = ' + pg.escapeLiteral('power user');
const partialQuery = pg.sqlUnsafe`AND ${role}`;
const name = 'root';
const results = await pg.query`SELECT * FROM users WHERE name = ${name} ${partialQuery}`;

But make sure to use methods like pg.escapeLiteral() to escape unsafe values yourself.


It’s best to use try/finally blocks whenever you dequeue a connection with pg.db(), to ensure efficient resource management.

try {
  const db = await pg.db();
  const tx = await db.begin();

  try {
    for (const user of ['Daniel', 'Isabell']) {
      await db.query`INSERT INTO users (name) VALUES (${user})`;
    await tx.commit();

  } finally {
    await tx.rollback();
} finally {
  await db.release();

The tx.rollback() call does nothing if tx.commit() has been called first.


To manage your database schema, there is also a minimal SQL based migration system built-in. A migration file is just a collection of sql blocks, with one or more statements, separated by comments of the form -- VERSION UP/DOWN.

-- 1 up
CREATE TABLE messages (message TEXT);
INSERT INTO messages VALUES ('I ♥ Mojolicious!');
-- 1 down
DROP TABLE messages;
-- 2 up ( can comment freely here...)
CREATE TABLE stuff (whatever INT);
-- 2 down

The idea is to let you migrate from any version, to any version, up and down. Migrations are very safe, because they are performed in transactions and only one can be performed at a time. If a single statement fails, the whole migration will fail and get rolled back. Every set of migrations has a name, which is stored together with the currently active version in an automatically created table named mojo_migrations.

import Path from '@mojojs/path';

// Load migrations from "migrations/myapp.sql" and migrate to the latest version
await pg.migrations.fromFile(Path.currentFile().sibling('migrations', 'myapp.sql'), {name: 'myapp'});
await pg.migrations.migrate();

// Use migrations to drop and recreate the schema
await pg.migrations.migrate(0);
await pg.migrations.migrate();

// Load migrations from a string
pg.migrations.fromString('-- 1 up\n...', {name: 'my_other_app'});

// Load migrations from a directory
await pg.migrations.fromDirecory(Path.currentFile().sibling('migrations'), {name: 'yet_another_app'});

To store your individual migration steps in separate SQL files you can use a directory structure like this. These files do not require special comments, because the version and migration direction are contained in the file names.

   |-- 1
   |   |-- up.sql
   |   `-- down.sql
   |-- 2
   |   `-- up.sql
   |-- 4
   |   |-- up.sql
   |   `-- down.sql
   `-- 5
       |-- up.sql
       `-- down.sql

Migrations are also compatible with Mojo::Pg, if you want to mix Perl and JavaScript code.


You can use events as well as async iterators for notifications.

// Send notifications
const db = await pg.db();
await db.notify('foo', 'just a message');

// Use an iterator to wait for incoming notifications
await db.listen('foo');
for await (const message of db) {
  console.log(`${}: ${message.payload}`);
await db.unlisten('foo');

// Use event to handle incoming notifications
await db.listen('bar');
db.on('notification', (message) => {
  console.log(`${}: ${message.payload}`);
await db.unlisten('bar');


This distribution also contains a great example you can use for inspiration. The well-structured blog application will show you how to apply the MVC design pattern in practice.


This package is designed to be compatible with the explicit resource management proposal and will support it as soon as the using keyword is available in Node.js.

// Multiple queries on the same connection (with automatic resource management)
using await const db = await pg.db();
const results = await db.query`SELECT 2`;
const results = await db.query`SELECT 3`;

Editor Support


All you need is Node.js 16.0.0 (or newer).

$ npm install @mojojs/pg


View Github