',children:[]},{type:"method",method:"inTable",example:"column.inTable(table)",description:'Sets the "table" where the foreign key column is located after calling column.references.',children:[]},{type:"method",method:"onDelete",example:"column.onDelete(command)",description:'Sets the SQL command to be run "onDelete".',children:[]},{type:"method",method:"onUpdate",example:"column.onUpdate(command)",description:'Sets the SQL command to be run "onUpdate".',children:[]},{type:"method",method:"defaultTo",example:"column.defaultTo(value)",description:"Sets the default value for the column on an insert.",children:[]},{type:"method",method:"unsigned",example:"column.unsigned()",description:"Specifies an integer as unsigned. No-op if this is chained off of a non-integer field.",children:[]},{type:"method",method:"notNullable",example:"column.notNullable()",description:"Adds a not null on the current column being created.",children:[]},{type:"method",method:"nullable",example:"column.nullable()",description:"Default on column creation, this explicitly sets a field to be nullable.",children:[]},{type:"method",method:"first",example:"column.first()",description:"Sets the column to be inserted on the first position, only used in MySQL alter tables.",children:[]},{type:"method",method:"after",example:"column.after(field)",description:"Sets the column to be inserted after another, only used in MySQL alter tables.",children:[]},{type:"method",method:"comment",example:"column.comment(value)",description:"Sets the comment for a column.",children:[]},{type:"code",content:"\n knex.schema.createTable('accounts', function(t) {\n t.increments().primary();\n t.string('email').unique().comment('This is the email field');\n });\n "},{type:"method",method:"collate",example:"column.collate(collation)",description:"Sets the collation for a column (only works in MySQL). Here is a list of all available collations: https://dev.mysql.com/doc/refman/5.5/en/charset-charsets.html",children:[]},{type:"code",content:"\n knex.schema.createTable('users', function(t) {\n t.increments();\n t.string('email').unique().collate('utf8_unicode_ci');\n });\n "}],e.exports=t.default},function(e,t){"use strict";t.__esModule=!0,t.default=[{type:"heading",
+size:"md",content:"Seed API",href:"Seeds-API"},{type:"text",content:["`knex.seed` is the class utilized by the knex seed CLI.","Each method takes an optional `config` object, which may specify the following properties:"]},{type:"list",content:["`directory`: a relative path to the directory containing the seed files (default `./seeds`)","`loadExtensions`: array of file extensions which knex will treat as seeds. For example, if you have typescript transpiled into javascript in the same folder, you want to execute only javascript seeds. In this case, set `loadExtensions` to `['.js']` (Notice the dot!) (default `['.co', '.coffee', '.eg', '.iced', '.js', '.litcoffee', '.ls', '.ts']`)"]},{type:"heading",size:"sm",content:"Methods",href:"Seeds-API-methods"},{type:"method",method:"make",example:"knex.seed.make(name, [config])",description:"Creates a new seed file, with the name of the seed file being added.",children:[]},{type:"method",method:"run",example:"knex.seed.run([config])",description:"Runs all seed files for the current environment.",children:[]}],e.exports=t.default},function(e,t){"use strict";t.__esModule=!0,t.default=[{type:"heading",size:"lg",content:"Support",href:"support"},{type:"text",content:"Have questions about the library? Come join us in the [#bookshelf freenode IRC](http://webchat.freenode.net/?channels=bookshelf) channel for support on knex.js and [bookshelf.js](http://bookshelfjs.org), or post an issue on [Stack Overflow](http://stackoverflow.com/questions/tagged/knex.js) or in the GitHub [issue tracker](https://github.com/tgriesser/knex/issues)."}],e.exports=t.default},function(e,t){"use strict";t.__esModule=!0,t.default=[{type:"heading",size:"lg",content:"Transactions",href:"Transactions"},{type:"text",content:["Transactions are an important feature of relational databases, as they allow correct recovery from failures and keep a database consistent even in cases of system failure. All queries within a transaction are executed on the same database connection, and run the entire set of queries as a single unit of work. Any failure will mean the database will rollback any queries executed on that connection to the pre-transaction state.",'Transactions are handled by passing a handler function into `knex.transaction`. The handler function accepts a single argument, an object which may be used in two ways:\n\n1. As the "promise aware" knex connection\n2. As an object passed into a query with [](#Builder-transacting)and eventually call commit or rollback.\n\nConsider these two examples:']},{type:"code",language:"js",content:"\n var Promise = require('bluebird');\n\n // Using trx as a query builder:\n knex.transaction(function(trx) {\n\n var books = [\n {title: 'Canterbury Tales'},\n {title: 'Moby Dick'},\n {title: 'Hamlet'}\n ];\n\n return trx\n .insert({name: 'Old Books'}, 'id')\n .into('catalogues')\n .then(function(ids) {\n return Promise.map(books, function(book) {\n book.catalogue_id = ids[0];\n\n // Some validation could take place here.\n\n return trx.insert(info).into('books');\n });\n });\n })\n .then(function(inserts) {\n console.log(inserts.length + ' new books saved.');\n })\n .catch(function(error) {\n // If we get here, that means that neither the 'Old Books' catalogues insert,\n // nor any of the books inserts will have taken place.\n console.error(error);\n });\n "},{type:"text",content:"And then this example:"},{type:"code",language:"js",content:"\n var Promise = require('bluebird');\n\n // Using trx as a transaction object:\n knex.transaction(function(trx) {\n\n var books = [\n {title: 'Canterbury Tales'},\n {title: 'Moby Dick'},\n {title: 'Hamlet'}\n ];\n\n knex.insert({name: 'Old Books'}, 'id')\n .into('catalogues')\n .transacting(trx)\n .then(function(ids) {\n return Promise.map(books, function(book) {\n book.catalogue_id = ids[0];\n\n // Some validation could take place here.\n\n return knex.insert(info).into('books').transacting(trx);\n });\n })\n .then(trx.commit)\n .catch(trx.rollback);\n })\n .then(function(inserts) {\n console.log(inserts.length + ' new books saved.');\n })\n .catch(function(error) {\n // If we get here, that means that neither the 'Old Books' catalogues insert,\n // nor any of the books inserts will have taken place.\n console.error(error);\n });\n "},{type:"text",content:["Throwing an error directly from the transaction handler function automatically rolls back the transaction, same as returning a rejected promise.","Notice that if a promise is not returned within the handler, it is up to you to ensure `trx.commit`, or `trx.rollback` are called, otherwise the transaction connection will hang."]}],e.exports=t.default},function(e,t){"use strict";t.__esModule=!0,t.default=[{type:"heading",size:"lg",content:"Upgrading 0.13 -> 0.14"},{type:"text",content:["[generic-pool](https://github.com/coopernurse/node-pool) was upgraded to v3. If you have specified `idleTimeoutMillis` or `softIdleTimeoutMillis` in the pool config then you will need to add `evictionRunIntervalMillis: 1000` when upgrading to 0.14.","See original [issue #2322](https://github.com/tgriesser/knex/issues/2322) for details."]}],e.exports=t.default},function(e,t){"use strict";t.__esModule=!0,t.default=[{type:"heading",size:"lg",content:"Utility",href:"Utility"},{type:"text",content:"A collection of utilities that the knex library provides for convenience."},{type:"heading",size:"md",content:"Batch Insert",href:"Utility-BatchInsert"},{type:"text",content:["The `batchInsert` utility will insert a batch of rows wrapped inside a transaction _(which is automatically created unless explicitly given a transaction using [transacting](#Builder-transacting))_, at a given `chunkSize`.","It's primarily designed to be used when you have thousands of rows to insert into a table.","By default, the `chunkSize` is set to 1000.","BatchInsert also allows for [returning values](#Builder-returning) and supplying transactions using [transacting](#Builder-transacting)."]},{type:"code",language:"js",content:"\n var rows = [{...}, {...}];\n var chunkSize = 30;\n knex.batchInsert('TableName', rows, chunkSize)\n .returning('id')\n .then(function(ids) { ... })\n .catch(function(error) { ... });\n\n knex.transaction(function(tr) {\n return knex.batchInsert('TableName', rows, chunkSize)\n .transacting(tr)\n })\n .then(function() { ... })\n .catch(function(error) { ... });\n "}],e.exports=t.default},function(e,t,n){"use strict";function r(e){return e&&e.__esModule?e:{default:e}}function i(e){return e.replace(/#(\d+)/g,function(e,t){return"["+e+"](https://github.com/tgriesser/knex/issues/"+t+")"})}function o(e){for(var t=s.default.lexer(e,{renderer:u}),n=[],r=void 0,o=void 0;t.length;){var a=t.shift();if("heading"===a.type){if(1===a.depth){var l=a.text.split("-").map(function(e){return e.trim()}),c=l[0],f=l[1];r=[],n.push({version:c,date:f,changes:r});continue}o=[],r.push({title:a.text,changes:o})}if("list_start"===a.type){for(var h=o||r;"list_end"!==a.type;)a=t.shift(),"text"===a.type&&h.push(i(a.text));o&&(o=null)}}return n}t.__esModule=!0,t.default=o;var a=n(340),s=r(a),u=new s.default.Renderer;e.exports=t.default},function(e,t,n){"use strict";function r(e){return e&&e.__esModule?e:{default:e}}function i(e,t){return t.map(function(t,n){var r=(0,a.default)({},t,{group:e,key:e+"-"+n});switch(r.children&&(r.children=i(e,r.children)),t.type){case"heading":return u.default.createElement(c.default,r);case"code":return u.default.createElement(h.default,r);case"method":return u.default.createElement(y.default,r);case"runnable":return u.default.createElement(b.default,r);case"list":return u.default.createElement(g.default,r);case"info":case"text":return u.default.createElement(d.default,r)}})}t.__esModule=!0;var o=n(245),a=r(o);t.default=i;var s=n(20),u=r(s),l=n(400),c=r(l),f=n(396),h=r(f),p=n(406),d=r(p),m=n(240),g=r(m),v=n(402),y=r(v),_=n(403),b=r(_);e.exports=t.default},function(e,t,n){e.exports={default:n(463),__esModule:!0}},function(e,t,n){e.exports={default:n(467),__esModule:!0}},function(e,t,n){e.exports={default:n(473),__esModule:!0}},function(e,t,n){e.exports={default:n(476),__esModule:!0}},function(e,t){"use strict";function n(e){var t=e.length;if(t%4>0)throw new Error("Invalid string. Length must be a multiple of 4");return"="===e[t-2]?2:"="===e[t-1]?1:0}function r(e){return 3*e.length/4-n(e)}function i(e){var t,r,i,o,a,s=e.length;o=n(e),a=new c(3*s/4-o),r=o>0?s-4:s;var u=0;for(t=0;t>16&255,a[u++]=i>>8&255,a[u++]=255&i;return 2===o?(i=l[e.charCodeAt(t)]<<2|l[e.charCodeAt(t+1)]>>4,a[u++]=255&i):1===o&&(i=l[e.charCodeAt(t)]<<10|l[e.charCodeAt(t+1)]<<4|l[e.charCodeAt(t+2)]>>2,a[u++]=i>>8&255,a[u++]=255&i),a}function o(e){return u[e>>18&63]+u[e>>12&63]+u[e>>6&63]+u[63&e]}function a(e,t,n){for(var r,i=[],a=t;ac?c:l+s));return 1===r?(t=e[n-1],i+=u[t>>2],i+=u[t<<4&63],i+="=="):2===r&&(t=(e[n-2]<<8)+e[n-1],i+=u[t>>10],i+=u[t>>4&63],i+=u[t<<2&63],i+="="),o.push(i),o.join("")}t.byteLength=r,t.toByteArray=i,t.fromByteArray=s;for(var u=[],l=[],c="undefined"!=typeof Uint8Array?Uint8Array:Array,f="ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/",h=0,p=f.length;h
Knex.js is a "batteries included" SQL query builder for Postgres, MSSQL, MySQL, MariaDB, SQLite3, and Oracle designed to be flexible, portable, and fun to use. It features both traditional node style callbacks as well as a promise interface for cleaner async flow control, a stream interface, full featured query and schema builders, transaction support (with savepoints), connection pooling and standardized responses between different query clients and dialects.
generic-pool was upgraded to v3. If you have specified idleTimeoutMillis or softIdleTimeoutMillis in the pool config then you will need to add evictionRunIntervalMillis: 1000 when upgrading to 0.14.
Knex can be used as an SQL query builder in both Node.JS and the browser, limited to WebSQL's constraints (like the inability to drop tables or read schemas). Composing SQL queries in the browser for execution on the server is highly discouraged, as this can be the cause of serious security vulnerabilities. The browser builds outside of WebSQL are primarily for learning purposes - for example, you can pop open the console and build queries on this page using the knex object.
Node.js
The primary target environment for Knex is Node.js, you will need to install the knex library, and then install the appropriate database library: pg for PostgreSQL, mysql for MySQL or MariaDB, sqlite3 for SQLite3, or mssql for MSSQL.
$ npm install knex --save
+
+# Then add one of the following (adding a --save) flag:
+$ npm install pg
+$ npm install sqlite3
+$ npm install mysql
+$ npm install mysql2
+$ npm install mariasql
+$ npm install strong-oracle
+$ npm install oracle
+$ npm install mssql
Browser
Knex can be built using a JavaScript build tool such as browserify or webpack. In fact, this documentation uses a webpack build which includes knex. View source on this page to see the browser build in-action (the global knex variable).
Initializing the Library
The knex module is itself a function which takes a configuration object for Knex, accepting a few parameters. The client parameter is required and determines which client adapter will be used with the library.
The connection options are passed directly to the appropriate database client to create the connection, and may be either an object, or a connection string:
Note: Knex's PostgreSQL client allows you to set the initial search path for each connection automatically using an additional option "searchPath" as shown below.
Initializing the library should normally only ever happen once in your application, as it creates a connection pool for the current database, you should use the instance returned from the initialize call throughout your library.
Specify the client for the particular flavour of SQL you are interested in.
var pg = require('knex')({client: 'pg'});
+knex('table').insert({a: 'b'}).returning('*').toString();
+// "insert into "table" ("a") values ('b')"
+
+pg('table').insert({a: 'b'}).returning('*').toString();
+// "insert into "table" ("a") values ('b') returning *"
Debugging
Passing a debug: true flag on your initialization object will turn on debugging for all queries.
Pooling
The client created by the configuration initializes a connection pool, using the generic-pool library. This connection pool has a default setting of a min: 2, max: 10 for the MySQL and PG libraries, and a single connection for sqlite3 (due to issues with utilizing multiple connections on a single file). To change the config settings for the pool, pass a pool option as one of the keys in the initialize block.
Checkout the generic-pool library for more information.
If you ever need to explicitly teardown the connection pool, you may use knex.destroy([callback]). You may use knex.destroy by passing a callback, or by chaining as a promise, just not both.
afterCreate
afterCreate callback (rawDriverConnection, done) is called when the pool aquires a new connection from the database server. done(err, connection) callback must be called for knex to be able to decide if the connection is ok or if it should be discarded right away from the pool.
var knex = require('knex')({
+ client: 'pg',
+ connection: {...},
+ pool: {
+ afterCreate: function (conn, done) {
+ // in this example we use pg driver's connection API
+ conn.query('SET timezone="UTC";', function (err) {
+ if (err) {
+ // first query failed, return error and don't try to make next query
+ done(err, conn);
+ } else {
+ // do the second query...
+ conn.query('SELECT set_limit(0.01);', function (err) {
+ // if err is not falsy, connection is discarded from pool
+ // if connection aquire was triggered by a query the error is passed to query promise
+ done(err, conn);
+ });
+ }
+ });
+ }
+ }
+});
acquireConnectionTimeout
acquireConnectionTimeout defaults to 60000ms and is used to determine how long knex should wait before throwing a timeout error when acquiring a connection is not possible. The most common cause for this is using up all the pool for transaction connections and then attempting to run queries outside of transactions while the pool is still full. The error thrown will provide information on the query the connection was for to simplify the job of locating the culprit.
Utilized by Oracledb. An array of types. The valid types are 'DATE', 'NUMBER' and 'CLOB'. When any column having one of the specified types is queried, the column data is returned as a string instead of the default representation.
For convenience, the any migration configuration may be specified when initializing the library. Read the Migrations section for more information and a full list of configuration options.
Hook for modifying returned rows, before passing them forward to user. One can do for example snake_case -> camelCase conversion for returned columns with this hook.
var knex = require('knex')({
+ client: 'mysql',
+ // overly simplified snake_case -> camelCase converter
+ postProcessResponse: (result) => {
+ // TODO: add special case for raw results (depends on dialect)
+ if (Array.isArray(result)) {
+ return result.map(row => convertToCamel(row));
+ } else {
+ return convertToCamel(result);
+ }
+ }
+});
wrapIdentifier
Knex supports transforming identifier names automatically to quoted versions for each dialect. For example 'Table.columnName as foo' for PostgreSQL is converted to "Table"."columnName" as "foo".
With wrapIdentifier one may override the way how identifiers are transformed. It can be used to override default functionality and for example to help doing camelCase -> snake_case conversion.
Conversion function wrapIdentifier(value, dialectImpl): string gets each part of the identifier as a single value and the second parameter is the original conversion function from the dialect implementation. For example knex('table').withSchema('foo').select('table.field as otherName').where('id', 1) will call wrapIdentifier converter for following values 'table', 'foo', 'table', 'field', 'otherName' and 'id'.
The heart of the library, the knex query builder is the interface used for building and executing standard SQL queries, such as select, insert, update, delete.
Identifier Syntax
In many places in APIs identifiers like table name or column name can be passed to methods.
Most commonly one needs just plain tableName.columnName, tableName or columnName, but in many cases one also needs to pass an alias how that identifier is referred later on in the query.
There are two ways to declare an alias for identifier. One can directly give as aliasName prefix for the identifier or oen can pass an object { aliasName: 'identifierName' }.
If in the object has multiple aliases { alias1: 'identifier1', alias2: 'identifier2' }, then all the aliased identifiers are expanded to comma separated list.
NOTE: identifier syntax has no place for selecting schema, so if you are doing schemaName.tableName, query might be rendered wrong. Use .withSchema('schemaName') instead.
The query builder starts off either by specifying a tableName you wish to query against, or by calling any method directly on the knex object. This kicks off a jQuery-like chain, with which you can call additional query builder methods as needed to construct the query, eventually calling any of the interface methods, to either convert toString, or execute the query with a promise, callback, or stream. Optional second argument for passing options:only: if true, the ONLY keyword is used before the tableName to discard inheriting tables' data. *NOTE: only supported in PostgreSQL for now.
timeout — .timeout(ms, options={cancel: boolean})
Sets a timeout for the query and will throw a TimeoutError if the timeout is exceeded. The error contains information about the query, bindings, and the timeout that was set. Useful for complex queries that you want to make sure are not taking too long to execute. Optional second argument for passing options:cancel: if true, cancel query if timeout is reached. *NOTE: only supported in MySQL and MariaDB for now.
knex.select().from('books').timeout(1000, {cancel: true}) // MySQL and MariaDB only Outputs: select * from`books`
select — .select([*columns])
Creates a select query, taking an optional array of columns for the query, eventually defaulting to * if none are specified when the query is built. The response of a select call will resolve with an array of objects selected from the database.
Allows for aliasing a subquery, taking the string you wish to name the current query. If the query is not a sub-query, it will be ignored.
knex.avg('sum_column1').from(function() {
+ this.sum('column1 as sum_column1').from('t1').groupBy('column1').as('t1')
+}).as('ignored_alias') Outputs: selectavg(`sum_column1`) from (selectsum(`column1`) as`sum_column1`from`t1`groupby`column1`) as`t1`
column — .column(columns)
Specifically set the columns to be selected on a select query, taking an array, an object or a list of column names. Passing an object will automatically alias the columns with the given keys.
from — .from([tableName], options={only: boolean})
Specifies the table used in the current query, replacing the current table name if one has already been specified. This is typically used in the sub-queries performed in the advanced where or union methods. Optional second argument for passing options:only: if true, the ONLY keyword is used before the tableName to discard inheriting tables' data. *NOTE: only supported in PostgreSQL for now.
Several methods exist to assist in dynamic where clauses. In many places functions may be used in place of values, constructing subqueries. In most places existing knex queries may be used to compose sub-queries, etc. Take a look at a few of the examples for each method for instruction on use:
Important: Supplying knex with an undefined value to any of the where functions will cause knex to throw an error during sql compilation. This is both for yours and our sake. Knex cannot know what to do with undefined values in a where clause, and generally it would be a programmatic error to supply one to begin with. The error will throw a message containing the type of query and the compiled query-string. Example:
knex('accounts')
+.where('login', undefined)
+.select()
+.toSQL() Error: Undefined binding(s) detected when compiling SELECT query: select * from `accounts` where `login` = ?
Several methods are provided which assist in building joins.
join — .join(table, first, [operator], second)
The join builder can be used to specify joins between tables, with the first argument being the joining table, the next three arguments being the first join column, the join operator and the second join column, respectively.
For grouped joins, specify a function as the second argument for the join query, and use on with orOn or andOn to create joins that are grouped with parentheses.
Creates a union query, taking an array or a list of callbacks to build the union statement, with optional boolean wrap. The queries will be individually wrapped in parentheses with a true wrap parameter.
Creates an insert query, taking either a hash of properties to be inserted into the row, or an array of inserts, to be executed as a single insert command. Resolves the promise / fulfills the callback with an array containing the first insert id of the inserted model, or an array containing all inserted ids for postgresql.
// Returns [1] in "mysql", "sqlite", "oracle"; [] in "postgresql" unless the 'returning' parameter is set.
+knex('books').insert({title: 'Slaughterhouse Five'}) Outputs: insertinto`books` (`title`) values ('Slaughterhouse Five')
Utilized by PostgreSQL, MSSQL, and Oracle databases, the returning method specifies which column should be returned by the insert and update methods. Passed column parameter may be a string or an array of strings. When passed in a string, makes the SQL result be reported as an array of values from the specified column. When passed in an array of strings, makes the SQL result be reported as an array of objects, each containing a single property for each of the specified columns.
Creates an update query, taking a hash of properties or a key/value pair to be updated based on the other query constraints. Resolves the promise / fulfills the callback with the number of affected rows for the query. If a key to be updated has value undefined it is ignored.
// Returns [1] in "mysql", "sqlite", "oracle"; [] in "postgresql" unless the 'returning' parameter is set.
+knex('books').update('title', 'Slaughterhouse Five') Outputs: update`books`set`title` = 'Slaughterhouse Five'
del / delete — .del()
Aliased to del as delete is a reserved word in JavaScript, this method deletes one or more rows, based on other conditions specified in the query. Resolves the promise / fulfills the callback with the number of affected rows for the query.
Used by knex.transaction, the transacting method may be chained to any query and passed the object you wish to join the query as part of the transaction for.
Dynamically added after a transaction is specified, the forShare adds a FOR SHARE in PostgreSQL and a LOCK IN SHARE MODE for MySQL during a select statement.
Performs a count on the specified column. Also accepts raw expressions. Note that in Postgres, count returns a bigint type which will be a String and not a Number (more info).
Clones the current query chain, useful for re-using partial query snippets in other queries without mutating the original.
modify — .modify(fn, *arguments)
Allows encapsulating and re-using query snippets and common behaviors as functions. The callback function should receive the query builder as its first argument, followed by the rest of the (optional) parameters passed to modify.
Returns an object with the column info about the current table, or an individual column if one is passed, returning an object with the following keys:defaultValue: the default value for the columntype: the column typemaxLength: the max length set for the columnnullable: whether the column may be null
Overrides the global debug setting for the current query chain. If enabled is omitted, query debugging will be turned on.
connection
(incomplete) - This feature was incorrectly documented as functional. If implemented, the method would set the db connection to use for the query without using the connection pool.
options — .options()
Allows for mixing in additional options as defined by database client specific libraries:
Transactions are an important feature of relational databases, as they allow correct recovery from failures and keep a database consistent even in cases of system failure. All queries within a transaction are executed on the same database connection, and run the entire set of queries as a single unit of work. Any failure will mean the database will rollback any queries executed on that connection to the pre-transaction state.
Transactions are handled by passing a handler function into knex.transaction. The handler function accepts a single argument, an object which may be used in two ways:
+
As the "promise aware" knex connection
+
As an object passed into a query with and eventually call commit or rollback.
+
+Consider these two examples:
varPromise = require('bluebird');
+
+// Using trx as a query builder:
+knex.transaction(function(trx) {
+
+ var books = [
+ {title: 'Canterbury Tales'},
+ {title: 'Moby Dick'},
+ {title: 'Hamlet'}
+ ];
+
+ return trx
+ .insert({name: 'Old Books'}, 'id')
+ .into('catalogues')
+ .then(function(ids) {
+ returnPromise.map(books, function(book) {
+ book.catalogue_id = ids[0];
+
+ // Some validation could take place here.
+
+ return trx.insert(info).into('books');
+ });
+ });
+})
+.then(function(inserts) {
+ console.log(inserts.length + ' new books saved.');
+})
+.catch(function(error) {
+ // If we get here, that means that neither the 'Old Books' catalogues insert,
+ // nor any of the books inserts will have taken place.
+ console.error(error);
+});
And then this example:
varPromise = require('bluebird');
+
+// Using trx as a transaction object:
+knex.transaction(function(trx) {
+
+ var books = [
+ {title: 'Canterbury Tales'},
+ {title: 'Moby Dick'},
+ {title: 'Hamlet'}
+ ];
+
+ knex.insert({name: 'Old Books'}, 'id')
+ .into('catalogues')
+ .transacting(trx)
+ .then(function(ids) {
+ returnPromise.map(books, function(book) {
+ book.catalogue_id = ids[0];
+
+ // Some validation could take place here.
+
+ return knex.insert(info).into('books').transacting(trx);
+ });
+ })
+ .then(trx.commit)
+ .catch(trx.rollback);
+})
+.then(function(inserts) {
+ console.log(inserts.length + ' new books saved.');
+})
+.catch(function(error) {
+ // If we get here, that means that neither the 'Old Books' catalogues insert,
+ // nor any of the books inserts will have taken place.
+ console.error(error);
+});
Throwing an error directly from the transaction handler function automatically rolls back the transaction, same as returning a rejected promise.
Notice that if a promise is not returned within the handler, it is up to you to ensure trx.commit, or trx.rollback are called, otherwise the transaction connection will hang.
Schema Builder
The knex.schema is a getter function, which returns a stateful object containing the query. Therefore be sure to obtain a new instance of the knex.schema for every query. These methods return promises.
withSchema — knex.schema.withSchema([schemaName])
Specifies the schema to be used when using the schema-building commands.
Creates a new table on the database if it doesn't exists on database, with a callback function to modify the table's structure, using the schema-building commands.
Run an arbitrary sql query in the schema builder chain.
knex.schema.raw("SET sql_mode='TRADITIONAL'")
+.table('users', function (table) {
+ table.dropColumn('name');
+ table.string('first_name');
+ table.string('last_name');
+}) Outputs: SET sql_mode='TRADITIONAL';
+altertable`users`add`first_name`varchar(255), add`last_name`varchar(255);
+altertable`users`drop`name`
Schema Building:
dropColumn — table.dropColumn(name)
Drops a column, specified by the column's name
dropColumns — table.dropColumns(*columns)
Drops multiple columns, taking a variable number of column names.
renameColumn — table.renameColumn(from, to)
Renames a column from one name to another.
increments — table.increments(name)
Adds an auto incrementing column, in PostgreSQL this is a serial. This will be used as the primary key for the table. Also available is a bigIncrements if you wish to add a bigint incrementing number (in PostgreSQL bigserial).
integer — table.integer(name)
Adds an integer column.
bigInteger — table.bigInteger(name)
In MySQL or PostgreSQL, adds a bigint column, otherwise adds a normal integer. Note that bigint data is returned as a string in queries because JavaScript may be unable to parse them without loss of precision.
text — table.text(name, [textType])
Adds a text column, with optional textType for MySql text datatype preference. textType may be mediumtext or longtext, otherwise defaults to text.
string — table.string(name, [length])
Adds a string column, with optional length defaulting to 255.
float — table.float(column, [precision], [scale])
Adds a float column, with optional precision (defaults to 8) and scale (defaults to 2).
Adds a created_at and updated_at column on the database, setting these each to dateTime types. When true is passed as the first argument a timestamp type is used. Both colums default to being not null and the current timestamp when true is passed as the second argument.
dropTimestamps — table.dropTimestamps()
Drops the columns created_at and updated_at from the table, which can be created via timestamps.
binary — table.binary(name, [length])
Adds a binary column, with optional length argument for MySQL.
enum / enu — table.enu(col, values)
Adds a enum column, (aliased to enu, as enum is a reserved word in JavaScript). Note that the second argument is an array of values. Example:
table.enu('column', ['value1', 'value2'])
json — table.json(name)
Adds a json column, using the built-in json type in postgresql, defaulting to a text column in older versions of postgresql or in unsupported databases. Note that when setting an array (or a value that could be an array) as the value of a json or jsonb column, you should use JSON.stringify() to convert your value to a string prior to passing it to the query builder, e.g.
This is because postgresql has a native array type which uses a syntax incompatible with json; knex has no way of knowing which syntax to use, and calling JSON.stringify() forces json-style syntax.
jsonb — table.jsonb(name)
Adds a jsonb column. Works similar to table.json(), but uses native jsonb type if possible.
uuid — table.uuid(name)
Adds a uuid column - this uses the built-in uuid type in postgresql, and falling back to a char(36) in other databases.
comment — table.comment(value)
Sets the comment for a table.
engine — table.engine(val)
Sets the engine for the database table, only available within a createTable call, and only applicable to MySQL.
charset — table.charset(val)
Sets the charset for the database table, only available within a createTable call, and only applicable to MySQL.
collate — table.collate(val)
Sets the collation for the database table, only available within a createTable call, and only applicable to MySQL.
inherits — table.inherits(val)
Sets the tables that this table inherits, only available within a createTable call, and only applicable to PostgreSQL.
specificType — table.specificType(name, type)
Sets a specific type for the column creation, if you'd like to add a column type that isn't supported here.
index — table.index(columns, [indexName], [indexType])
Adds an index to a table over the given columns. A default index name using the columns is used unless indexName is specified. The indexType can be optionally specified for PostgreSQL.
dropIndex — table.dropIndex(columns, [indexName])
Drops an index from a table. A default index name using the columns is used unless indexName is specified (in which case columns is ignored).
unique — table.unique(columns, [indexName])
Adds an unique index to a table over the given columns. A default index name using the columns is used unless indexName is specified.
Adds a foreign key constraint to a table for an existing column using table.foreign(column).references(column) or multiple columns using table.foreign(columns).references(columns). A default key name using the columns is used unless foreignKeyName is specified. You can also chain onDelete and/or onUpdate to set the reference option (RESTRICT, CASCADE, SET NULL, NO ACTION) for the operation. Note, this is the same as column.references(column) but works for existing columns.
knex.schema.table('users', function (table) {
+ table.integer('user_id').unsigned()
+ table.foreign('user_id').references('Items.user_id_in_items')
+})
Drops a foreign key constraint from a table. A default foreign key name using the columns is used unless foreignKeyName is specified (in which case columns is ignored).
Drops a unique key constraint from a table. A default unique key name using the columns is used unless indexName is specified (in which case columns is ignored).
dropPrimary — table.dropPrimary([constraintName])
Drops the primary key constraint on a table. Defaults to tablename_pkey unless constraintName is specified.
Chainable Methods:
The following three methods may be chained on the schema building methods, as modifiers to the column.
alter — column.alter()
Marks the column as an alter / modify, instead of the default add. Note: This only works in .alterTable() and is not supported by SQlite. Alter is not done incrementally over older column type so if you like to add notNull and keep the old default value, the alter statement must contain both .notNull().defaultTo(1).alter(). If one just tries to add .notNull().alter() the old default value will be dropped.
knex.schema.alterTable('user', function(t) {
+ t.increments().primary(); // add
+ // drops previous default value from column, change type to string and add not nullable constraint
+ t.string('username', 35).notNullable().alter();
+ // drops both not null contraint and the default value
+ t.integer('age').alter();
+});
index — column.index([indexName], [indexType])
Specifies a field as an index. If an indexName is specified, it is used in place of the standard index naming convention of tableName_columnName. The indexType can be optionally specified for PostgreSQL. No-op if this is chained off of a field that cannot be indexed.
primary — column.primary([constraintName])
When called on a single column it will set that column as the primary key for a table.
+ To create a compound primary key, pass an array of column names:
+ `table.primary(['column1', 'column2'])`.
+ Constraint name defaults to `tablename_pkey` unless `constraintName` is specified.
+
unique — column.unique()
Sets the column as unique.
references — column.references(column)
Sets the "column" that the current column references as a foreign key. "column" can either be "." syntax, or just the column name followed up with a call to inTable to specify the table.
inTable — column.inTable(table)
Sets the "table" where the foreign key column is located after calling column.references.
onDelete — column.onDelete(command)
Sets the SQL command to be run "onDelete".
onUpdate — column.onUpdate(command)
Sets the SQL command to be run "onUpdate".
defaultTo — column.defaultTo(value)
Sets the default value for the column on an insert.
unsigned — column.unsigned()
Specifies an integer as unsigned. No-op if this is chained off of a non-integer field.
notNullable — column.notNullable()
Adds a not null on the current column being created.
nullable — column.nullable()
Default on column creation, this explicitly sets a field to be nullable.
first — column.first()
Sets the column to be inserted on the first position, only used in MySQL alter tables.
after — column.after(field)
Sets the column to be inserted after another, only used in MySQL alter tables.
comment — column.comment(value)
Sets the comment for a column.
knex.schema.createTable('accounts', function(t) {
+ t.increments().primary();
+ t.string('email').unique().comment('This is the email field');
+});
Sometimes you may need to use a raw expression in a query. Raw query object may be injected pretty much anywhere you want, and using proper bindings can ensure your values are escaped properly, preventing SQL-injection attacks.
Raw Parameter Binding:
One can paramterize sql given to knex.raw(sql, bindings). Parameters can be positional named. One can also choose if parameter should be treated as value or as sql identifier e.g. in case of 'TableName.ColumnName' reference.
knex('users')
+.select(knex.raw('count(*) as user_count, status'))
+.where(knex.raw(1))
+.orWhere(knex.raw('status <> ?', [1]))
+.groupBy('status') Outputs: selectcount(*) as user_count, statusfrom`users`where1orstatus <> 1groupby`status`
Positional bindings ? are interpreted as values and ?? are interpreted as identifiers.
Named bindings such as :name are interpreted as values and :name: interpreted as identifiers. Named bindings are processed so long as the value is anything other than undefined.
knex('users')
+.where(knex.raw(':name: = :thisGuy or :name: = :otherGuy or :name: = :undefinedBinding', {
+ name: 'users.name',
+ thisGuy: 'Bob',
+ otherGuy: 'Jay',
+ undefinedBinding: undefined
+})) Error: Undefined binding(s) detected when compiling RAW query: `users`.`name` = ? or `users`.`name` = ? or `users`.`name` = :undefinedBinding
For simpler queries where one only has a single binding, .raw can accept said binding as its second parameter.
knex('users')
+.where(
+ knex.raw('LOWER("login") = ?', 'knex')
+)
+.orWhere(
+ knex.raw('accesslevel = ?', 1)
+)
+.orWhere(
+ knex.raw('updtime = ?', newDate.UTC('01-01-2016'))
+) Error: Date.UTC is not a constructor
Note that due to ambiguity, arrays must be passed as arguments within a containing array.
knex.raw('select * from users where id in (?)', [1, 2, 3]);
+// Error: Expected 3 bindings, saw 1
+
+knex.raw('select * from users where id in (?)', [[1, 2, 3]]) Outputs: select * fromuserswhereidin (1, 2, 3)
To prevent replacement of ? one can use the escape sequence \\?.
Raw expressions are created by using knex.raw(sql, [bindings]) and passing this as a value for any value in the query chain.
knex('users')
+.select(knex.raw('count(*) as user_count, status'))
+.where(knex.raw(1))
+.orWhere(knex.raw('status <> ?', [1]))
+.groupBy('status') Outputs: selectcount(*) as user_count, statusfrom`users`where1orstatus <> 1groupby`status`
Raw Queries:
The knex.raw may also be used to build a full query and execute it, as a standard query builder query would be executed. The benefit of this is that it uses the connection pool and provides a standard interface for the different client libraries.
knex.raw('select * from users where id = ?', [1]).then(function(resp) { ... });
Note that the response will be whatever the underlying sql library would typically return on a normal query, so you may need to look at the documentation for the base library the queries are executing against to determine how to handle the response.
Wrapped Queries:
The raw query builder also comes with a wrap method, which allows wrapping the query in a value:
var subcolumn = knex.raw('select avg(salary) from employee where dept_no = e.dept_no')
+.wrap('(', ') avg_sal_dept');
+
+knex.select('e.lastname', 'e.salary', subcolumn)
+.from('employee as e')
+.whereRaw('dept_no = e.dept_no') Outputs: select`e`.`lastname`, `e`.`salary`, (selectavg(salary) from employee where dept_no = e.dept_no) avg_sal_dept from`employee`as`e`where dept_no = e.dept_no
Note that the example above be achieved more easily using the as method.
A collection of utilities that the knex library provides for convenience.
Batch Insert
The batchInsert utility will insert a batch of rows wrapped inside a transaction (which is automatically created unless explicitly given a transaction using transacting), at a given chunkSize.
It's primarily designed to be used when you have thousands of rows to insert into a table.
Knex.js provides several options to deal with query output. The following methods are present on the query builder, schema builder, and the raw builder:
Promises
Promises are the preferred way of dealing with queries in knex, as they allow you to return values from a fulfillment handler, which in turn become the value of the promise. The main benefit of promises are the ability to catch thrown errors without crashing the node app, making your code behave like a .try / .catch / .finally in synchronous code.
Coerces the current query builder chain into a promise state, accepting the resolve and reject handlers as specified by the Promises/A+ spec. As stated in the spec, more than one call to the then method for the current query chain will resolve with the same value, in the order they were called; the query will not be executed multiple times.
Executes side effects on the resolved response, ultimately returning a promise that fulfills with the original value. A thrown error or rejected promise will cause the promise to transition into a rejected state.
// Using only .then()
+query.then(function(x) {
+ doSideEffectsHere(x);
+ return x;
+});
+
+// Using .tap()
+promise.tap(doSideEffectsHere);
map — .map(mapper)
A passthrough to Bluebird's map implementation with the result set.
If you'd prefer a callback interface over promises, the asCallback function accepts a standard node style callback for executing the query chain. Note that as with the then method, subsequent calls to the same query chain will return the same result.
Streams are a powerful way of piping data through as it comes in, rather than all at once. You can read more about streams here at substack's stream handbook. See the following for example uses of stream & pipe. If you wish to use streams with PostgreSQL, you must also install the pg-query-stream module. On an HTTP server, make sure to manually close your streams if a request is aborted.
stream — .stream([options], [callback])
If called with a callback, the callback is passed the stream and a promise is returned. Otherwise, the readable stream is returned.
Pipe a stream for the current query to a writableStream.
var stream = knex.select('*').from('users').pipe(writableStream);
Events
query
A query event is fired just before a query takes place, providing data about the query, including the connection's __cid property and any other information about the query as described in toSQL. Useful for logging all queries throughout your application.
A query-error event is fired when an error occurs when running a query, providing the error object and data about the query, including the connection's __cid property and any other information about the query as described in toSQL. Useful for logging all query errors throughout your application.
knex.select(['NonExistentColumn'])
+.from('users')
+.on('query-error', function(error, obj) {
+ app.log(error);
+})
+.then(function() { // ... })
+.catch(function(error) {
+ // Same error object as the query-error event provides.
+});
query-response
A query-response event is fired when a successful query has been run, providing the response of the query and data about the query, including the connection's __cid property and any other information about the query as described in toSQL, and finally the query builder used for the query.
knex.select('*')
+.from('users')
+.on('query-response', function(response, obj, builder) {
+ // ...
+})
+.then(function(response) {
+ // Same response as the emitted event
+})
+.catch(function(error) { });
Other
toString — .toString()
Returns an array of query strings filled out with the correct values based on bindings, etc. Useful for debugging, but should not be used to create queries for running them against DB.
var toStringQuery = knex.select('*').from('users').where('id', 1).toString();
+
+// Outputs: console.log(toStringQuery);
+// select * from "users" where "id" = 1
toSQL — .toSQL() and toSQL().toNative()
Returns an array of query strings filled out with the correct values based on bindings, etc. Useful for debugging and building queries for running them manually with DB driver. .toSQL().toNative() outputs object with sql string and bindings in a dialects format in the same way that knex internally sends them to unterlying DB driver.
knex.select('*').from('users')
+ .where(knex.raw('id = ?', [1]))
+ .toSQL()
+// Outputs:
+// {
+// bindings: [1],
+// method: 'select',
+// sql: 'select * from "users" where id = ?',
+// options: undefined,
+// toNative: function () {}
+// }
+
+knex.select('*').from('users')
+ .where(knex.raw('id = ?', [1]))
+ .toSQL().toNative()
+// Outputs for postgresql dialect:
+// {
+// bindings: [1],
+// sql: 'select * from "users" where id = $1',
+// }
Migrations
Migrations allow for you to define sets of schema changes so upgrading a database is a breeze.
Migration CLI
The migration CLI is bundled with the knex install, and is driven by the node-liftoff module. To install globally, run:
$ npm install knex -g
Migrations use a knexfile, which specify various configuration settings for the module. To create a new knexfile, run the following:
$ knex init
+
+# or for .coffee
+
+$ knex init -x coffee
will create a sample knexfile.js - the file which contains our various database configurations. Once you have a knexfile.js, you can use the migration tool to create migration files to the specified directory (default migrations). Creating new migration files can be achieved by running:
$ knex migrate:make migration_name
Once you have finished writing the migrations, you can update the database matching your NODE_ENV by running:
$ knex migrate:latest
You can also pass the --env flag or set NODE_ENV to select an alternative environment:
$ knex migrate:latest --env production
+
+# or
+
+$ NODE_ENV=production knex migrate:latest
To rollback the last batch of migrations:
$ knex migrate:rollback
Seed files
Seed files allow you to populate your database with test or seed data independent of your migration files.
Seed CLI
To create a seed file, run:
$ knex seed:make seed_name
Seed files are created in the directory specified in your knexfile.js for the current environment. A sample seed configuration looks like:
If no seeds.directory is defined, files are created in ./seeds. Note that the seed directory needs to be a relative path. Absolute paths are not supported (nor is it good practice).
To run seed files, execute:
$ knex seed:run
Seed files are executed in alphabetical order. Unlike migrations, every seed file will be executed when you run the command. You should design your seed files to reset tables as needed before inserting data.
knexfile.js
A knexfile.js or knexfile.coffee generally contains all of the configuration for your database. It can optionally provide different configuration for different environments. You may pass a --knexfile option to any of the command line statements to specify an alternate path to your knexfile.
knex.migrate is the class utilized by the knex migrations cli.
Each method takes an optional config object, which may specify the following properties:
directory: a relative path to the directory containing the migration files (default ./migrations)
extension: the file extension used for the generated migration files (default js)
tableName: the table name used for storing the migration state (default knex_migrations)
disableTransactions: don't run migrations inside transactions (default false)
loadExtensions: array of file extensions which knex will treat as migrations. For example, if you have typescript transpiled into javascript in the same folder, you want to execute only javascript migrations. In this case, set loadExtensions to ['.js'] (Notice the dot!) (default ['.co', '.coffee', '.eg', '.iced', '.js', '.litcoffee', '.ls', '.ts'])
Transactions in migrations
By default, each migration is run inside a transaction. Whenever needed, one can disable transactions for all migrations via the common migration config option config.disableTransactions or per-migration, via exposing a boolean property config.transaction from a migration file:
Retrieves and returns the current migration version, as a promise. If there aren't any migrations run yet, returns "none" as the value for the currentVersion.
Notes about locks
A lock system is there to prevent multiple processes from running the same migration batch in the same time. When a batch of migrations is about to be run, the migration system first tries to get a lock using a SELECT ... FOR UPDATE statement (preventing race conditions from happening). If it can get a lock, the migration batch will run. If it can't, it will wait until the lock is released.
Please note that if your process unfortunately crashes, the lock will have to be manually removed in order to let migrations run again. The locks are saved in a table called "tableName_lock"; it has single one column called is_locked that you need to set to 0 in order to release the lock.
Seed API
knex.seed is the class utilized by the knex seed CLI.
Each method takes an optional config object, which may specify the following properties:
directory: a relative path to the directory containing the seed files (default ./seeds)
loadExtensions: array of file extensions which knex will treat as seeds. For example, if you have typescript transpiled into javascript in the same folder, you want to execute only javascript seeds. In this case, set loadExtensions to ['.js'] (Notice the dot!) (default ['.co', '.coffee', '.eg', '.iced', '.js', '.litcoffee', '.ls', '.ts'])
Methods
make — knex.seed.make(name, [config])
Creates a new seed file, with the name of the seed file being added.
How do I help contribute? Glad you ask! Pull requests, or feature requests, though not always implemented, are a great way to help make Knex even better than it is now. If you're looking for something specific to help out with, there's a number of unit tests that aren't implemented yet, the library could never have too many of those. If you want to submit a fix or feature, take a look at the Contributing readme in the Github and go ahead and open a ticket.
How do I debug? Knex is beginning to make use of the debug module internally, so you can set the DEBUG environment variable to knex:* to see all debugging, or select individual namespaces DEBUG=knex:query,knex:tx to constrain a bit.
If you pass {debug: true} as one of the options in your initialize settings, you can see all of the query calls being made. Sometimes you need to dive a bit further into the various calls and see what all is going on behind the scenes. I'd recommend node-inspector, which allows you to debug code with debugger statements like you would in the browser.
At the start of your application code will catch any errors not otherwise caught in the normal promise chain handlers, which is very helpful in debugging.
How do I run the test suite? The test suite looks for an environment variable called KNEX_TEST for the path to the database configuration. If you run the following command:
$ export KNEX_TEST='/path/to/your/knex_config.js'
+$ npm test
replacing with the path to your config file, and the config file is valid, the test suite should run properly.
My tests are failing because slow DB connection and short test timeouts! How to extend test timeouts? Sometimes, e.g. when running CI on travis, test suite's default timeout of 5 seconds might be too short. In such cases an alternative test timeout value in milliseconds can be specified using the KNEX_TEST_TIMEOUT environment variable.
$ export KNEX_TEST_TIMEOUT=30000
+$ npm test
Can I use Knex outside of Node.js Yes. While the WebSQL spec is deprecated, there is still an adapter that provides support. You will need to use a build tool like browserify or webpack for a browser build.
Dropped indexes feature now functions correctly, (#278)
0.5.14 — May 6, 2014
Remove the charset encoding if it's utf8 for mysql, as it's the default but also currently causes some issues in recent versions of node-mysql
0.5.13 — April 2, 2014
Fix regression in array bindings for postgresql (#228)
0.5.12 — Mar 31, 2014
Add more operators for where clauses, including && (#226)
0.5.11 — Mar 25, 2014
.where(col, 'is', null) or .where(col, 'is not', null) are not supported (#221).
Case insensitive where operators now allowed (#212).
Fix bug in increment/decrement truncating to an integer (#210).
Disconnected connections are now properly handled & removed from the pool (#206).
Internal tweaks to binding concatenations for performance (#207)
0.5.10 — Mar 19, 2014
Add the .exec method to the internal promise shim
0.5.9 — Mar 18, 2014
Remove error'ed connections from the connection pool (#206), added support for node-postgres-pure (pg.js) (#200)
0.5.8 — Feb 27, 2014
Fix for chaining on forUpdate / forShare, adding map & reduce from bluebird
0.5.7 — Feb 18, 2014
Fix for a null limit / offset breaking query chain (#182)
0.5.6 — Feb 5, 2014
Bump bluebird dependency to ~1.0.0, fixing regression in Bluebird 1.0.2 (#176)
0.5.5 — Jan 28, 2014
Fix for the exit code on the migrations cli (#151).
The init method in knex.migrate now uses this.config if one isn't passed in (#156)
0.5.4 — Jan 7, 2014
Fix for using raw statements in defaultTo schema builder methods (#146)
0.5.3 — Jan 2, 2014
Fix for incorrectly formed sql when aggregates are used with columns (#144)
0.5.2 — Dec 18, 2013
Adding passthrough "catch", "finally" to bluebird implementations, use bluebird's "nodeify" internally for exec
0.5.1 — Dec 12, 2013
The returning in PostgreSQL may now accept * or an array of columns to return. If either of these are passed, the response will be an array of objects rather than an array of values. Updates may also now use a returning value. (#132)
Added bigint and bigserial type to PostgreSQL. (#111)
Several fixes for migrations, including migration file path fixes, passing a Promise constructor to the migration up and down methods, allowing the "knex" module to be used globally, file ordering on migrations, and other small improvements. (#112-115, #125, #135)
Fix for documentation of hasColumn, ensure that hasColumn works with MySQL (#87).
More cleanup of error messages, showing the original error message concatenated with the sql and bindings
0.4.8 — Oct 2, 2013
Connections are no longer pushed back into the pool if they never existed to begin with (#85)
0.4.7 — Sep 27, 2013
The column is now a documented method on the builder api, and takes either an individual column or an array of columns to select
0.4.6 — Sep 25, 2013
Standardizing handling of errors for easier debugging, as noted in (#39)
0.4.5 — Sep 24, 2013
Fix for hasTable always returning true in MySQL (#82), fix where sql queries were duplicated with multiple calls on toSql with the schema builder
0.4.4 — Sep 22, 2013
Fix for debug method not properly debugging individual queries
0.4.3 — Sep 18, 2013
Fix for underscore not being defined in various grammar files
0.4.2 — Sep 17, 2013
Fix for an error being thrown when an initialized ClientBase instance was passed into Knex.initialize. pool.destroy now optionally accepts a callback to notify when it has completed draining and destroying all connections
0.4.1 — Sep 16, 2013
Cleanup from the 0.4.0 release, fix a potential exploit in "where" clauses pointed out by Andri Möll, fix for clients not being properly released from the pool #70, fix for where("foo", "<>", null) doing an "IS NULL" statement
0.4.0 — Sep 13, 2013 Breaking Changes:
Global state is no longer stored in the library, an instance is returned from Knex.initialize, so you will need to call this once and then reference this knex client elsewhere in your application.
Lowercasing of knex.raw, knex.transaction, and knex.schema.
Created columns are now nullable by default, unless notNullable is chained as an option.
Keys created with increments are now assumed to be unsigned (MySQL) by default.
The destroyAllNow is no longer called by the library on process.exit event. If you need to call it explicitly yourself, you may use knex.client.destroyPool
0.2.6 — Aug 29, 2013
Reject the transaction promise if the transaction "commit" fails, (#50)
0.2.5 — Aug 25, 2013
Fix error if a callback isn't specified for exec, (#49)
0.2.4 — Aug 22, 2013
Fix SQLite3 delete not returning affected row count, (#45)
0.2.3 — Aug 22, 2013
Fix insert with default values in PostgreSQL and SQLite3, (#44)
0.2.2 — Aug 20, 2013
Allowing Raw queries to be passed as the primary table names
0.2.1 — Aug 13, 2013
Fix for an array passed to insert being mutated
0.2.0 — Aug 7, 2013 Breaking changes:
hasTable now returns a boolean rather than a failed promise.
Changed syntax for insert in postgresql, where the id is not assumed on inserts (#18). The second parameter of insert is now required to return an array of insert id's for the last insert.
The timestamp method on the schema builder now uses a dateTime rather than a timestamp
0.1.8 — July 7, 2013
Somehow missing the != operator. Using .find rather than .where in getCommandsByName(#22)
0.1.7 — June 12, 2013
Ensures unhandled errors in the exec callback interface are re-thrown
0.1.6 — June 9, 2013
Renaming beforeCreate to afterCreate. Better handling of errors in the connection pooling
0.1.5 — June 9, 2013
Added the ability to specify beforeCreate and beforeDestroy hooks on the initialize's options.pool to perform any necessary database setup/teardown on connections before use (#14). where and having may now accept Knex.Raw instances, for consistency (#15). Added an orHaving method to the builder. The ability to specify bindings on Raw queries has been removed
0.1.4 — May 22, 2013
defaultTo now accepts "false" for boolean columns, allows for empty strings as default values
0.1.3 — May 18, 2013
Enabling table aliases (#11). Fix for issues with transactions not functioning (#12)
0.1.2 — May 15, 2013
Bug fixes for groupBy (#7). Mysql using collation, charset config settings in createTable. Added engine on schemaBuilder specifier (#6). Other doc fixes, tests
0.1.1 — May 14, 2013
Bug fixes for sub-queries, minor changes to initializing "main" instance, adding "pg" as a valid parameter for the client name in the connection settings