Need help with seriate?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

LeanKit-Labs
123 Stars 50 Forks MIT License 168 Commits 40 Opened issues

Description

A cross platform node module for Microsoft SQL Server based on node-mssql

Services available

!
?

Need anything else?

Contributors list

Seriate

=========

A cross platform node module for Microsoft SQL Server based on node-mssql

Connections and Pooling

Each connection configuration provided creates a connection pool for the parameters specified by the configuration. When connecting to a single database, use

setDefault
to specify a global/ambient connection pool and all future API calls made without a connection configuration or name argument will use that connection pool. If connections to multiple databases are required, provide an name in the configuration when calling
addConnection
and use the name in place of that configuration when making an API call targeting that connection.

Connection Configuration

A connection configuration provides information on how to connect to the datbase server and how to handle connection pooling. Seriate will not create multiple connection pools for the same database; i.e. if different pooling parameters are provided but all other parameters are the same as a previous configuration, Seriate just uses the previously created pool without any changes made.

Note: the name property is optional and only useful when managing multiple connection pools to different databases.

SQL authentication example (default instance)

{
    "name": "default",
    "host": "127.0.0.1",
    "user": "nodejs",
    "password": "mypassword",
    "database": "master",
    "pool": {
        "max": 10,
        "min": 4,
        "idleTimeoutMillis": 30000
    }
}

SQL authentication example (named instance)

{
    "name": "default",
    "host": "127.0.0.1",
    "port": "12345", // find the port of your instance in the SQL Server Configuration Manager
    "user": "nodejs",
    "password": "mypassword",
    "database": "master",
    "pool": {
        "max": 10,
        "min": 4,
        "idleTimeoutMillis": 30000
    }
}

The port for your named instance can be found in SQL Server Configuration Manager under:

SQL Server Network Configuration ->
 Protocols for [instance name] ->
 TCP/IP (properties) ->
 IP Addresses tab ->
 IPAll group ->
 TCP Dynamic Ports value
.

Alternatively, you can specify the server as

server: yourServer\\instanceName
, but there are some caveats:
  • The SQL Server Browser service must be running
  • You cannot use
    .
    as the server name. It must be a name that can be resolved via DNS (ex.
    localhost
    )
  • There may be local firewall configuration settings that need adjustment

Trusted/NTLM authentication example

{
    "name": "default",
    "host": "127.0.0.1",
    "user": "windowsUser",
    "password": "windowsUserPassword",
    "database": "master",
    "domain": "yourDomain", // should be machine name if the server is not in an AD domain
    "pool": {
        "max": 10,
        "min": 4,
        "idleTimeoutMillis": 30000
    }
}

Connectivity Events

Seriate emits connectivity events from the top level library:

  • connected
  • closed
  • failed

Each event includes a

name
property that represents which connection the event is happening on. If the event is failed, an
error
property will have the error that caused the failure.

API

Sql type constants are exposed in both Pascal Case and all capitals off of the library. See the listing at the end of this document.

Summary

The follow methods are exposed on the module:

|method name | description | |--------------------|--------------------------| |

addConnection( config )
| Adds a named connection pool for use in other API calls. |
setDefault( config )
| Sets an ambient/global connection pool for all commands that don't specify a connection. |
getPlainContext( [connection] )
| returns transaction-less database context. |
getTransactionContext( [connection] )
| returns a context associating one transaction with a connection. |
executeTransaction( [connection,] queryOptions )
| shortcut method to execute one command on a transaction context |
execute( [connection,] queryOptions )
| shortcut method to execute one command on a plain (transaction-less) context. |
first( [connection,] queryOptions )
| shortcut method that returns only the first row of a result set (calls
execute
under the hood) |
fromFile( path )
| Allows you to read a
.sql
file instead of in-lining your SQL in your JavaScript. |
useMetrics( metronic [, namespace ]  )
| Pass metronic instance to
seriate
and optionally set namespace.

NOTE: The

connection
argument can be omitted to use a default config, an existing configuration name or a new configuration.

addConnection( config )

Creates a named connection pool for future use. If the name is omitted and no default has been specified, it has the same effect as calling

setDefault
.

setDefault( config )

Creates a connection pool with the name "default" for future use. If other API calls are made with no

connection
argument, the connection pool created by this call will be used.

getPlainContext( [connection] )

This method returns a

SqlContext
instance, and allows you to add 1 or more steps to the context, with each step representing a query/command that should be executed in the database. Steps are given an
alias
(which is used to identify the result set returned), and the query details can be provided as an object (which we"ll see below), or a callback that takes an
execute
continuation (which is used to process your query options argument). Let's take a look at the first approach:

Specifying a step using just the query options object argument

Here's an example of using a plain context to read a table:

// set a default connection pool
sql.setDefault( {
    user: "username",
    password: "pwd",
    host: "127.0.0.1",
    database: "master"
} );

sql.getPlainContext() .step( "readUsers", { query: "select * From sys.sysusers" // optionally you could do this if the // above query were in a readUsers.sql file // query: sql.fromFile( "readUsers" ); } ) .end( function( sets ){ // sets has a "readUsers" property // which contains the results of the query } ) .error( function( err ){ console.log( err ); } );

Note that the

SqlContext
instance returns from
getPlainContext
has a
step
method with a signature of
(alias, queryOptions)
. The possible values for a
queryOptions
argument are:
{
    query: "plain sql query here",
    procedure: "stored procedure name to execute",
    preparedSql: "prepared SQL statement",
    params: {
        param1Name: {
            type: sql.NVARCHAR,
            val: "paramValue"
        },
        param2Name: {
            type: sql.Int,
            val: 123
        },
        param3Name: "param3Value"
    },
    multiple: false,
    bulkLoadTable: {
        name: "TableToBulkLoad",
        columns: {
            id: {
                type: sql.INT,
                nullable: false
            },
            name: {
                type: sql.NVARCHAR(100),
                nullable: true
            }
        },
        rows: [
            {
                id: 246,
                name: "Teddy Pendergrass"
            },
            ...
        ]
    }
}

You can only use one of the four sql-related fields:

query
,
procedure
,
preparedSql
, or
bulkLoadTable
. The module infers from which one you use as to how it should execute (and it's checked in that order). If your query takes params, you can provide a
params
object, where each key represents a parameter name and the value can be an object that provides the
val
and
type
(types are pulled from the
mssql
module), or the value can be a primitive value which will be passed as the parameter value. If multiple recordsets are expected from a
query
or
preparedSql
, set the
multiple
field to
true
. Multiple recordsets are automatically supported when executing a stored procedure. (NOTE: if you use multiple recordsets, your result set for a step will be an array of records sets (i.e. - nested arrays), rather than an array of a single record set).

If you need to bulk load a table, you can do that with the

bulkLoadTable
option. Just specify the
name
of the table to load, definitions of
columns
keyed by name, and an array of
rows
as shown above. Each column must have a
type
and may optionally have a
nullable
boolean, which must match the nullability of the corresponding database column. If you omit
nullable
for a column, it is assumed to be
true
. If the table does not exist, it will be created. If you need to bulk load a temporary table, just specify a name with a
#
at the beginning of its name. In this case you must use a transaction context instead of a plain context. The temporary table will automatically be dropped at the end of the transaction. The result for a
bulkLoadTable
step is the number of rows inserted.

When bulk loading a temporary table, by default seriate will assume it should be a new table, and it will first drop any temporary table found with that name. If you want to override this behavior and add rows to an existing temporary table, just add a

useExisting
of
true
to your
bulkLoadTable
object.

The

end
method of a
SqlContext
instance takes a callback which receives a
sets
argument. The
sets
argument contains the dataset(s) from each step (using the step
alias
as the property name). The
error
method allows you to pass a callback that will receive an error notfication if anything fails. Note that calling
end
or
error
is what starts the unit of work handled by the context.

Specifying a step using a callback that takes an
execute
continuation

Here's an example of using a plain context to read a table, and then use data from that read to determine details about the next step:

// add a named connection pool
sql.addConnection( {
    name: "example-1",
    user: "username",
    password: "pwd",
    host: "127.0.0.1",
    database: "master"
} );

sql.getPlainContext( "example-1" ) .step( "readUsers", { query: "select * From sys.sysusers" } ) .step( "usersTransforms", function( execute, data ) { // data will contain a readUsers property with // the prior step's results in an array. You can // use this approach with a callback to dynamically // determine what's fed to this step's executable // action. Let's pretend we fished out a particular // user from the readUsers step and then did this: const userId = getUserIdFrom( data.readUsers ); execute( { procedure: "GetExtendedUserInfo", params: { userid: userId } } ) } ) .end( function( sets ) { // sets has a "readUsers" property // which contains the results of the query } ) .error( function( err ) { console.log( err ); } );

The above example shows both

step
approaches side-by-side.

Inserting multiple rows from an object array with the
asTable
property

Using the

asTable
property you can define a parameter as an array of objects, and its name will refer to a table variable with a column for each declared attribute. The
asTable
property should be set to an object with a key for each property and its desired SQL type as its value.
.step( "insertChildren", {
    query: "INSERT INTO Children (Id, ParentId, FirstName, MiddleName) " +
                "SELECT id, @parentId, firstName, middleName FROM @children",
    params: {
        parentId: {
            val: 123,
            type: sql.INT
        },
        children: {
            val: [
                { id: 1, firstName: "James", middleName: "Paul"},
                { id: 2, firstName: "John", middleName: "Winston" },
                { id: 3, firstName: "George", middleName: "Harold" },
                { id: 4, firstName: "Richard", middleName: "Parkin" }
            ],
            asTable: {
                id: sql.INT,
                firstName: sql.NVARCHAR(50),
                middleName: sql.NVARCHAR(50)
            }
        }
    }
} )

Inserting multiple rows from a value array with
asTable: true

Providing the

asTable
property the Boolean value
true
, you can define a parameter as an array of scalar values, and its name will refer to a table variable with a single column named
value
. The column will be of the parameter's specified type.
.step( "insertChildren", {
    query: "INSERT INTO ParentChild (ParentId, ChildId) " +
                "SELECT @parentId, value FROM @childIds",
    params: {
        parentId: {
            val: 123,
            type: sql.INT
        },
        childIds: {
            val: [ 10, 15, 24, 48 ],
            type: sql.INT,
            asTable: true
        }
    }
} )

Passing a value list with
asList: true

Providing the

asList
property the Boolean value
true
, you can define a parameter list for use in an
IN
clause. The parameter name you provide will be transformed into a list of parameters, one for each provided value. These parameters will be of the parameter's specified type.
.step( "selectChildren", {
    query: "SELECT * FROM ParentChild WHERE ParentId IN @parentIds"
    params: {
        parentIds: {
            val: [ 123, 456, 789 ],
            type: sql.INT,
            asList: true
        }
    }
} )

Getting results in streams

If you are expecting a particularly large result set, you might want to receive your results in a readable stream instead of an array of objects. To get a step's result returned as a stream in object mode, put

stream: true
on its
queryOptions
argument. The stream will emit two types of objects,
recordset
objects and
row
objects.

At the beginning of each result set will come a single

recordset
object, recognizable by its single
recordset
property whose value describes the columns of the coming rows. This object is keyed by column names with object values describing each column.
{
    recordset: {
        column1: {
            index: 0,
            name: 'column1',
            length: undefined,
            type: [sql.Int],
            scale: undefined,
            precision: undefined,
            nullable: false,
            caseSensitive: false,
            identity: false,
            readOnly: true
        },
        ...
    }
}

Following each

recordset
object will be zero or more
row
objects, each recognizable by its single
row
property whose value is simply an object of values keyed by column name.
{
    row: {
        column1: 123,
        ...
    }
}

Streamed results can be very useful for piping into transforms, writing directly to the response stream, writing to a file, etc.

getTransactionContext( [connection], [dataForHooks] )

The

getTransactionContext
method returns a
TransactionContext
instance - which for the most part is nearly identical to a
SqlContext
instance - however, a transaction is started as the context begins its work, and you have the option to commit or rollback in the
end
method's callback. For example:
// note: name defaults to "default"
// if no prior default connection was
// specified.
sql.addConnection( {
    user: "username",
    password: "pwd",
    host: "127.0.0.1",
    database: "master",
    // optional code to run as the first step in a transaction
    atTransactionStart( dataForHooks ) {
        return {
            procedure: "UpdateTracking",
            params: {
                userId: {
                    val: dataforHooks.userId,
                    type: sql.INT
                }
            }
        };
    },
    // optional code to run as the last step in a transaction
    atTransactionEnd( dataForHooks ) {
        return {
            procedure: "ClearTracking"
        };
    }
})

sql.getTransactionContext( null, { userId: "1234" } ) .step( "readUsers", { query: "select * From sys.sysusers" } ) .step( "usersTransforms", function( execute, data ) { // data will contain a readUsers property with // the prior step's results in an array. You can // use this approach with a callback to dynamically // determine what's fed to this step's executable // action. Let's pretend we fished out a particular // user from the readUsers step and then did this: const userId = getUserIdFrom( data.readUsers ); execute( { procedure: "GetExtendedUserInfo", params: { userid: userId } } ) } ) .end( function( result ) { // the result arg contains a sets property // with all the dataset results from the steps // in this context, but it also contains a transaction // member that contains a commit and rollback method. // Calling either commit or rollback returns a promise. result.transaction .commit() .then(function() { console.log( "Yay, we"re not afraid of commitment..." ); }, function( err ) { console.log( "O NOES! An error" ); } ); } ) .error( function( err ){ console.log( err ); } );

You can see that the main difference between a

SqlContext
and
TransactionContext
is that the argument passed to the
end
callback contains more than just the
sets
(data sets) in the
TransactionContext
. A
TransactionContext
does not automatically call
commit
for you - that's in your hands (for now). However, if an error occurs, it will call
rollback
.

executeTransaction( [connection,] queryOptions, [dataForHooks] )

This is a shortcut method to getting a

TransactionContext
instance to execute one step. It returns a promise, and the
result
argument that's normally fed to the
end
method's callback is passed to the success handler of the promise, and any errors are passed to the error handler. For example:
// re-use of this variable to API calls
// will result in the same underlying pool
// being used.
const connection = {
    user: "username",
    password: "pwd",
    host: "127.0.0.1",
    database: "master",
    // optional code to run as the first step in a transaction
    atTransactionStart( dataForHooks ) {
        return {
            procedure: "UpdateTracking",
            params: {
                userId: {
                    val: dataforHooks.userId,
                    type: sql.INT
                }
            }
        };
    },
    // optional code to run as the last step in a transaction
    atTransactionEnd( dataForHooks ) {
        return {
            procedure: "ClearTracking"
        };
    }
};

sql.executeTransaction( connection, { procedure: "UpdateCustomer", params: { customerid: { val: id, type: sql.INT }, balance: { val: 45334, type: sql.MONEY } } } ).then( function( data ) { // you can choose to commit or rollback here // data.result is your result set // data also contains a transaction prop // with commit/rollback methods return data.transaction .commit() .then( function() { console.log( "Updated customer balance...." ); } ); }, function( err ) { console.log( err ); } );

execute( [connection,] queryOptions )

This is a shortcut method to getting a

SqlContext
instance to execute one step. It returns a promise, and the query result is passed to the success handler of the promise, and any errors are passed to the error handler. For example:
// re-use of this variable to API calls
// will result in the same underlying pool
// being used.
const connection = {
    user: "username",
    password: "pwd",
    host: "127.0.0.1",
    database: "master"
};

sql.execute( connection, { name: "selectFromSomeTableById", preparedSql: "select * from someTable where id = @id", params: { id: { val: 123, type: sql.INT } } } ).then( function( data ) { //data is the query result set }, function( err ) { console.log( err ); } );

useMetrics( metrics [, namespace ] )

Seriate will record metrics around each database call but only if a metronic instance is provided via this call. The optional namespace argument allows you to control the namespace used for all metrics captured.

Note: when calling

execute
,
executeTransaction
or
first
, provide a step name via a
name
property in the options.

So How Does This Work?

Read this far, eh? Great - now we can talk about how it works under the hood.

Contexts as Acceptor-style FSMs

The

SqlContext
constructor function is derived from
machina.Fsm
(the constructor function used to create a finite state machine in the
machina
library).
SqlContext
starts with only the following states:
uninitialized
,
connecting
,
done
and
error
. The FSM loosely acts as an acceptor-style FSM - with each step resulting in either a success or error result, determining if it continues on to the eventual
done
state, or lands in
error
.

Adding Steps -> Adding States to the FSM

When you use the

step
call to add a step to the context, the
alias
you provide becomes a new state name which is added to the underlying FSM, and the action you specify via the
queryOptions
argument (or the alternative syntax using the callback that takes the
execute
continuation), becomes the entry action for this new state. Success and error input handlers are also added to this state, and are triggered based on the results of the call to the database that occurs in the entry action. The new state names you add are queued up in a
pipeline
array, so that the order in which you added them can be maintained once the FSM begins executing.

Starting Execution

Once you"ve added the steps you want executed as part of the unit of work, you have to use the

end
or
error
calls to start execution. Under the hood, calling either
end
or
error
results in the FSM being told to handle a
start
input. Let's pretend you have a
SqlContext
, and you"ve added two steps to it:
readCustomer
and
updateContact
. The FSM starts in the uninitialized state, and once it receives the
start
input, proceeds to the
connecting
state where it opens a connection to the database. Assuming a successful connection, it then proceeds to the
readUsers
state added by our first step. Again, assuming success, the next state would be
updateContact
, and finally the
done
state. As the FSM enters the
done
state, it emits the result to the callback passed to
end
. If any state encounters an error along the way, it transitions to the
error
state. Upon entering the
error
state, the FSM emits the error to the callback passed to
error
. Here's a directed graph showing the scenario I just described (the states that are always part of the FSM are in blue, and the ones added by the :

TransactionContext Flow

The

TransactionContext
is mostly identical to the SqlContext, except that it has an extra state after
connecting
- the
startingTransaction
state. Here's the same scenario we described above in a
TransactionContext
:

Getting Set Up/Testing/etc.

  • You"ll need to run
    npm install
    at the root of this project once you clone it to install the dependencies.
  • To run unit tests:
    npm test
  • To run integration tests:
    npm run test:int
  • To run the example module:
    npm run example

Please note that in order to run the integration tests, you will need to create a

local-config.json
file in the
spec/integration
directory, matching something similar to this:
{
    "host": "10.0.1.16",
    "user": "username",
    "password": "pwd",
    "database": "master",
    "pool": {
        "max": 10,
        "min": 2,
        "idleTimeoutMillis": 30000
    }
}

Sql Constants

Pascal and Upper case properties:

javascript
VarChar                 VARCHAR
NVarChar                NVARCHAR
Text                    TEXT
Int                     INT
BigInt                  BIGINT
TinyInt                 TINYINT
SmallInt                SMALLINT
Bit                     BIT
Float                   FLOAT
Numeric                 NUMERIC
Decimal                 DECIMAL
Real                    REAL
Date                    DATE
DateTime                DATETIME
DateTime2               DATETIME2
DateTimeOffset          DATETIMEOFFSET
SmallDateTime           SMALLDATETIME
Time                    TIME
UniqueIdentifier        UNIQUEIDENTIFIER
SmallMoney              SMALLMONEY
Money                   MONEY
Binary                  BINARY
VarBinary               VARBINARY
Image                   IMAGE
Xml                     XML
Char                    CHAR
NChar                   NCHAR
NText                   NTEXT
                        TVP
                        UDT
Geography               GEOGRAPHY
Geometry                GEOMETRY

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.