Advanced Usage

# Advanced Usage

# Data Driven Tests

Execute the same scenario on a different data set.

Let's say you want to test login for different user accounts. In this case, you need to create a datatable and fill it in with credentials. Then use Data().Scenario to include this data and generate multiple scenarios:

// Define data table inside a test or load from another module
let accounts = new DataTable(['login', 'password']); //
accounts.add(['davert', '123456']); // adding records to a table
accounts.add(['admin', '123456']);

// You can skip some data. But add them to report as skipped (just like with usual scenarios):
accounts.xadd(['admin', '23456'])

// Pass dataTable to Data()
// Use special param `current` to get current data set
Data(accounts).Scenario('Test Login', ({ I, current }) => {
  I.fillField('Username', current.login); // current is reserved!
  I.fillField('Password', current.password);'Sign In');
  I.see('Welcome '+ current.login);

// Also you can set only for Data tests. It will launch executes only the current test but with all data options
Data(accounts).only.Scenario('Test Login', ({ I, current }) => {
  I.fillField('Username', current.login); // current is reserved!
  I.fillField('Password', current.password);'Sign In');
  I.see('Welcome '+ current.login);

Important: you can't use name current for pageObjects or helpers in data scenarios

This will produce 2 tests with different data sets. Current data set is appended to a test name in output:

βœ“ Test Login | {"login":"davert","password":"123456"}
βœ“ Test Login | {"login":"admin","password":"123456"}
S Test Login | {"login":"admin","password":"23456"}
// You can filter your data table
Data(accounts.filter(account => account.login == 'admin')
.Scenario('Test Login', ({ I, current }) => {
  I.fillField('Username', current.login);
  I.fillField('Password', current.password);'Sign In');
  I.see('Welcome '+ current.login);

This will limit data sets accoring passed function:

βœ“ Test Login | {"login":"admin","password":"123456"}
S Test Login | {"login":"admin","password":"23456"}

Data sets can also be defined with array, generator, or a function.

Data(function*() {
  yield { user: 'davert'};
  yield { user: 'andrey'};
}).Scenario() // ...

HINT: If you don't use DataTable. add toString() method to each object added to data set, so the data could be pretty printed in a test name

# Tags

Append @tag to your test name, so

Scenario('update user profile @slow')

Alternativly, use tag method of Scenario to set additional tags:

Scenario('update user profile', ({  }) => {
  // test goes here

All tests with @tag could be executed with --grep '@tag' option.

codeceptjs run --grep '@slow'

Use regex for more flexible filtering:

  • --grep '(?=.*@smoke2)(?=.*@smoke3)' - run tests with @smoke2 and @smoke3 in name
  • --grep "\@smoke2|\@smoke3" - run tests with @smoke2 or @smoke3 in name
  • --grep '((?=.*@smoke2)(?=.*@smoke3))|@smoke4' - run tests with (@smoke2 and @smoke3) or @smoke4 in name
  • --grep '(?=.*@smoke2)^(?!.*@smoke3)' - run tests with @smoke2 but without @smoke3 in name
  • --grep '(?=.*)^(?!.*@smoke4)' - run all tests except @smoke4

# Debug

CodeceptJS provides a debug mode in which additional information is printed. It can be turned on with --debug flag.

npx codeceptjs run --debug

to receive even more information turn on --verbose flag:

npx codeceptjs run --verbose

You can pause execution and enter interactive console mode by calling pause() inside your test.

To see a complete internal debug of CodeceptJS use DEBUG env variable:

DEBUG=codeceptjs:* npx codeceptjs run

For an interactive debugging use NodeJS debugger. In WebStorm:

node $NODE_DEBUG_OPTION ./node_modules/.bin/codeceptjs run

For Visual Studio Code, add the following configuration in launch.json:

  "type": "node",
  "request": "launch",
  "name": "codeceptjs",
  "args": ["run", "--grep", "@your_test_tag"],
  "program": "${workspaceFolder}/node_modules/codeceptjs/bin/codecept.js"

# Test Options

Features and Scenarios have their options that can be set by passing a hash after their names:

Feature('My feature', {key: val});

Scenario('My scenario', {key: val},({ I }) => {});

You can use this options for build your own plugins (opens new window) with event listners (opens new window). Example:

  // for test
  event.dispatcher.on(event.test.before, (test) => {
    if (test.opts.key) {
  // or for suite
  event.dispatcher.on(event.suite.before, (suite) => {
    if (suite.opts.key) {

# Timeout

Tests can get stuck due to various reasons such as network connection issues, crashed browser, etc. This can make tests process hang. To prevent these situations timeouts can be used. Timeouts can be set explicitly for flaky parts of code, or implicitly in a config.

Previous timeout implementation was disabled as it had no effect when dealing with steps and promises.

# Steps Timeout

It is possible to limit a step execution to specified time with I.limitTime command. It will set timeout in seconds for the next executed step:

// limit clicking to 5 seconds

It is possible to set a timeout for all steps implicitly (except waiters) using stepTimeout plugin.

# Tests Timeout

Test timeout can be set in seconds via Scenario options:

// limit test to 20 seconds
Scenario('slow test that should be stopped', { timeout: 20 }, ({ I }) => {
  // ...

This timeout can be set globally in codecept.conf.js in seconds:

exports.config = {

  // each test must not run longer than 5 mins
  timeout: 300,


# Suites Timeout

A timeout for a group of tests can be set on Feature level via options.

// limit all tests in this suite to 30 seconds
Feature('flaky tests', { timeout: 30 })

# Timeout Confguration

Updated in 3.4

Timeout rules can be set globally via config.

To set a timeout for all running tests provide a number of seconds to timeout config option:

// inside codecept.conf.js or codecept.conf.ts
timeout: 30, // limit all tests in all suites to 30 secs

It is possible to tune this configuration for a different groups of tests passing options as array and using grep option to filter tests:

// inside codecept.conf.js or codecept.conf.ts

timeout: [
  10, // default timeout is 10secs  

  // but increase timeout for slow tests
    grep: '@slow',
    Feature: 50

ℹ️ grep value can be string or regexp

It is possible to set a timeout for Scenario or Feature:

// inside codecept.conf.js or codecept.conf.ts
timeout: [

  // timeout for Feature with @slow in title
    grep: '@slow',
    Feature: 50
  // timeout for Scenario with 'flaky0' .. `flaky1` in title
    // regexp can be passed to grep
    grep: /flaky[0-9]/,
    Scenario: 10

  // timeout for all suites
    Feature: 20

Global timeouts will be overridden by explicit timeouts of a test or steps.

# Disable Timeouts

To execute tests ignoring all timeout settings use --no-timeouts option:

npx codeceptjs run --no-timeouts

# Dynamic Configuration

Helpers can be reconfigured per scenario or per feature. This might be useful when some tests should be executed with different settings than others. In order to reconfigure tests use .config() method of Scenario or Feature.

Scenario('should be executed in firefox', ({ I }) => {
  // I.amOnPage(..)
}).config({ browser: 'firefox' })

In this case config overrides current config of the first helper. To change config of specific helper pass two arguments: helper name and config values:

Scenario('should create data via v2 version of API', ({ I }) => {
  // I.amOnPage(..)
}).config('REST', { endpoint: '' })

Config can also be set by a function, in this case you can get a test object and specify config values based on it. This is very useful when running tests against cloud providers, like BrowserStack. This function can also be asynchronous.

Scenario('should report to BrowserStack', ({ I }) => {
  // I.amOnPage(..)
}).config((test) => {
  return { desiredCapabilities: {
    project: test.suite.title,
    name: test.title,

Config changes can be applied to all tests in suite:

Feature('Admin Panel').config({ url: '' });

Please note that some config changes can't be applied on the fly. For instance, if you set restart: false in your config and then changing value browser won't take an effect as browser is already started and won't be closed untill all tests finish.

Configuration changes will be reverted after a test or a suite.

Last Updated: 11/7/2023, 6:18:35 PM