Skip to content

monch1962/wilee

Repository files navigation

wilee

App designed to execute REST API functional test cases that have been encoded as JSON documents

Build Status

Build Status Build Status Build Status Build Status Build Status Build Status Build Status Build Status Build Status Build Status Build Status Build Status

Quick start

Assuming you've got Go installed:

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ go build wilee.go
$ APP="https://jsonplaceholder.typicode.com" bin/wilee < demo/test-cases/jsonplaceholder-test.json

To cross-compile for Mac

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ make clean
$ make mac

standalone wilee executable will be in bin/ directory

To cross-compile for Windows

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ make clean
$ make windows

standalone wilee executable will be in bin/ directory

To cross-compile for Linux

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ make clean
$ make linux

standalone wilee executable will be in bin/ directory

To install as a Lambda function

Assuming you have the serverless framework installed (https://www.serverless.com) and your AWS credentials setup:

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ make clean
$ make lambda
$ serverless deploy

wilee will be deployed as a Lambda function in your AWS account

Getting help

Let's see what a very simple wilee test case looks like...

Getting started with Test-Driven Development (TDD)

Let's assume we're at the planning stage for a sprint, and the decision is made to implement a new REST API to return details about Australian Prime Ministers. Let's see what we can do during that planning session to start creating executable test cases...

We can very quickly create a new wilee test template for the functionality we're about to start building. Let's create a very simple test to check whether the initial functionality is working as expected Note that we're not checking anything about the output of this test case, except that it returns a HTTP 200 response. All this test does is a GET /prime-ministers/1 and confirm a 200 response code.

Big deal, right? Except that now we've got a small piece of scope that everyone can agree on, and that we can now check into our CI pipeline to run whether new code gets checked in. Took about 5 seconds to create the test, and now we can start to flesh it out a bit.

Now I want to tag this new test as a progression test case, meaning it covers new functionality being built in this sprint. That's pretty simple

Creating negative tests from working tests

One of the strengths of having test cases written in JSON is that they're highly mutable. This means that we can run test cases through tools to mutate them into other test cases. There's any number of ways to mutate JSON into something else; the only limit is your imagination.

Some simple utilities are provided to turn passing tests into tests that should return an error code. For example

$ cat demo/test-cases/jsonplaceholder-test.json | test-mutators/mutate-naughty-body.sh

should mutate a "good" test case into one with a malformed body, where we want to see it return a HTTP 422 result

Why use wilee?

wilee is been created out of frustration at currently available integration test tools. In my opinion, nearly all of these tools do a very good job of solving the wrong set of problems. In 2018, my set of non-negotiable requirements are:

  • good testers are hard to find and tend to be expensive; you really want them to be as productive as possible, as quickly as possible
  • most of the value of good testers comes in their ability to identify what needs to be tested, work out how to test it & explain to others why exactly that set of things needs to be tested. Their ability to actually create automation test cases is somewhere further down the list
  • installing an automation test toolset should be as fast, simple and idiot-proof as possible, on any platform from a tester's laptop to a CI/CD pipeline
  • people talk about test-driven development (TDD) as though it's easily achievable; it's a great idea, but for TDD to work, it has to be possible to create new automation tests within seconds or minutes, not hours or days
    • adding new test cases should be a no-brainer, near-zero-time task that can be performed before development starts
    • with practice, a scrum team should be able to create new automation tests during sprint planning (WTF?)
  • test cases should be highly mutable by design; test cases are almost never static, and need to evolve over the lifetime of a project
  • test cases should be atomic (e.g. single file per test case, with no external dependencies)
  • existing tools tend to focus on test case creation, but the real problem is what comes after that: test lifecycle management
  • test execution should be extremely fast; I don't want to wait while some big clunky framework gets compiled and/or spun up, I want my results instantly
  • test execution should normally be performed as part of a CI/CD pipeline, automatically triggered via e.g. a git branch update. Generally speaking, people should only be running automation test cases when they're shaking out problems with the tests themselves
  • test execution should be highly auditable and support any audit framework
  • creating and maintaining test suites should be a trivial exercise
  • avoid test frameworks as far as possible, as they tie you into a specific way of working
  • testing tools should support current and future best practice processes. There's no way of knowing what "future best practice processes" will be, and IT people are notoriously bad when it comes to selecting which tools to use
  • I work with code in many different languages; over the past 12 months, the list includes Java, Ruby, Python, Clojure, Golang and probably a few others. I don't want a different testing toolkit for each of those languages

What's wrong with other SIT tools in this space?

This is always going to be a highly opinionated list, but here goes...

  • high level of expertise required to create automation test cases
  • even with that high level of expertise, creating new test cases tends to be sloooow
  • some come with a large set of dependencies
  • focused primarily on test creation, rather than lifecycle management, productivity, agility, maintaining flexibility, auditability, future proofing, CI/CD friendliness, ...
  • tend to lock you into to specific processes, toolsets, frameworks, programming languages, etc.
  • can be a challenge to interface with test management tools such as JIRA

What does wilee bring to the table?

wilee sets out to be the smallest possible viable framework for API integration testing. The entire wilee "framework" is one file - the wilee executable you built above

wilee seeks to utilise the fantastic command-line tools that have been developed since Unix first appeared in the 1960s. Tools like 'bash', 'cat', and 'jq' in particular; these tools are free, they have just about every conceivable bug shaken out, and your developers probably already know how to use them.

wilee makes it easy to leverage the brilliance that is Docker containers, along with whatever container orchestration framework you happen to be using (Docker Compose, Kubernetes, etc.). Docker makes loads of stuff easier, including SIT. wilee won't force you to use Docker, but Docker makes life a lot easier.

wilee tries to make it simple to integrate with whatever test management tools you might be using - that's JIRA, ALM and the like - using bash and wilee itself. If it has an API, then wilee can talk to it

Design decisions

wilee is deliberately opinionated, which has driven the following design decisions:

  • test cases are written in JSON, and test execution results are presented in JSON
  • wilee is written in Go
  • command line interface only
  • minimal feature sets

Now let's look at the implications of these decisions...

Why JSON?

JSON is:

  • language and toolset agnostic - "everything understands JSON"
  • easy for humans to write and reason about - "have I get test coverage for this?"
  • JSON actually makes a great DSL for writing test cases & capturing test results
  • tests become atomic (1 JSON per test case, 1 JSON per test case execution)
  • easy to mutate existing test cases to create new tests (e.g. generative testing, fuzzy testing)
  • easy to evolve existing test cases to deal with new data, changes in field content, adding/deleting fields, etc.
  • opens up new options for managing test lifecycle
  • easy to extend
  • provides a path for repurposing functional automation tests as performance tests via Artillery

Why Go?

Go:

  • compiles down to a single, small executable file on any platform, with no additional runtime
    • makes it easy to setup and reset "test runner" environments
    • makes it very easy to deploy into containers, which is where I prefer to use for test environments
    • makes it easy to integrate with CI/CD pipelines
  • creates executables that have very small RAM footprints
  • is very fast to execute (e.g. no startup delay as with JVM languages)
  • offers native support for multithreading via goroutines, which makes it easy to run lots of integration tests simultaneously
  • has cross compilation is built into the compiler, which is very convenient for creating Docker executables from a non-Linux workstation

Why command line interface only?

  • Every OS supports execution from the command line, and I want a solution that works on every OS
  • Easy to scale from running a single test to running 1000s of tests
  • Easy to leverage different test case & results storage options

Why a minimal feature set?

wilee attempts to follow the Unix philosophy as documented at https://en.Wikipedia.org/wiki/Unix_philosophy

  • “Write programs that do one thing and do it well”
  • “Expect the output of every program to become the input to another, as yet unknown, program”
  • “Write programs to work together”
  • “Write programs to handle text streams, because that is a universal interface”
  • “Use shell scripts to increase leverage and portability”
  • “Developers should design programs so that they do not print unnecessary output”
  • “Developers should design for the future by making their protocols extensible”

Why do you...?

  • Pass in the base URL via an environment variable, rather than put it in the JSON? Because I want to be able to use the same set of tests across many different test environment instances; because I want to do test execution from within immutable Docker containers, and passing in env vars to Docker containers is a very convenient configuration pattern
  • Read test cases from STDIN rather than from files? Purely for flexibility - I may not want to store my test cases in files on disc

Getting Started

These instructions will get you a copy of the project up and running on your local machine for development and testing purposes. See deployment for notes on how to deploy the project on a live system.

Download and Compile

$ git clone https://github.com/monch1962/wilee
$ cd wilee
$ go build wilee.go

You should then see a new wilee executable file. It'll be named 'wilee.exe' if you're running Windows, or 'wilee' if you're running pretty much anything else.

You may want to built wilee for a different system to that you're currently using - for example, you might be sitting in front of a Mac, but you want to build wilee to run within a Linux Docker container. In that case, change the last command to:

$ OS=linux ARCH=amd64 go build wilee.go

and you'll get a file named 'wilee' that'll run on your Linux container. It WON'T run on your Mac, because the binary won't be compatible.

Quick test

To quickly test your new wilee executable, try running on your non-Windows box

$ APP="https://jsonplaceholder.typicode.com" wilee < demo/jsonplaceholder.typicode.com/test-cases/jsonplaceholder-test.json

Assuming you've got Internet access, this will run a test against the https://jsonplaceholder.typicode.com site, and return a blob of JSON. Feel free to scroll through the output to see what's there...

Prerequisites

You'll probably want to define a set of environment variables before executing tests using wilee.

An important one is to set the target app server that you wish to test

APP="https://server:port"

wilee can either read a single test case from stdin and write results to stdout (where they can be processed by e.g. jq), or you can supply a set of test cases to be executed. In the latter case, all test cases will be executed simultaneously

TESTCASE=test-cases/jsonplaceholder-test[012].json

Installing

A step by step series of examples that tell you have to get a development env running

Say what the step will be

Give the example

And repeat

until finished

End with an example of getting some data out of the system or using it for a little demo

Running the tests

Explain how to run the automated tests for this system

Break down into end to end tests

Explain what these tests test and why

Give an example

And coding style tests

Explain what these tests test and why

Give an example

Deployment

Add additional notes about how to deploy this on a live system

Built With

Contributing

Versioning

Authors

See also the list of contributors who participated in this project.

License

This project is licensed under the MIT License - see the LICENSE.md file for details

Acknowledgments

  • Hat tip to anyone who's code was used
  • Inspiration
  • etc