Skip to content

Commit

Permalink
Fix: typos (#659)
Browse files Browse the repository at this point in the history
  • Loading branch information
omahs authored Jul 13, 2023
1 parent 144681b commit 3ce98d1
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion BUILD.md
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ Run the following command to retrieve the image:
docker image pull ghcr.io/delvtech/elf-simulations/dcdefc6153fc
```

If you you get a 401 error, you'll need to make a GH Personal Access Token with write accesss.
If you get a 401 error, you'll need to make a GH Personal Access Token with write access.

You can optionally provide a tag

Expand Down
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ Please refer to [BUILD.md](https://github.com/delvtech/elf-simulations/blob/main

## Testing

Testing is achieved with [py.test](https://docs.pytest.org/en/latest/contents.html). You can run all tests from the repository root directory by runing `python -m pytest`, or you can pick a specific test in the `tests/` folder with `python -m pytest tests/{test_file.py}`.
Testing is achieved with [py.test](https://docs.pytest.org/en/latest/contents.html). You can run all tests from the repository root directory by running `python -m pytest`, or you can pick a specific test in the `tests/` folder with `python -m pytest tests/{test_file.py}`.

## Coverage

Expand Down Expand Up @@ -77,7 +77,7 @@ evm_bots.py bots_config.default.json
```

## Data pipeline
The data pipeline queries the running chain and exports data to a postgres database. The `infra` repostitory spins up a local postgres instance via Docker, and the data piepline will point to this by default. Optinally, you can also configure the backend database by specifying the following environmental variables (for example, in a `.env` file in the base of the repo):
The data pipeline queries the running chain and exports data to a postgres database. The `infra` repository spins up a local postgres instance via Docker, and the data pipeline will point to this by default. Optionally, you can also configure the backend database by specifying the following environmental variables (for example, in a `.env` file in the base of the repo):

```bash
POSTGRES_USER="admin"
Expand Down

0 comments on commit 3ce98d1

Please sign in to comment.