From cb9d2b4c5c7da1e668aeac536ec9f55dbc8c9932 Mon Sep 17 00:00:00 2001 From: fryorcraken Date: Tue, 20 Aug 2024 15:21:11 +1000 Subject: [PATCH 1/2] chore: use submodule nph in CI to check lint --- .github/workflows/ci.yml | 37 ++++++++++++++++++++++++------------- Makefile | 10 ++++++---- vendor/nph | 2 +- 3 files changed, 31 insertions(+), 18 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 464a2fe8e4..43f3cc15f4 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -117,7 +117,7 @@ jobs: export MAKEFLAGS="-j1" export NIMFLAGS="--colors:off -d:chronicles_colors:none" - + make V=1 LOG_LEVEL=DEBUG QUICK_AND_DIRTY_COMPILER=1 POSTGRES=$postgres_enabled test testwakunode2 build-docker-image: @@ -141,25 +141,36 @@ jobs: nim_wakunode_image: ${{ needs.build-docker-image.outputs.image }} test_type: node-optional debug: waku* - + lint: name: "Lint" runs-on: ubuntu-latest + needs: build steps: - - name: Checkout - uses: actions/checkout@v4 + - name: Checkout code + uses: actions/checkout@v3 + + - name: Get submodules hash + id: submodules + run: | + echo "hash=$(git submodule status | awk '{print $1}' | sort | shasum -a 256 | sed 's/[ -]*//g')" >> $GITHUB_OUTPUT + + - name: Cache submodules + uses: actions/cache@v3 with: - fetch-depth: 2 # In PR, has extra merge commit: ^1 = PR, ^2 = base + path: | + vendor/ + .git/modules + key: ${{ runner.os }}-vendor-modules-${{ steps.submodules.outputs.hash }} + + - name: Build nph + run: | + make build-nph - name: Check nph formatting - # Pin nph to a specific version to avoid sudden style differences. - # Updating nph version should be accompanied with running the new - # version on the fluffy directory. run: | - VERSION="v0.5.1" - ARCHIVE="nph-linux_x64.tar.gz" - curl -L "https://github.com/arnetheduck/nph/releases/download/${VERSION}/${ARCHIVE}" -o ${ARCHIVE} - tar -xzf ${ARCHIVE} shopt -s extglob # Enable extended globbing - ./nph examples waku tests tools apps *.@(nim|nims|nimble) + NPH=$(make print-nph-path) + echo "using nph at ${NPH}" + "${NPH}" examples waku tests tools apps *.@(nim|nims|nimble) git diff --exit-code diff --git a/Makefile b/Makefile index 075c1b8d54..df4cba206b 100644 --- a/Makefile +++ b/Makefile @@ -270,12 +270,10 @@ networkmonitor: | build deps librln ############ ## Format ## ############ -.PHONY: build-nph clean-nph install-nph +.PHONY: build-nph install-nph clean-nph print-nph-path build-nph: -ifeq ("$(wildcard $(NPH))","") - $(ENV_SCRIPT) nim c vendor/nph/src/nph.nim -endif + $(ENV_SCRIPT) nim c vendor/nph/src/nph.nim GIT_PRE_COMMIT_HOOK := .git/hooks/pre-commit @@ -294,6 +292,10 @@ nph/%: build-nph clean-nph: rm -f $(NPH) +# To avoid hardcoding nph binary location in several places +print-nph-path: + echo "$(NPH)" + clean: | clean-nph ################### diff --git a/vendor/nph b/vendor/nph index de5cd4823e..31bdced07d 160000 --- a/vendor/nph +++ b/vendor/nph @@ -1 +1 @@ -Subproject commit de5cd4823e63424adb58ef3717524348ae6c4d87 +Subproject commit 31bdced07d3dc3d254669bd94210101c701deeda From 993f6e1a3743ee2dfb8a202b9025e699b7a9d5e0 Mon Sep 17 00:00:00 2001 From: fryorcraken Date: Tue, 10 Sep 2024 11:08:04 +1000 Subject: [PATCH 2/2] chore: reformat with nph v0.6.0 --- apps/liteprotocoltester/liteprotocoltester.nim | 3 +-- apps/liteprotocoltester/tester_config.nim | 3 +-- tests/testlib/wakunode.nim | 6 ++---- tests/wakunode_rest/test_rest_store.nim | 12 ++++-------- waku/node/peer_manager/waku_peer_store.nim | 3 +-- waku/waku_api/rest/filter/types.nim | 12 ++++-------- waku/waku_api/rest/legacy_store/types.nim | 3 +-- waku/waku_api/rest/lightpush/types.nim | 3 +-- waku/waku_api/rest/relay/types.nim | 3 +-- .../waku_archive/driver/sqlite_driver/queries.nim | 12 ++++-------- .../driver/sqlite_driver/queries.nim | 15 +++++---------- waku/waku_keystore/utils.nim | 3 +-- 12 files changed, 26 insertions(+), 52 deletions(-) diff --git a/apps/liteprotocoltester/liteprotocoltester.nim b/apps/liteprotocoltester/liteprotocoltester.nim index a4589b60b6..a109a7bb0e 100644 --- a/apps/liteprotocoltester/liteprotocoltester.nim +++ b/apps/liteprotocoltester/liteprotocoltester.nim @@ -75,8 +75,7 @@ when isMainModule: wnconf: WakuNodeConf, sources: auto ) {.gcsafe, raises: [ConfigurationError].} = echo "Loading secondary configuration file into WakuNodeConf" - sources.addConfigFile(Toml, configFile) - , + sources.addConfigFile(Toml, configFile), ) except CatchableError: error "Loading Waku configuration failed", error = getCurrentExceptionMsg() diff --git a/apps/liteprotocoltester/tester_config.nim b/apps/liteprotocoltester/tester_config.nim index 5683b544f5..ccfdcc24c8 100644 --- a/apps/liteprotocoltester/tester_config.nim +++ b/apps/liteprotocoltester/tester_config.nim @@ -159,8 +159,7 @@ proc load*(T: type LiteProtocolTesterConf, version = ""): ConfResult[T] = secondarySources = proc( conf: LiteProtocolTesterConf, sources: auto ) {.gcsafe, raises: [ConfigurationError].} = - sources.addConfigFile(Envvar, InputFile("liteprotocoltester")) - , + sources.addConfigFile(Envvar, InputFile("liteprotocoltester")), ) ok(conf) except CatchableError: diff --git a/tests/testlib/wakunode.nim b/tests/testlib/wakunode.nim index eb679859b8..b2cff54e30 100644 --- a/tests/testlib/wakunode.nim +++ b/tests/testlib/wakunode.nim @@ -140,14 +140,12 @@ proc newTestWakuNode*( if secureKey != "": some(secureKey) else: - none(string) - , + none(string), secureCert = if secureCert != "": some(secureCert) else: - none(string) - , + none(string), agentString = agentString, ) diff --git a/tests/wakunode_rest/test_rest_store.nim b/tests/wakunode_rest/test_rest_store.nim index 32e8151db3..24248ee4b3 100644 --- a/tests/wakunode_rest/test_rest_store.nim +++ b/tests/wakunode_rest/test_rest_store.nim @@ -297,8 +297,7 @@ procSuite "Waku Rest API - Store v3": if reqHash.isSome(): reqHash.get().toRestStringWakuMessageHash() else: - "" - , # base64-encoded digest. Empty ignores the field. + "", # base64-encoded digest. Empty ignores the field. "true", # ascending "7", # page size. Empty implies default page size. ) @@ -790,8 +789,7 @@ procSuite "Waku Rest API - Store v3": if reqHash.isSome(): reqHash.get().toRestStringWakuMessageHash() else: - "" - , # base64-encoded digest. Empty ignores the field. + "", # base64-encoded digest. Empty ignores the field. "true", # ascending "3", # page size. Empty implies default page size. ) @@ -827,8 +825,7 @@ procSuite "Waku Rest API - Store v3": if reqHash.isSome(): reqHash.get().toRestStringWakuMessageHash() else: - "" - , # base64-encoded digest. Empty ignores the field. + "", # base64-encoded digest. Empty ignores the field. ) check: @@ -850,8 +847,7 @@ procSuite "Waku Rest API - Store v3": if reqHash.isSome(): reqHash.get().toRestStringWakuMessageHash() else: - "" - , # base64-encoded digest. Empty ignores the field. + "", # base64-encoded digest. Empty ignores the field. "true", # ascending "5", # page size. Empty implies default page size. ) diff --git a/waku/node/peer_manager/waku_peer_store.nim b/waku/node/peer_manager/waku_peer_store.nim index a7db829f29..09d6ebc658 100644 --- a/waku/node/peer_manager/waku_peer_store.nim +++ b/waku/node/peer_manager/waku_peer_store.nim @@ -55,8 +55,7 @@ proc get*(peerStore: PeerStore, peerId: PeerID): RemotePeerInfo = if peerStore[ENRBook][peerId] != default(enr.Record): some(peerStore[ENRBook][peerId]) else: - none(enr.Record) - , + none(enr.Record), protocols: peerStore[ProtoBook][peerId], agent: peerStore[AgentBook][peerId], protoVersion: peerStore[ProtoVersionBook][peerId], diff --git a/waku/waku_api/rest/filter/types.nim b/waku/waku_api/rest/filter/types.nim index 7497d6a78e..0506a7a741 100644 --- a/waku/waku_api/rest/filter/types.nim +++ b/waku/waku_api/rest/filter/types.nim @@ -58,8 +58,7 @@ proc toFilterWakuMessage*(msg: WakuMessage): FilterWakuMessage = if msg.meta.len > 0: some(base64.encode(msg.meta)) else: - none(Base64String) - , + none(Base64String), ephemeral: some(msg.ephemeral), ) @@ -239,8 +238,7 @@ proc readValue*( if pubsubTopic.isNone() or pubsubTopic.get() == "": none(string) else: - some(pubsubTopic.get()) - , + some(pubsubTopic.get()), contentFilters: contentFilters.get(), ) @@ -315,8 +313,7 @@ proc readValue*( if pubsubTopic.isNone() or pubsubTopic.get() == "": none(string) else: - some(pubsubTopic.get()) - , + some(pubsubTopic.get()), contentFilters: contentFilters.get(), ) @@ -364,8 +361,7 @@ proc readValue*( if pubsubTopic.isNone() or pubsubTopic.get() == "": none(string) else: - some(pubsubTopic.get()) - , + some(pubsubTopic.get()), contentFilters: contentFilters.get(), ) diff --git a/waku/waku_api/rest/legacy_store/types.nim b/waku/waku_api/rest/legacy_store/types.nim index 0ae36e05e5..eee3ac2d82 100644 --- a/waku/waku_api/rest/legacy_store/types.nim +++ b/waku/waku_api/rest/legacy_store/types.nim @@ -110,8 +110,7 @@ proc toStoreResponseRest*(histResp: HistoryResponse): StoreResponseRest = if message.meta.len > 0: some(base64.encode(message.meta)) else: - none(Base64String) - , + none(Base64String), ) var storeWakuMsgs: seq[StoreWakuMessage] diff --git a/waku/waku_api/rest/lightpush/types.nim b/waku/waku_api/rest/lightpush/types.nim index 3d8d0715fe..f499600b7a 100644 --- a/waku/waku_api/rest/lightpush/types.nim +++ b/waku/waku_api/rest/lightpush/types.nim @@ -62,7 +62,6 @@ proc readValue*( if pubsubTopic.isNone() or pubsubTopic.get() == "": none(string) else: - some(pubsubTopic.get()) - , + some(pubsubTopic.get()), message: message.get(), ) diff --git a/waku/waku_api/rest/relay/types.nim b/waku/waku_api/rest/relay/types.nim index 98957c747e..1ffb59384e 100644 --- a/waku/waku_api/rest/relay/types.nim +++ b/waku/waku_api/rest/relay/types.nim @@ -34,8 +34,7 @@ proc toRelayWakuMessage*(msg: WakuMessage): RelayWakuMessage = if msg.meta.len > 0: some(base64.encode(msg.meta)) else: - none(Base64String) - , + none(Base64String), ephemeral: some(msg.ephemeral), ) diff --git a/waku/waku_archive/driver/sqlite_driver/queries.nim b/waku/waku_archive/driver/sqlite_driver/queries.nim index 0a167937ec..6fafc06ebe 100644 --- a/waku/waku_archive/driver/sqlite_driver/queries.nim +++ b/waku/waku_archive/driver/sqlite_driver/queries.nim @@ -82,8 +82,7 @@ proc createTable*(db: SqliteDatabase): DatabaseResult[void] = ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -98,8 +97,7 @@ proc createOldestMessageTimestampIndex*(db: SqliteDatabase): DatabaseResult[void ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -184,8 +182,7 @@ proc deleteMessagesOlderThanTimestamp*( ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -206,8 +203,7 @@ proc deleteOldestMessagesNotWithinLimit*( ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() diff --git a/waku/waku_archive_legacy/driver/sqlite_driver/queries.nim b/waku/waku_archive_legacy/driver/sqlite_driver/queries.nim index 76d9755e56..47f1d86ae7 100644 --- a/waku/waku_archive_legacy/driver/sqlite_driver/queries.nim +++ b/waku/waku_archive_legacy/driver/sqlite_driver/queries.nim @@ -96,8 +96,7 @@ proc createTable*(db: SqliteDatabase): DatabaseResult[void] = ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -112,8 +111,7 @@ proc createOldestMessageTimestampIndex*(db: SqliteDatabase): DatabaseResult[void ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -127,8 +125,7 @@ proc createHistoryQueryIndex*(db: SqliteDatabase): DatabaseResult[void] = ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -226,8 +223,7 @@ proc deleteMessagesOlderThanTimestamp*( ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() @@ -248,8 +244,7 @@ proc deleteOldestMessagesNotWithinLimit*( ?db.query( query, proc(s: ptr sqlite3_stmt) = - discard - , + discard, ) return ok() diff --git a/waku/waku_keystore/utils.nim b/waku/waku_keystore/utils.nim index 422b96106c..88c10207e2 100644 --- a/waku/waku_keystore/utils.nim +++ b/waku/waku_keystore/utils.nim @@ -9,8 +9,7 @@ proc hasKeys*(data: JsonNode, keys: openArray[string]): bool = return all( keys, proc(key: string): bool = - return data.hasKey(key) - , + return data.hasKey(key), ) # Safely saves a Keystore's JsonNode to disk.