diff --git a/.github/ISSUE_TEMPLATE/new-team-member.md b/.github/ISSUE_TEMPLATE/new-team-member.md
index 9b446cdd97..f82d3017e9 100644
--- a/.github/ISSUE_TEMPLATE/new-team-member.md
+++ b/.github/ISSUE_TEMPLATE/new-team-member.md
@@ -36,7 +36,7 @@ Slack Username:
 
 - [ ] Added to Slack channels:
 
-  - [ ] #data-analyses
+  - [ ] #data-analysis
   - [ ] #data-office-hours
   - [ ] #data
   - [ ] #ct-bdat-internal
diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md
index 2c4845c68e..68a278f117 100644
--- a/CONTRIBUTING.md
+++ b/CONTRIBUTING.md
@@ -1,7 +1,7 @@
 # Contribution guidelines
 
 These guidelines are meant to provide a foundation for collaboration in Cal-ITP's data services repos,
-primarily [#data-infra](https://github.com/cal-itp/data-infra).
+primarily [data-infra](https://github.com/cal-itp/data-infra).
 
 ## Issues
 
diff --git a/airflow/dags/dags.py b/airflow/dags/dags.py
index 3956c277a3..74b249f84c 100644
--- a/airflow/dags/dags.py
+++ b/airflow/dags/dags.py
@@ -7,7 +7,7 @@
 
 import airflow  # noqa
 
-# pointed at #data-infra-notify as of 2022-02-01
+# pointed at #alerts-data-infra as of 2024-02-05
 CALITP_SLACK_URL_KEY = "CALITP_SLACK_URL"
 
 # DAG Directories =============================================================
diff --git a/docs/airflow/dags-maintenance.md b/docs/airflow/dags-maintenance.md
index fefd87cc2c..e8d87c90a6 100644
--- a/docs/airflow/dags-maintenance.md
+++ b/docs/airflow/dags-maintenance.md
@@ -6,7 +6,7 @@ We use [Airflow](https://airflow.apache.org/) to orchestrate our data ingest pro
 
 ## Monitoring DAGs
 
-When an Airflow DAG task fails, it will alert to the `#data-infra-alerts` channel in Slack.
+When an Airflow DAG task fails, it will alert to the `#alerts-data-infra` channel in Slack.
 
 In that case, someone should respond according to the considerations described below and in each individual DAG's documentation (available in [each DAG subdirectory](https://github.com/cal-itp/data-infra/tree/main/airflow/dags) in the `data-infra` GitHub repository).
 
diff --git a/docs/transit_database/transitdatabase.md b/docs/transit_database/transitdatabase.md
index b62b9dfa57..aeabd0fe63 100644
--- a/docs/transit_database/transitdatabase.md
+++ b/docs/transit_database/transitdatabase.md
@@ -1,6 +1,6 @@
 # Transit Database (Airtable)
 
-The Cal-ITP Airtable Transit Database stores key relationships about how transit services are organized and operated in California as well as how well they are performing. See Evan or post in the `#airtable-data` Slack channel to get a link and gain access.
+The Cal-ITP Airtable Transit Database stores key relationships about how transit services are organized and operated in California as well as how well they are performing. See Evan or post in the `#data-airtable` Slack channel to get a link and gain access.
 
 Important Airtable documentation is maintained elsewhere:
 
diff --git a/runbooks/workflow/disk-space.md b/runbooks/workflow/disk-space.md
index 836a081cd5..0f66033d1a 100644
--- a/runbooks/workflow/disk-space.md
+++ b/runbooks/workflow/disk-space.md
@@ -1,6 +1,6 @@
 # Resolving Disk Space Usage and Pod Offset Errors
 
-We have an [alert](https://monitoring.calitp.org/alerting/grafana/Geo72Nf4z/view) that will trigger when any Kubernetes volume is more than 80% full, placing a message in the #data-infra-alerts channel of the Cal-ITP Slack group. The resolution path will depend on the affected service.
+We have an [alert](https://monitoring.calitp.org/alerting/grafana/Geo72Nf4z/view) that will trigger when any Kubernetes volume is more than 80% full, placing a message in the `#alerts-data-infra` channel of the Cal-ITP Slack group. The resolution path will depend on the affected service.
 
 After following any of these specific runbook sections, check the Disk Usage section of the general [Kubernetes dashboard](https://monitoring.calitp.org/d/oWe9aYxmk/1-kubernetes-deployment-statefulset-daemonset-metrics) in Grafana to verify the disk space consumption has decreased to a safe level for any impacted volumes. The alert should also show as resolved in Slack a couple minutes after recovery.