mirror of
https://github.com/domainaware/parsedmarc.git
synced 2026-03-17 03:55:59 +00:00
Compare commits
47 Commits
9.0.0
...
copilot/re
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2174f23eb5 | ||
|
|
febbb107c4 | ||
|
|
9a64b494e7 | ||
|
|
e93209c766 | ||
|
|
d1c22466be | ||
|
|
3d1b2522d3 | ||
|
|
af9ad568ec | ||
|
|
748164d177 | ||
|
|
487e5e1149 | ||
|
|
73010cf964 | ||
|
|
a4a5475aa8 | ||
|
|
dab78880df | ||
|
|
fb54e3b742 | ||
|
|
6799f10364 | ||
|
|
445c9565a4 | ||
|
|
4b786846ae | ||
|
|
23ae563cd8 | ||
|
|
cdd000e675 | ||
|
|
7d58abc67b | ||
|
|
a18ae439de | ||
|
|
d7061330a8 | ||
|
|
9d5654b8ec | ||
|
|
a0e0070dd0 | ||
|
|
cf3b7f2c29 | ||
|
|
d312522ab7 | ||
|
|
888d717476 | ||
|
|
1127f65fbb | ||
|
|
d017dfcddf | ||
|
|
5fae99aacc | ||
|
|
ba57368ac3 | ||
|
|
dc6ee5de98 | ||
|
|
158d63d205 | ||
|
|
f1933b906c | ||
|
|
4b98d795ff | ||
|
|
b1356f7dfc | ||
|
|
1969196e1a | ||
|
|
553f15f6a9 | ||
|
|
1fc9f638e2 | ||
|
|
48bff504b4 | ||
|
|
681b7cbf85 | ||
|
|
0922d6e83a | ||
|
|
baf3f95fb1 | ||
|
|
a51f945305 | ||
|
|
55dbf8e3db | ||
|
|
00267c9847 | ||
|
|
51356175e1 | ||
|
|
3be10d30dd |
10
.github/workflows/docker.yml
vendored
10
.github/workflows/docker.yml
vendored
@@ -24,11 +24,11 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v5
|
||||||
|
|
||||||
- name: Docker meta
|
- name: Docker meta
|
||||||
id: meta
|
id: meta
|
||||||
uses: docker/metadata-action@v3
|
uses: docker/metadata-action@v5
|
||||||
with:
|
with:
|
||||||
images: |
|
images: |
|
||||||
${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
||||||
@@ -40,16 +40,14 @@ jobs:
|
|||||||
type=semver,pattern={{major}}.{{minor}}
|
type=semver,pattern={{major}}.{{minor}}
|
||||||
|
|
||||||
- name: Log in to the Container registry
|
- name: Log in to the Container registry
|
||||||
# https://github.com/docker/login-action/releases/tag/v2.0.0
|
uses: docker/login-action@v3
|
||||||
uses: docker/login-action@49ed152c8eca782a232dede0303416e8f356c37b
|
|
||||||
with:
|
with:
|
||||||
registry: ${{ env.REGISTRY }}
|
registry: ${{ env.REGISTRY }}
|
||||||
username: ${{ github.actor }}
|
username: ${{ github.actor }}
|
||||||
password: ${{ secrets.GITHUB_TOKEN }}
|
password: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
|
||||||
- name: Build and push Docker image
|
- name: Build and push Docker image
|
||||||
# https://github.com/docker/build-push-action/releases/tag/v3.0.0
|
uses: docker/build-push-action@v6
|
||||||
uses: docker/build-push-action@e551b19e49efd4e98792db7592c17c09b89db8d8
|
|
||||||
with:
|
with:
|
||||||
context: .
|
context: .
|
||||||
push: ${{ github.event_name == 'release' }}
|
push: ${{ github.event_name == 'release' }}
|
||||||
|
|||||||
12
.github/workflows/python-tests.yml
vendored
12
.github/workflows/python-tests.yml
vendored
@@ -15,7 +15,7 @@ jobs:
|
|||||||
|
|
||||||
services:
|
services:
|
||||||
elasticsearch:
|
elasticsearch:
|
||||||
image: elasticsearch:8.18.2
|
image: elasticsearch:8.19.7
|
||||||
env:
|
env:
|
||||||
discovery.type: single-node
|
discovery.type: single-node
|
||||||
cluster.name: parsedmarc-cluster
|
cluster.name: parsedmarc-cluster
|
||||||
@@ -33,15 +33,15 @@ jobs:
|
|||||||
python-version: ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
python-version: ["3.9", "3.10", "3.11", "3.12", "3.13"]
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v5
|
||||||
- name: Set up Python ${{ matrix.python-version }}
|
- name: Set up Python ${{ matrix.python-version }}
|
||||||
uses: actions/setup-python@v5
|
uses: actions/setup-python@v6
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python-version }}
|
python-version: ${{ matrix.python-version }}
|
||||||
- name: Install system dependencies
|
- name: Install system dependencies
|
||||||
run: |
|
run: |
|
||||||
sudo apt-get update
|
sudo apt-get -q update
|
||||||
sudo apt-get install -y libemail-outlook-message-perl
|
sudo apt-get -qy install libemail-outlook-message-perl
|
||||||
- name: Install Python dependencies
|
- name: Install Python dependencies
|
||||||
run: |
|
run: |
|
||||||
python -m pip install --upgrade pip
|
python -m pip install --upgrade pip
|
||||||
@@ -65,6 +65,6 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
hatch build
|
hatch build
|
||||||
- name: Upload coverage to Codecov
|
- name: Upload coverage to Codecov
|
||||||
uses: codecov/codecov-action@v4
|
uses: codecov/codecov-action@v5
|
||||||
with:
|
with:
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
|
|||||||
13
.vscode/launch.json
vendored
13
.vscode/launch.json
vendored
@@ -19,20 +19,11 @@
|
|||||||
"console": "integratedTerminal"
|
"console": "integratedTerminal"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "sample.eml",
|
"name": "sample",
|
||||||
"type": "debugpy",
|
"type": "debugpy",
|
||||||
"request": "launch",
|
"request": "launch",
|
||||||
"module": "parsedmarc.cli",
|
"module": "parsedmarc.cli",
|
||||||
"args": ["samples/private/sample.eml"]
|
"args": ["samples/private/sample"]
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "find_sus_domains.py",
|
|
||||||
"type": "debugpy",
|
|
||||||
"request": "launch",
|
|
||||||
"program": "find_sus_domains.py",
|
|
||||||
"args": ["-i", "unknown_domains.txt", "-o", "sus_domains.csv"],
|
|
||||||
"cwd": "${workspaceFolder}/parsedmarc/resources/maps",
|
|
||||||
"console": "integratedTerminal"
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name": "sortlists.py",
|
"name": "sortlists.py",
|
||||||
|
|||||||
16
.vscode/settings.json
vendored
16
.vscode/settings.json
vendored
@@ -1,4 +1,14 @@
|
|||||||
{
|
{
|
||||||
|
"[python]": {
|
||||||
|
"editor.defaultFormatter": "charliermarsh.ruff",
|
||||||
|
"editor.formatOnSave": true,
|
||||||
|
|
||||||
|
// Let Ruff handle lint fixes + import sorting on save
|
||||||
|
"editor.codeActionsOnSave": {
|
||||||
|
"source.fixAll.ruff": "explicit",
|
||||||
|
"source.organizeImports.ruff": "explicit"
|
||||||
|
}
|
||||||
|
},
|
||||||
"markdownlint.config": {
|
"markdownlint.config": {
|
||||||
"MD024": false
|
"MD024": false
|
||||||
},
|
},
|
||||||
@@ -36,6 +46,7 @@
|
|||||||
"exampleuser",
|
"exampleuser",
|
||||||
"expiringdict",
|
"expiringdict",
|
||||||
"fieldlist",
|
"fieldlist",
|
||||||
|
"GELF",
|
||||||
"genindex",
|
"genindex",
|
||||||
"geoip",
|
"geoip",
|
||||||
"geoipupdate",
|
"geoipupdate",
|
||||||
@@ -65,17 +76,20 @@
|
|||||||
"mailrelay",
|
"mailrelay",
|
||||||
"mailsuite",
|
"mailsuite",
|
||||||
"maxdepth",
|
"maxdepth",
|
||||||
|
"MAXHEADERS",
|
||||||
"maxmind",
|
"maxmind",
|
||||||
"mbox",
|
"mbox",
|
||||||
"mfrom",
|
"mfrom",
|
||||||
"michaeldavie",
|
"michaeldavie",
|
||||||
"mikesiegel",
|
"mikesiegel",
|
||||||
|
"Mimecast",
|
||||||
"mitigations",
|
"mitigations",
|
||||||
"MMDB",
|
"MMDB",
|
||||||
"modindex",
|
"modindex",
|
||||||
"msgconvert",
|
"msgconvert",
|
||||||
"msgraph",
|
"msgraph",
|
||||||
"MSSP",
|
"MSSP",
|
||||||
|
"multiprocess",
|
||||||
"Munge",
|
"Munge",
|
||||||
"ndjson",
|
"ndjson",
|
||||||
"newkey",
|
"newkey",
|
||||||
@@ -86,6 +100,7 @@
|
|||||||
"nosniff",
|
"nosniff",
|
||||||
"nwettbewerb",
|
"nwettbewerb",
|
||||||
"opensearch",
|
"opensearch",
|
||||||
|
"opensearchpy",
|
||||||
"parsedmarc",
|
"parsedmarc",
|
||||||
"passsword",
|
"passsword",
|
||||||
"Postorius",
|
"Postorius",
|
||||||
@@ -123,6 +138,7 @@
|
|||||||
"truststore",
|
"truststore",
|
||||||
"Übersicht",
|
"Übersicht",
|
||||||
"uids",
|
"uids",
|
||||||
|
"Uncategorized",
|
||||||
"unparasable",
|
"unparasable",
|
||||||
"uper",
|
"uper",
|
||||||
"urllib",
|
"urllib",
|
||||||
|
|||||||
606
CHANGELOG.md
606
CHANGELOG.md
File diff suppressed because it is too large
Load Diff
31
README.md
31
README.md
@@ -23,11 +23,10 @@ ProofPoint Email Fraud Defense, and Valimail.
|
|||||||
|
|
||||||
## Help Wanted
|
## Help Wanted
|
||||||
|
|
||||||
This project is maintained by one developer. Please consider
|
This project is maintained by one developer. Please consider reviewing the open
|
||||||
reviewing the open
|
[issues](https://github.com/domainaware/parsedmarc/issues) to see how you can
|
||||||
[issues](https://github.com/domainaware/parsedmarc/issues) to see how
|
contribute code, documentation, or user support. Assistance on the pinned
|
||||||
you can contribute code, documentation, or user support. Assistance on
|
issues would be particularly helpful.
|
||||||
the pinned issues would be particularly helpful.
|
|
||||||
|
|
||||||
Thanks to all
|
Thanks to all
|
||||||
[contributors](https://github.com/domainaware/parsedmarc/graphs/contributors)!
|
[contributors](https://github.com/domainaware/parsedmarc/graphs/contributors)!
|
||||||
@@ -42,6 +41,24 @@ Thanks to all
|
|||||||
- Consistent data structures
|
- Consistent data structures
|
||||||
- Simple JSON and/or CSV output
|
- Simple JSON and/or CSV output
|
||||||
- Optionally email the results
|
- Optionally email the results
|
||||||
- Optionally send the results to Elasticsearch, Opensearch, and/or Splunk, for use
|
- Optionally send the results to Elasticsearch, Opensearch, and/or Splunk, for
|
||||||
with premade dashboards
|
use with premade dashboards
|
||||||
- Optionally send reports to Apache Kafka
|
- Optionally send reports to Apache Kafka
|
||||||
|
|
||||||
|
## Python Compatibility
|
||||||
|
|
||||||
|
This project supports the following Python versions, which are either actively maintained or are the default versions
|
||||||
|
for RHEL or Debian.
|
||||||
|
|
||||||
|
| Version | Supported | Reason |
|
||||||
|
|---------|-----------|------------------------------------------------------------|
|
||||||
|
| < 3.6 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.6 | ❌ | Used in RHEL 8, but not supported by project dependencies |
|
||||||
|
| 3.7 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.8 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.9 | ✅ | Supported until August 2026 (Debian 11); May 2032 (RHEL 9) |
|
||||||
|
| 3.10 | ✅ | Actively maintained |
|
||||||
|
| 3.11 | ✅ | Actively maintained; supported until June 2028 (Debian 12) |
|
||||||
|
| 3.12 | ✅ | Actively maintained; supported until May 2035 (RHEL 10) |
|
||||||
|
| 3.13 | ✅ | Actively maintained; supported until June 2030 (Debian 13) |
|
||||||
|
| 3.14 | ❌ | Not currently supported due to Not currently supported due to [this imapclient bug](https://github.com/mjs/imapclient/issues/618)|
|
||||||
|
|||||||
3
build.sh
3
build.sh
@@ -9,12 +9,11 @@ fi
|
|||||||
. venv/bin/activate
|
. venv/bin/activate
|
||||||
pip install .[build]
|
pip install .[build]
|
||||||
ruff format .
|
ruff format .
|
||||||
ruff check .
|
|
||||||
cd docs
|
cd docs
|
||||||
make clean
|
make clean
|
||||||
make html
|
make html
|
||||||
touch build/html/.nojekyll
|
touch build/html/.nojekyll
|
||||||
if [ -d "./../parsedmarc-docs" ]; then
|
if [ -d "../../parsedmarc-docs" ]; then
|
||||||
cp -rf build/html/* ../../parsedmarc-docs/
|
cp -rf build/html/* ../../parsedmarc-docs/
|
||||||
fi
|
fi
|
||||||
cd ..
|
cd ..
|
||||||
|
|||||||
@@ -1,8 +1,6 @@
|
|||||||
version: '3.7'
|
|
||||||
|
|
||||||
services:
|
services:
|
||||||
elasticsearch:
|
elasticsearch:
|
||||||
image: docker.elastic.co/elasticsearch/elasticsearch:8.3.1
|
image: docker.elastic.co/elasticsearch/elasticsearch:8.19.7
|
||||||
environment:
|
environment:
|
||||||
- network.host=127.0.0.1
|
- network.host=127.0.0.1
|
||||||
- http.host=0.0.0.0
|
- http.host=0.0.0.0
|
||||||
@@ -14,7 +12,7 @@ services:
|
|||||||
- xpack.security.enabled=false
|
- xpack.security.enabled=false
|
||||||
- xpack.license.self_generated.type=basic
|
- xpack.license.self_generated.type=basic
|
||||||
ports:
|
ports:
|
||||||
- 127.0.0.1:9200:9200
|
- "127.0.0.1:9200:9200"
|
||||||
ulimits:
|
ulimits:
|
||||||
memlock:
|
memlock:
|
||||||
soft: -1
|
soft: -1
|
||||||
@@ -30,7 +28,7 @@ services:
|
|||||||
retries: 24
|
retries: 24
|
||||||
|
|
||||||
opensearch:
|
opensearch:
|
||||||
image: opensearchproject/opensearch:2.18.0
|
image: opensearchproject/opensearch:2
|
||||||
environment:
|
environment:
|
||||||
- network.host=127.0.0.1
|
- network.host=127.0.0.1
|
||||||
- http.host=0.0.0.0
|
- http.host=0.0.0.0
|
||||||
@@ -41,7 +39,7 @@ services:
|
|||||||
- bootstrap.memory_lock=true
|
- bootstrap.memory_lock=true
|
||||||
- OPENSEARCH_INITIAL_ADMIN_PASSWORD=${OPENSEARCH_INITIAL_ADMIN_PASSWORD}
|
- OPENSEARCH_INITIAL_ADMIN_PASSWORD=${OPENSEARCH_INITIAL_ADMIN_PASSWORD}
|
||||||
ports:
|
ports:
|
||||||
- 127.0.0.1:9201:9200
|
- "127.0.0.1:9201:9200"
|
||||||
ulimits:
|
ulimits:
|
||||||
memlock:
|
memlock:
|
||||||
soft: -1
|
soft: -1
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ from parsedmarc import __version__
|
|||||||
# -- Project information -----------------------------------------------------
|
# -- Project information -----------------------------------------------------
|
||||||
|
|
||||||
project = "parsedmarc"
|
project = "parsedmarc"
|
||||||
copyright = "2018 - 2023, Sean Whalen and contributors"
|
copyright = "2018 - 2025, Sean Whalen and contributors"
|
||||||
author = "Sean Whalen and contributors"
|
author = "Sean Whalen and contributors"
|
||||||
|
|
||||||
# The version info for the project you're documenting, acts as replacement for
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
|
|||||||
@@ -45,6 +45,24 @@ and Valimail.
|
|||||||
with premade dashboards
|
with premade dashboards
|
||||||
- Optionally send reports to Apache Kafka
|
- Optionally send reports to Apache Kafka
|
||||||
|
|
||||||
|
## Python Compatibility
|
||||||
|
|
||||||
|
This project supports the following Python versions, which are either actively maintained or are the default versions
|
||||||
|
for RHEL or Debian.
|
||||||
|
|
||||||
|
| Version | Supported | Reason |
|
||||||
|
|---------|-----------|------------------------------------------------------------|
|
||||||
|
| < 3.6 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.6 | ❌ | Used in RHEL 8, but not supported by project dependencies |
|
||||||
|
| 3.7 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.8 | ❌ | End of Life (EOL) |
|
||||||
|
| 3.9 | ✅ | Supported until August 2026 (Debian 11); May 2032 (RHEL 9) |
|
||||||
|
| 3.10 | ✅ | Actively maintained |
|
||||||
|
| 3.11 | ✅ | Actively maintained; supported until June 2028 (Debian 12) |
|
||||||
|
| 3.12 | ✅ | Actively maintained; supported until May 2035 (RHEL 10) |
|
||||||
|
| 3.13 | ✅ | Actively maintained; supported until June 2030 (Debian 13) |
|
||||||
|
| 3.14 | ❌ | Not currently supported due to [this imapclient bug](https://github.com/mjs/imapclient/issues/618)|
|
||||||
|
|
||||||
```{toctree}
|
```{toctree}
|
||||||
:caption: 'Contents'
|
:caption: 'Contents'
|
||||||
:maxdepth: 2
|
:maxdepth: 2
|
||||||
|
|||||||
@@ -199,7 +199,7 @@ sudo apt-get install libemail-outlook-message-perl
|
|||||||
[geoipupdate releases page on github]: https://github.com/maxmind/geoipupdate/releases
|
[geoipupdate releases page on github]: https://github.com/maxmind/geoipupdate/releases
|
||||||
[ip to country lite database]: https://db-ip.com/db/download/ip-to-country-lite
|
[ip to country lite database]: https://db-ip.com/db/download/ip-to-country-lite
|
||||||
[license keys]: https://www.maxmind.com/en/accounts/current/license-key
|
[license keys]: https://www.maxmind.com/en/accounts/current/license-key
|
||||||
[maxmind geoipupdate page]: https://dev.maxmind.com/geoip/geoipupdate/
|
[maxmind geoipupdate page]: https://dev.maxmind.com/geoip/updating-databases/
|
||||||
[maxmind geolite2 country database]: https://dev.maxmind.com/geoip/geolite2-free-geolocation-data
|
[maxmind geolite2 country database]: https://dev.maxmind.com/geoip/geolite2-free-geolocation-data
|
||||||
[registering for a free geolite2 account]: https://www.maxmind.com/en/geolite2/signup
|
[registering for a free geolite2 account]: https://www.maxmind.com/en/geolite2/signup
|
||||||
[to comply with various privacy regulations]: https://blog.maxmind.com/2019/12/18/significant-changes-to-accessing-and-using-geolite2-databases/
|
[to comply with various privacy regulations]: https://blog.maxmind.com/2019/12/18/significant-changes-to-accessing-and-using-geolite2-databases/
|
||||||
|
|||||||
@@ -23,6 +23,8 @@ of the report schema.
|
|||||||
"report_id": "9391651994964116463",
|
"report_id": "9391651994964116463",
|
||||||
"begin_date": "2012-04-27 20:00:00",
|
"begin_date": "2012-04-27 20:00:00",
|
||||||
"end_date": "2012-04-28 19:59:59",
|
"end_date": "2012-04-28 19:59:59",
|
||||||
|
"timespan_requires_normalization": false,
|
||||||
|
"original_timespan_seconds": 86399,
|
||||||
"errors": []
|
"errors": []
|
||||||
},
|
},
|
||||||
"policy_published": {
|
"policy_published": {
|
||||||
@@ -39,8 +41,10 @@ of the report schema.
|
|||||||
"source": {
|
"source": {
|
||||||
"ip_address": "72.150.241.94",
|
"ip_address": "72.150.241.94",
|
||||||
"country": "US",
|
"country": "US",
|
||||||
"reverse_dns": "adsl-72-150-241-94.shv.bellsouth.net",
|
"reverse_dns": null,
|
||||||
"base_domain": "bellsouth.net"
|
"base_domain": null,
|
||||||
|
"name": null,
|
||||||
|
"type": null
|
||||||
},
|
},
|
||||||
"count": 2,
|
"count": 2,
|
||||||
"alignment": {
|
"alignment": {
|
||||||
@@ -74,7 +78,10 @@ of the report schema.
|
|||||||
"result": "pass"
|
"result": "pass"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
},
|
||||||
|
"normalized_timespan": false,
|
||||||
|
"interval_begin": "2012-04-28 00:00:00",
|
||||||
|
"interval_end": "2012-04-28 23:59:59"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
@@ -83,8 +90,10 @@ of the report schema.
|
|||||||
### CSV aggregate report
|
### CSV aggregate report
|
||||||
|
|
||||||
```text
|
```text
|
||||||
xml_schema,org_name,org_email,org_extra_contact_info,report_id,begin_date,end_date,errors,domain,adkim,aspf,p,sp,pct,fo,source_ip_address,source_country,source_reverse_dns,source_base_domain,count,spf_aligned,dkim_aligned,dmarc_aligned,disposition,policy_override_reasons,policy_override_comments,envelope_from,header_from,envelope_to,dkim_domains,dkim_selectors,dkim_results,spf_domains,spf_scopes,spf_results
|
xml_schema,org_name,org_email,org_extra_contact_info,report_id,begin_date,end_date,normalized_timespan,errors,domain,adkim,aspf,p,sp,pct,fo,source_ip_address,source_country,source_reverse_dns,source_base_domain,source_name,source_type,count,spf_aligned,dkim_aligned,dmarc_aligned,disposition,policy_override_reasons,policy_override_comments,envelope_from,header_from,envelope_to,dkim_domains,dkim_selectors,dkim_results,spf_domains,spf_scopes,spf_results
|
||||||
draft,acme.com,noreply-dmarc-support@acme.com,http://acme.com/dmarc/support,9391651994964116463,2012-04-27 20:00:00,2012-04-28 19:59:59,,example.com,r,r,none,none,100,0,72.150.241.94,US,adsl-72-150-241-94.shv.bellsouth.net,bellsouth.net,2,True,False,True,none,,,example.com,example.com,,example.com,none,fail,example.com,mfrom,pass
|
draft,acme.com,noreply-dmarc-support@acme.com,http://acme.com/dmarc/support,9391651994964116463,2012-04-28 00:00:00,2012-04-28 23:59:59,False,,example.com,r,r,none,none,100,0,72.150.241.94,US,,,,,2,True,False,True,none,,,example.com,example.com,,example.com,none,fail,example.com,mfrom,pass
|
||||||
|
draft,acme.com,noreply-dmarc-support@acme.com,http://acme.com/dmarc/support,9391651994964116463,2012-04-28 00:00:00,2012-04-28 23:59:59,False,,example.com,r,r,none,none,100,0,72.150.241.94,US,,,,,2,True,False,True,none,,,example.com,example.com,,example.com,none,fail,example.com,mfrom,pass
|
||||||
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Sample forensic report output
|
## Sample forensic report output
|
||||||
|
|||||||
@@ -4,47 +4,50 @@
|
|||||||
|
|
||||||
```text
|
```text
|
||||||
usage: parsedmarc [-h] [-c CONFIG_FILE] [--strip-attachment-payloads] [-o OUTPUT]
|
usage: parsedmarc [-h] [-c CONFIG_FILE] [--strip-attachment-payloads] [-o OUTPUT]
|
||||||
[--aggregate-json-filename AGGREGATE_JSON_FILENAME]
|
[--aggregate-json-filename AGGREGATE_JSON_FILENAME] [--forensic-json-filename FORENSIC_JSON_FILENAME]
|
||||||
[--forensic-json-filename FORENSIC_JSON_FILENAME]
|
[--smtp-tls-json-filename SMTP_TLS_JSON_FILENAME] [--aggregate-csv-filename AGGREGATE_CSV_FILENAME]
|
||||||
[--aggregate-csv-filename AGGREGATE_CSV_FILENAME]
|
[--forensic-csv-filename FORENSIC_CSV_FILENAME] [--smtp-tls-csv-filename SMTP_TLS_CSV_FILENAME]
|
||||||
[--forensic-csv-filename FORENSIC_CSV_FILENAME]
|
[-n NAMESERVERS [NAMESERVERS ...]] [-t DNS_TIMEOUT] [--offline] [-s] [-w] [--verbose] [--debug]
|
||||||
[-n NAMESERVERS [NAMESERVERS ...]] [-t DNS_TIMEOUT] [--offline]
|
[--log-file LOG_FILE] [--no-prettify-json] [-v]
|
||||||
[-s] [--verbose] [--debug] [--log-file LOG_FILE] [-v]
|
[file_path ...]
|
||||||
[file_path ...]
|
|
||||||
|
|
||||||
Parses DMARC reports
|
Parses DMARC reports
|
||||||
|
|
||||||
positional arguments:
|
positional arguments:
|
||||||
file_path one or more paths to aggregate or forensic report
|
file_path one or more paths to aggregate or forensic report files, emails, or mbox files'
|
||||||
files, emails, or mbox files'
|
|
||||||
|
|
||||||
optional arguments:
|
options:
|
||||||
-h, --help show this help message and exit
|
-h, --help show this help message and exit
|
||||||
-c CONFIG_FILE, --config-file CONFIG_FILE
|
-c CONFIG_FILE, --config-file CONFIG_FILE
|
||||||
a path to a configuration file (--silent implied)
|
a path to a configuration file (--silent implied)
|
||||||
--strip-attachment-payloads
|
--strip-attachment-payloads
|
||||||
remove attachment payloads from forensic report output
|
remove attachment payloads from forensic report output
|
||||||
-o OUTPUT, --output OUTPUT
|
-o OUTPUT, --output OUTPUT
|
||||||
write output files to the given directory
|
write output files to the given directory
|
||||||
--aggregate-json-filename AGGREGATE_JSON_FILENAME
|
--aggregate-json-filename AGGREGATE_JSON_FILENAME
|
||||||
filename for the aggregate JSON output file
|
filename for the aggregate JSON output file
|
||||||
--forensic-json-filename FORENSIC_JSON_FILENAME
|
--forensic-json-filename FORENSIC_JSON_FILENAME
|
||||||
filename for the forensic JSON output file
|
filename for the forensic JSON output file
|
||||||
--aggregate-csv-filename AGGREGATE_CSV_FILENAME
|
--smtp-tls-json-filename SMTP_TLS_JSON_FILENAME
|
||||||
filename for the aggregate CSV output file
|
filename for the SMTP TLS JSON output file
|
||||||
--forensic-csv-filename FORENSIC_CSV_FILENAME
|
--aggregate-csv-filename AGGREGATE_CSV_FILENAME
|
||||||
filename for the forensic CSV output file
|
filename for the aggregate CSV output file
|
||||||
-n NAMESERVERS [NAMESERVERS ...], --nameservers NAMESERVERS [NAMESERVERS ...]
|
--forensic-csv-filename FORENSIC_CSV_FILENAME
|
||||||
nameservers to query
|
filename for the forensic CSV output file
|
||||||
-t DNS_TIMEOUT, --dns_timeout DNS_TIMEOUT
|
--smtp-tls-csv-filename SMTP_TLS_CSV_FILENAME
|
||||||
number of seconds to wait for an answer from DNS
|
filename for the SMTP TLS CSV output file
|
||||||
(default: 2.0)
|
-n NAMESERVERS [NAMESERVERS ...], --nameservers NAMESERVERS [NAMESERVERS ...]
|
||||||
--offline do not make online queries for geolocation or DNS
|
nameservers to query
|
||||||
-s, --silent only print errors and warnings
|
-t DNS_TIMEOUT, --dns_timeout DNS_TIMEOUT
|
||||||
--verbose more verbose output
|
number of seconds to wait for an answer from DNS (default: 2.0)
|
||||||
--debug print debugging information
|
--offline do not make online queries for geolocation or DNS
|
||||||
--log-file LOG_FILE output logging to a file
|
-s, --silent only print errors
|
||||||
-v, --version show program's version number and exit
|
-w, --warnings print warnings in addition to errors
|
||||||
|
--verbose more verbose output
|
||||||
|
--debug print debugging information
|
||||||
|
--log-file LOG_FILE output logging to a file
|
||||||
|
--no-prettify-json output JSON in a single line without indentation
|
||||||
|
-v, --version show program's version number and exit
|
||||||
```
|
```
|
||||||
|
|
||||||
:::{note}
|
:::{note}
|
||||||
@@ -169,7 +172,7 @@ The full set of configuration options are:
|
|||||||
IDLE response or the number of seconds until the next
|
IDLE response or the number of seconds until the next
|
||||||
mail check (Default: `30`)
|
mail check (Default: `30`)
|
||||||
- `since` - str: Search for messages since certain time. (Examples: `5m|3h|2d|1w`)
|
- `since` - str: Search for messages since certain time. (Examples: `5m|3h|2d|1w`)
|
||||||
Acceptable units - {"m":"minutes", "h":"hours", "d":"days", "w":"weeks"}).
|
Acceptable units - {"m":"minutes", "h":"hours", "d":"days", "w":"weeks"}.
|
||||||
Defaults to `1d` if incorrect value is provided.
|
Defaults to `1d` if incorrect value is provided.
|
||||||
- `imap`
|
- `imap`
|
||||||
- `host` - str: The IMAP server hostname or IP address
|
- `host` - str: The IMAP server hostname or IP address
|
||||||
@@ -254,7 +257,7 @@ The full set of configuration options are:
|
|||||||
:::
|
:::
|
||||||
- `user` - str: Basic auth username
|
- `user` - str: Basic auth username
|
||||||
- `password` - str: Basic auth password
|
- `password` - str: Basic auth password
|
||||||
- `apiKey` - str: API key
|
- `api_key` - str: API key
|
||||||
- `ssl` - bool: Use an encrypted SSL/TLS connection
|
- `ssl` - bool: Use an encrypted SSL/TLS connection
|
||||||
(Default: `True`)
|
(Default: `True`)
|
||||||
- `timeout` - float: Timeout in seconds (Default: 60)
|
- `timeout` - float: Timeout in seconds (Default: 60)
|
||||||
@@ -277,7 +280,7 @@ The full set of configuration options are:
|
|||||||
:::
|
:::
|
||||||
- `user` - str: Basic auth username
|
- `user` - str: Basic auth username
|
||||||
- `password` - str: Basic auth password
|
- `password` - str: Basic auth password
|
||||||
- `apiKey` - str: API key
|
- `api_key` - str: API key
|
||||||
- `ssl` - bool: Use an encrypted SSL/TLS connection
|
- `ssl` - bool: Use an encrypted SSL/TLS connection
|
||||||
(Default: `True`)
|
(Default: `True`)
|
||||||
- `timeout` - float: Timeout in seconds (Default: 60)
|
- `timeout` - float: Timeout in seconds (Default: 60)
|
||||||
|
|||||||
File diff suppressed because one or more lines are too long
File diff suppressed because it is too large
Load Diff
@@ -10,7 +10,6 @@ from glob import glob
|
|||||||
import logging
|
import logging
|
||||||
import math
|
import math
|
||||||
import yaml
|
import yaml
|
||||||
from collections import OrderedDict
|
|
||||||
import json
|
import json
|
||||||
from ssl import CERT_NONE, create_default_context
|
from ssl import CERT_NONE, create_default_context
|
||||||
from multiprocessing import Pipe, Process
|
from multiprocessing import Pipe, Process
|
||||||
@@ -77,6 +76,7 @@ def cli_parse(
|
|||||||
always_use_local_files,
|
always_use_local_files,
|
||||||
reverse_dns_map_path,
|
reverse_dns_map_path,
|
||||||
reverse_dns_map_url,
|
reverse_dns_map_url,
|
||||||
|
normalize_timespan_threshold_hours,
|
||||||
conn,
|
conn,
|
||||||
):
|
):
|
||||||
"""Separated this function for multiprocessing"""
|
"""Separated this function for multiprocessing"""
|
||||||
@@ -91,6 +91,7 @@ def cli_parse(
|
|||||||
nameservers=nameservers,
|
nameservers=nameservers,
|
||||||
dns_timeout=dns_timeout,
|
dns_timeout=dns_timeout,
|
||||||
strip_attachment_payloads=sa,
|
strip_attachment_payloads=sa,
|
||||||
|
normalize_timespan_threshold_hours=normalize_timespan_threshold_hours,
|
||||||
)
|
)
|
||||||
conn.send([file_results, file_path])
|
conn.send([file_results, file_path])
|
||||||
except ParserError as error:
|
except ParserError as error:
|
||||||
@@ -591,7 +592,7 @@ def _main():
|
|||||||
elasticsearch_monthly_indexes=False,
|
elasticsearch_monthly_indexes=False,
|
||||||
elasticsearch_username=None,
|
elasticsearch_username=None,
|
||||||
elasticsearch_password=None,
|
elasticsearch_password=None,
|
||||||
elasticsearch_apiKey=None,
|
elasticsearch_api_key=None,
|
||||||
opensearch_hosts=None,
|
opensearch_hosts=None,
|
||||||
opensearch_timeout=60,
|
opensearch_timeout=60,
|
||||||
opensearch_number_of_shards=1,
|
opensearch_number_of_shards=1,
|
||||||
@@ -603,7 +604,7 @@ def _main():
|
|||||||
opensearch_monthly_indexes=False,
|
opensearch_monthly_indexes=False,
|
||||||
opensearch_username=None,
|
opensearch_username=None,
|
||||||
opensearch_password=None,
|
opensearch_password=None,
|
||||||
opensearch_apiKey=None,
|
opensearch_api_key=None,
|
||||||
kafka_hosts=None,
|
kafka_hosts=None,
|
||||||
kafka_username=None,
|
kafka_username=None,
|
||||||
kafka_password=None,
|
kafka_password=None,
|
||||||
@@ -659,6 +660,7 @@ def _main():
|
|||||||
webhook_forensic_url=None,
|
webhook_forensic_url=None,
|
||||||
webhook_smtp_tls_url=None,
|
webhook_smtp_tls_url=None,
|
||||||
webhook_timeout=60,
|
webhook_timeout=60,
|
||||||
|
normalize_timespan_threshold_hours=24.0,
|
||||||
)
|
)
|
||||||
args = arg_parser.parse_args()
|
args = arg_parser.parse_args()
|
||||||
|
|
||||||
@@ -674,8 +676,11 @@ def _main():
|
|||||||
if "general" in config.sections():
|
if "general" in config.sections():
|
||||||
general_config = config["general"]
|
general_config = config["general"]
|
||||||
if "silent" in general_config:
|
if "silent" in general_config:
|
||||||
if general_config["silent"].lower() == "false":
|
opts.silent = general_config.getboolean("silent")
|
||||||
opts.silent = False
|
if "normalize_timespan_threshold_hours" in general_config:
|
||||||
|
opts.normalize_timespan_threshold_hours = general_config.getfloat(
|
||||||
|
"normalize_timespan_threshold_hours"
|
||||||
|
)
|
||||||
if "index_prefix_domain_map" in general_config:
|
if "index_prefix_domain_map" in general_config:
|
||||||
with open(general_config["index_prefix_domain_map"]) as f:
|
with open(general_config["index_prefix_domain_map"]) as f:
|
||||||
index_prefix_domain_map = yaml.safe_load(f)
|
index_prefix_domain_map = yaml.safe_load(f)
|
||||||
@@ -723,11 +728,11 @@ def _main():
|
|||||||
)
|
)
|
||||||
exit(-1)
|
exit(-1)
|
||||||
if "save_aggregate" in general_config:
|
if "save_aggregate" in general_config:
|
||||||
opts.save_aggregate = general_config["save_aggregate"]
|
opts.save_aggregate = general_config.getboolean("save_aggregate")
|
||||||
if "save_forensic" in general_config:
|
if "save_forensic" in general_config:
|
||||||
opts.save_forensic = general_config["save_forensic"]
|
opts.save_forensic = general_config.getboolean("save_forensic")
|
||||||
if "save_smtp_tls" in general_config:
|
if "save_smtp_tls" in general_config:
|
||||||
opts.save_smtp_tls = general_config["save_smtp_tls"]
|
opts.save_smtp_tls = general_config.getboolean("save_smtp_tls")
|
||||||
if "debug" in general_config:
|
if "debug" in general_config:
|
||||||
opts.debug = general_config.getboolean("debug")
|
opts.debug = general_config.getboolean("debug")
|
||||||
if "verbose" in general_config:
|
if "verbose" in general_config:
|
||||||
@@ -798,8 +803,9 @@ def _main():
|
|||||||
if "ssl" in imap_config:
|
if "ssl" in imap_config:
|
||||||
opts.imap_ssl = imap_config.getboolean("ssl")
|
opts.imap_ssl = imap_config.getboolean("ssl")
|
||||||
if "skip_certificate_verification" in imap_config:
|
if "skip_certificate_verification" in imap_config:
|
||||||
imap_verify = imap_config.getboolean("skip_certificate_verification")
|
opts.imap_skip_certificate_verification = imap_config.getboolean(
|
||||||
opts.imap_skip_certificate_verification = imap_verify
|
"skip_certificate_verification"
|
||||||
|
)
|
||||||
if "user" in imap_config:
|
if "user" in imap_config:
|
||||||
opts.imap_user = imap_config["user"]
|
opts.imap_user = imap_config["user"]
|
||||||
else:
|
else:
|
||||||
@@ -975,8 +981,12 @@ def _main():
|
|||||||
opts.elasticsearch_username = elasticsearch_config["user"]
|
opts.elasticsearch_username = elasticsearch_config["user"]
|
||||||
if "password" in elasticsearch_config:
|
if "password" in elasticsearch_config:
|
||||||
opts.elasticsearch_password = elasticsearch_config["password"]
|
opts.elasticsearch_password = elasticsearch_config["password"]
|
||||||
|
# Until 8.20
|
||||||
if "apiKey" in elasticsearch_config:
|
if "apiKey" in elasticsearch_config:
|
||||||
opts.elasticsearch_apiKey = elasticsearch_config["apiKey"]
|
opts.elasticsearch_apiKey = elasticsearch_config["apiKey"]
|
||||||
|
# Since 8.20
|
||||||
|
if "api_key" in elasticsearch_config:
|
||||||
|
opts.elasticsearch_apiKey = elasticsearch_config["api_key"]
|
||||||
|
|
||||||
if "opensearch" in config:
|
if "opensearch" in config:
|
||||||
opensearch_config = config["opensearch"]
|
opensearch_config = config["opensearch"]
|
||||||
@@ -1011,8 +1021,12 @@ def _main():
|
|||||||
opts.opensearch_username = opensearch_config["user"]
|
opts.opensearch_username = opensearch_config["user"]
|
||||||
if "password" in opensearch_config:
|
if "password" in opensearch_config:
|
||||||
opts.opensearch_password = opensearch_config["password"]
|
opts.opensearch_password = opensearch_config["password"]
|
||||||
|
# Until 8.20
|
||||||
if "apiKey" in opensearch_config:
|
if "apiKey" in opensearch_config:
|
||||||
opts.opensearch_apiKey = opensearch_config["apiKey"]
|
opts.opensearch_apiKey = opensearch_config["apiKey"]
|
||||||
|
# Since 8.20
|
||||||
|
if "api_key" in opensearch_config:
|
||||||
|
opts.opensearch_apiKey = opensearch_config["api_key"]
|
||||||
|
|
||||||
if "splunk_hec" in config.sections():
|
if "splunk_hec" in config.sections():
|
||||||
hec_config = config["splunk_hec"]
|
hec_config = config["splunk_hec"]
|
||||||
@@ -1169,7 +1183,9 @@ def _main():
|
|||||||
)
|
)
|
||||||
opts.gmail_api_scopes = _str_to_list(opts.gmail_api_scopes)
|
opts.gmail_api_scopes = _str_to_list(opts.gmail_api_scopes)
|
||||||
if "oauth2_port" in gmail_api_config:
|
if "oauth2_port" in gmail_api_config:
|
||||||
opts.gmail_api_oauth2_port = gmail_api_config.get("oauth2_port", 8080)
|
opts.gmail_api_oauth2_port = gmail_api_config.getint(
|
||||||
|
"oauth2_port", 8080
|
||||||
|
)
|
||||||
|
|
||||||
if "maildir" in config.sections():
|
if "maildir" in config.sections():
|
||||||
maildir_api_config = config["maildir"]
|
maildir_api_config = config["maildir"]
|
||||||
@@ -1271,11 +1287,11 @@ def _main():
|
|||||||
es_smtp_tls_index = "{0}{1}".format(prefix, es_smtp_tls_index)
|
es_smtp_tls_index = "{0}{1}".format(prefix, es_smtp_tls_index)
|
||||||
elastic.set_hosts(
|
elastic.set_hosts(
|
||||||
opts.elasticsearch_hosts,
|
opts.elasticsearch_hosts,
|
||||||
opts.elasticsearch_ssl,
|
use_ssl=opts.elasticsearch_ssl,
|
||||||
opts.elasticsearch_ssl_cert_path,
|
ssl_cert_path=opts.elasticsearch_ssl_cert_path,
|
||||||
opts.elasticsearch_username,
|
username=opts.elasticsearch_username,
|
||||||
opts.elasticsearch_password,
|
password=opts.elasticsearch_password,
|
||||||
opts.elasticsearch_apiKey,
|
api_key=opts.elasticsearch_api_key,
|
||||||
timeout=opts.elasticsearch_timeout,
|
timeout=opts.elasticsearch_timeout,
|
||||||
)
|
)
|
||||||
elastic.migrate_indexes(
|
elastic.migrate_indexes(
|
||||||
@@ -1303,11 +1319,11 @@ def _main():
|
|||||||
os_smtp_tls_index = "{0}{1}".format(prefix, os_smtp_tls_index)
|
os_smtp_tls_index = "{0}{1}".format(prefix, os_smtp_tls_index)
|
||||||
opensearch.set_hosts(
|
opensearch.set_hosts(
|
||||||
opts.opensearch_hosts,
|
opts.opensearch_hosts,
|
||||||
opts.opensearch_ssl,
|
use_ssl=opts.opensearch_ssl,
|
||||||
opts.opensearch_ssl_cert_path,
|
ssl_cert_path=opts.opensearch_ssl_cert_path,
|
||||||
opts.opensearch_username,
|
username=opts.opensearch_username,
|
||||||
opts.opensearch_password,
|
password=opts.opensearch_password,
|
||||||
opts.opensearch_apiKey,
|
api_key=opts.opensearch_api_key,
|
||||||
timeout=opts.opensearch_timeout,
|
timeout=opts.opensearch_timeout,
|
||||||
)
|
)
|
||||||
opensearch.migrate_indexes(
|
opensearch.migrate_indexes(
|
||||||
@@ -1445,6 +1461,7 @@ def _main():
|
|||||||
opts.always_use_local_files,
|
opts.always_use_local_files,
|
||||||
opts.reverse_dns_map_path,
|
opts.reverse_dns_map_path,
|
||||||
opts.reverse_dns_map_url,
|
opts.reverse_dns_map_url,
|
||||||
|
opts.normalize_timespan_threshold_hours,
|
||||||
child_conn,
|
child_conn,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
@@ -1463,7 +1480,7 @@ def _main():
|
|||||||
pbar.update(counter - pbar.n)
|
pbar.update(counter - pbar.n)
|
||||||
|
|
||||||
for result in results:
|
for result in results:
|
||||||
if type(result[0]) is ParserError:
|
if isinstance(result[0], ParserError) or result[0] is None:
|
||||||
logger.error("Failed to parse {0} - {1}".format(result[1], result[0]))
|
logger.error("Failed to parse {0} - {1}".format(result[1], result[0]))
|
||||||
else:
|
else:
|
||||||
if result[0]["report_type"] == "aggregate":
|
if result[0]["report_type"] == "aggregate":
|
||||||
@@ -1495,6 +1512,7 @@ def _main():
|
|||||||
reverse_dns_map_path=opts.reverse_dns_map_path,
|
reverse_dns_map_path=opts.reverse_dns_map_path,
|
||||||
reverse_dns_map_url=opts.reverse_dns_map_url,
|
reverse_dns_map_url=opts.reverse_dns_map_url,
|
||||||
offline=opts.offline,
|
offline=opts.offline,
|
||||||
|
normalize_timespan_threshold_hours=opts.normalize_timespan_threshold_hours,
|
||||||
)
|
)
|
||||||
aggregate_reports += reports["aggregate_reports"]
|
aggregate_reports += reports["aggregate_reports"]
|
||||||
forensic_reports += reports["forensic_reports"]
|
forensic_reports += reports["forensic_reports"]
|
||||||
@@ -1513,7 +1531,7 @@ def _main():
|
|||||||
if opts.imap_skip_certificate_verification:
|
if opts.imap_skip_certificate_verification:
|
||||||
logger.debug("Skipping IMAP certificate verification")
|
logger.debug("Skipping IMAP certificate verification")
|
||||||
verify = False
|
verify = False
|
||||||
if opts.imap_ssl is False:
|
if not opts.imap_ssl:
|
||||||
ssl = False
|
ssl = False
|
||||||
|
|
||||||
mailbox_connection = IMAPConnection(
|
mailbox_connection = IMAPConnection(
|
||||||
@@ -1604,6 +1622,7 @@ def _main():
|
|||||||
test=opts.mailbox_test,
|
test=opts.mailbox_test,
|
||||||
strip_attachment_payloads=opts.strip_attachment_payloads,
|
strip_attachment_payloads=opts.strip_attachment_payloads,
|
||||||
since=opts.mailbox_since,
|
since=opts.mailbox_since,
|
||||||
|
normalize_timespan_threshold_hours=opts.normalize_timespan_threshold_hours,
|
||||||
)
|
)
|
||||||
|
|
||||||
aggregate_reports += reports["aggregate_reports"]
|
aggregate_reports += reports["aggregate_reports"]
|
||||||
@@ -1614,7 +1633,7 @@ def _main():
|
|||||||
logger.exception("Mailbox Error")
|
logger.exception("Mailbox Error")
|
||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
results = OrderedDict(
|
results = dict(
|
||||||
[
|
[
|
||||||
("aggregate_reports", aggregate_reports),
|
("aggregate_reports", aggregate_reports),
|
||||||
("forensic_reports", forensic_reports),
|
("forensic_reports", forensic_reports),
|
||||||
@@ -1666,6 +1685,7 @@ def _main():
|
|||||||
reverse_dns_map_path=opts.reverse_dns_map_path,
|
reverse_dns_map_path=opts.reverse_dns_map_path,
|
||||||
reverse_dns_map_url=opts.reverse_dns_map_url,
|
reverse_dns_map_url=opts.reverse_dns_map_url,
|
||||||
offline=opts.offline,
|
offline=opts.offline,
|
||||||
|
normalize_timespan_threshold_hours=opts.normalize_timespan_threshold_hours,
|
||||||
)
|
)
|
||||||
except FileExistsError as error:
|
except FileExistsError as error:
|
||||||
logger.error("{0}".format(error.__str__()))
|
logger.error("{0}".format(error.__str__()))
|
||||||
|
|||||||
@@ -1,2 +1,3 @@
|
|||||||
__version__ = "8.19.1"
|
__version__ = "9.0.5"
|
||||||
|
|
||||||
USER_AGENT = f"parsedmarc/{__version__}"
|
USER_AGENT = f"parsedmarc/{__version__}"
|
||||||
|
|||||||
@@ -1,6 +1,9 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
from collections import OrderedDict
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Optional, Union, Any
|
||||||
|
|
||||||
|
|
||||||
from elasticsearch_dsl.search import Q
|
from elasticsearch_dsl.search import Q
|
||||||
from elasticsearch_dsl import (
|
from elasticsearch_dsl import (
|
||||||
@@ -67,6 +70,8 @@ class _AggregateReportDoc(Document):
|
|||||||
date_range = Date()
|
date_range = Date()
|
||||||
date_begin = Date()
|
date_begin = Date()
|
||||||
date_end = Date()
|
date_end = Date()
|
||||||
|
normalized_timespan = Boolean()
|
||||||
|
original_timespan_seconds = Integer
|
||||||
errors = Text()
|
errors = Text()
|
||||||
published_policy = Object(_PublishedPolicy)
|
published_policy = Object(_PublishedPolicy)
|
||||||
source_ip_address = Ip()
|
source_ip_address = Ip()
|
||||||
@@ -87,15 +92,15 @@ class _AggregateReportDoc(Document):
|
|||||||
dkim_results = Nested(_DKIMResult)
|
dkim_results = Nested(_DKIMResult)
|
||||||
spf_results = Nested(_SPFResult)
|
spf_results = Nested(_SPFResult)
|
||||||
|
|
||||||
def add_policy_override(self, type_, comment):
|
def add_policy_override(self, type_: str, comment: str):
|
||||||
self.policy_overrides.append(_PolicyOverride(type=type_, comment=comment))
|
self.policy_overrides.append(_PolicyOverride(type=type_, comment=comment))
|
||||||
|
|
||||||
def add_dkim_result(self, domain, selector, result):
|
def add_dkim_result(self, domain: str, selector: str, result: _DKIMResult):
|
||||||
self.dkim_results.append(
|
self.dkim_results.append(
|
||||||
_DKIMResult(domain=domain, selector=selector, result=result)
|
_DKIMResult(domain=domain, selector=selector, result=result)
|
||||||
)
|
)
|
||||||
|
|
||||||
def add_spf_result(self, domain, scope, result):
|
def add_spf_result(self, domain: str, scope: str, result: _SPFResult):
|
||||||
self.spf_results.append(_SPFResult(domain=domain, scope=scope, result=result))
|
self.spf_results.append(_SPFResult(domain=domain, scope=scope, result=result))
|
||||||
|
|
||||||
def save(self, **kwargs):
|
def save(self, **kwargs):
|
||||||
@@ -131,21 +136,21 @@ class _ForensicSampleDoc(InnerDoc):
|
|||||||
body = Text()
|
body = Text()
|
||||||
attachments = Nested(_EmailAttachmentDoc)
|
attachments = Nested(_EmailAttachmentDoc)
|
||||||
|
|
||||||
def add_to(self, display_name, address):
|
def add_to(self, display_name: str, address: str):
|
||||||
self.to.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.to.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_reply_to(self, display_name, address):
|
def add_reply_to(self, display_name: str, address: str):
|
||||||
self.reply_to.append(
|
self.reply_to.append(
|
||||||
_EmailAddressDoc(display_name=display_name, address=address)
|
_EmailAddressDoc(display_name=display_name, address=address)
|
||||||
)
|
)
|
||||||
|
|
||||||
def add_cc(self, display_name, address):
|
def add_cc(self, display_name: str, address: str):
|
||||||
self.cc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.cc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_bcc(self, display_name, address):
|
def add_bcc(self, display_name: str, address: str):
|
||||||
self.bcc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.bcc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_attachment(self, filename, content_type, sha256):
|
def add_attachment(self, filename: str, content_type: str, sha256: str):
|
||||||
self.attachments.append(
|
self.attachments.append(
|
||||||
_EmailAttachmentDoc(
|
_EmailAttachmentDoc(
|
||||||
filename=filename, content_type=content_type, sha256=sha256
|
filename=filename, content_type=content_type, sha256=sha256
|
||||||
@@ -197,15 +202,15 @@ class _SMTPTLSPolicyDoc(InnerDoc):
|
|||||||
|
|
||||||
def add_failure_details(
|
def add_failure_details(
|
||||||
self,
|
self,
|
||||||
result_type,
|
result_type: Optional[str] = None,
|
||||||
ip_address,
|
ip_address: Optional[str] = None,
|
||||||
receiving_ip,
|
receiving_ip: Optional[str] = None,
|
||||||
receiving_mx_helo,
|
receiving_mx_helo: Optional[str] = None,
|
||||||
failed_session_count,
|
failed_session_count: Optional[int] = None,
|
||||||
sending_mta_ip=None,
|
sending_mta_ip: Optional[str] = None,
|
||||||
receiving_mx_hostname=None,
|
receiving_mx_hostname: Optional[str] = None,
|
||||||
additional_information_uri=None,
|
additional_information_uri: Optional[str] = None,
|
||||||
failure_reason_code=None,
|
failure_reason_code: Union[str, int, None] = None,
|
||||||
):
|
):
|
||||||
_details = _SMTPTLSFailureDetailsDoc(
|
_details = _SMTPTLSFailureDetailsDoc(
|
||||||
result_type=result_type,
|
result_type=result_type,
|
||||||
@@ -235,13 +240,14 @@ class _SMTPTLSReportDoc(Document):
|
|||||||
|
|
||||||
def add_policy(
|
def add_policy(
|
||||||
self,
|
self,
|
||||||
policy_type,
|
policy_type: str,
|
||||||
policy_domain,
|
policy_domain: str,
|
||||||
successful_session_count,
|
successful_session_count: int,
|
||||||
failed_session_count,
|
failed_session_count: int,
|
||||||
policy_string=None,
|
*,
|
||||||
mx_host_patterns=None,
|
policy_string: Optional[str] = None,
|
||||||
failure_details=None,
|
mx_host_patterns: Optional[list[str]] = None,
|
||||||
|
failure_details: Optional[str] = None,
|
||||||
):
|
):
|
||||||
self.policies.append(
|
self.policies.append(
|
||||||
policy_type=policy_type,
|
policy_type=policy_type,
|
||||||
@@ -259,24 +265,25 @@ class AlreadySaved(ValueError):
|
|||||||
|
|
||||||
|
|
||||||
def set_hosts(
|
def set_hosts(
|
||||||
hosts,
|
hosts: Union[str, list[str]],
|
||||||
use_ssl=False,
|
*,
|
||||||
ssl_cert_path=None,
|
use_ssl: Optional[bool] = False,
|
||||||
username=None,
|
ssl_cert_path: Optional[str] = None,
|
||||||
password=None,
|
username: Optional[str] = None,
|
||||||
apiKey=None,
|
password: Optional[str] = None,
|
||||||
timeout=60.0,
|
api_key: Optional[str] = None,
|
||||||
|
timeout: Optional[float] = 60.0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Sets the Elasticsearch hosts to use
|
Sets the Elasticsearch hosts to use
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
hosts (str): A single hostname or URL, or list of hostnames or URLs
|
hosts (Union[str, list[str]]): A single hostname or URL, or list of hostnames or URLs
|
||||||
use_ssl (bool): Use a HTTPS connection to the server
|
use_ssl (bool): Use an HTTPS connection to the server
|
||||||
ssl_cert_path (str): Path to the certificate chain
|
ssl_cert_path (str): Path to the certificate chain
|
||||||
username (str): The username to use for authentication
|
username (str): The username to use for authentication
|
||||||
password (str): The password to use for authentication
|
password (str): The password to use for authentication
|
||||||
apiKey (str): The Base64 encoded API key to use for authentication
|
api_key (str): The Base64 encoded API key to use for authentication
|
||||||
timeout (float): Timeout in seconds
|
timeout (float): Timeout in seconds
|
||||||
"""
|
"""
|
||||||
if not isinstance(hosts, list):
|
if not isinstance(hosts, list):
|
||||||
@@ -289,14 +296,14 @@ def set_hosts(
|
|||||||
conn_params["ca_certs"] = ssl_cert_path
|
conn_params["ca_certs"] = ssl_cert_path
|
||||||
else:
|
else:
|
||||||
conn_params["verify_certs"] = False
|
conn_params["verify_certs"] = False
|
||||||
if username:
|
if username and password:
|
||||||
conn_params["http_auth"] = username + ":" + password
|
conn_params["http_auth"] = username + ":" + password
|
||||||
if apiKey:
|
if api_key:
|
||||||
conn_params["api_key"] = apiKey
|
conn_params["api_key"] = api_key
|
||||||
connections.create_connection(**conn_params)
|
connections.create_connection(**conn_params)
|
||||||
|
|
||||||
|
|
||||||
def create_indexes(names, settings=None):
|
def create_indexes(names: list[str], settings: Optional[dict[str, Any]] = None):
|
||||||
"""
|
"""
|
||||||
Create Elasticsearch indexes
|
Create Elasticsearch indexes
|
||||||
|
|
||||||
@@ -319,7 +326,10 @@ def create_indexes(names, settings=None):
|
|||||||
raise ElasticsearchError("Elasticsearch error: {0}".format(e.__str__()))
|
raise ElasticsearchError("Elasticsearch error: {0}".format(e.__str__()))
|
||||||
|
|
||||||
|
|
||||||
def migrate_indexes(aggregate_indexes=None, forensic_indexes=None):
|
def migrate_indexes(
|
||||||
|
aggregate_indexes: Optional[list[str]] = None,
|
||||||
|
forensic_indexes: Optional[list[str]] = None,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Updates index mappings
|
Updates index mappings
|
||||||
|
|
||||||
@@ -366,18 +376,18 @@ def migrate_indexes(aggregate_indexes=None, forensic_indexes=None):
|
|||||||
|
|
||||||
|
|
||||||
def save_aggregate_report_to_elasticsearch(
|
def save_aggregate_report_to_elasticsearch(
|
||||||
aggregate_report,
|
aggregate_report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[str] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: Optional[int] = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: Optional[int] = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed DMARC aggregate report to Elasticsearch
|
Saves a parsed DMARC aggregate report to Elasticsearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
aggregate_report (OrderedDict): A parsed forensic report
|
aggregate_report (dict): A parsed forensic report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
||||||
@@ -395,15 +405,11 @@ def save_aggregate_report_to_elasticsearch(
|
|||||||
domain = aggregate_report["policy_published"]["domain"]
|
domain = aggregate_report["policy_published"]["domain"]
|
||||||
begin_date = human_timestamp_to_datetime(metadata["begin_date"], to_utc=True)
|
begin_date = human_timestamp_to_datetime(metadata["begin_date"], to_utc=True)
|
||||||
end_date = human_timestamp_to_datetime(metadata["end_date"], to_utc=True)
|
end_date = human_timestamp_to_datetime(metadata["end_date"], to_utc=True)
|
||||||
begin_date_human = begin_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
|
||||||
end_date_human = end_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
|
||||||
if monthly_indexes:
|
if monthly_indexes:
|
||||||
index_date = begin_date.strftime("%Y-%m")
|
index_date = begin_date.strftime("%Y-%m")
|
||||||
else:
|
else:
|
||||||
index_date = begin_date.strftime("%Y-%m-%d")
|
index_date = begin_date.strftime("%Y-%m-%d")
|
||||||
aggregate_report["begin_date"] = begin_date
|
|
||||||
aggregate_report["end_date"] = end_date
|
|
||||||
date_range = [aggregate_report["begin_date"], aggregate_report["end_date"]]
|
|
||||||
|
|
||||||
org_name_query = Q(dict(match_phrase=dict(org_name=org_name)))
|
org_name_query = Q(dict(match_phrase=dict(org_name=org_name)))
|
||||||
report_id_query = Q(dict(match_phrase=dict(report_id=report_id)))
|
report_id_query = Q(dict(match_phrase=dict(report_id=report_id)))
|
||||||
@@ -425,6 +431,9 @@ def save_aggregate_report_to_elasticsearch(
|
|||||||
try:
|
try:
|
||||||
existing = search.execute()
|
existing = search.execute()
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
|
begin_date_human = begin_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
||||||
|
end_date_human = end_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
||||||
|
|
||||||
raise ElasticsearchError(
|
raise ElasticsearchError(
|
||||||
"Elasticsearch's search for existing report \
|
"Elasticsearch's search for existing report \
|
||||||
error: {}".format(error_.__str__())
|
error: {}".format(error_.__str__())
|
||||||
@@ -450,6 +459,17 @@ def save_aggregate_report_to_elasticsearch(
|
|||||||
)
|
)
|
||||||
|
|
||||||
for record in aggregate_report["records"]:
|
for record in aggregate_report["records"]:
|
||||||
|
begin_date = human_timestamp_to_datetime(record["interval_begin"], to_utc=True)
|
||||||
|
end_date = human_timestamp_to_datetime(record["interval_end"], to_utc=True)
|
||||||
|
normalized_timespan = record["normalized_timespan"]
|
||||||
|
|
||||||
|
if monthly_indexes:
|
||||||
|
index_date = begin_date.strftime("%Y-%m")
|
||||||
|
else:
|
||||||
|
index_date = begin_date.strftime("%Y-%m-%d")
|
||||||
|
aggregate_report["begin_date"] = begin_date
|
||||||
|
aggregate_report["end_date"] = end_date
|
||||||
|
date_range = [aggregate_report["begin_date"], aggregate_report["end_date"]]
|
||||||
agg_doc = _AggregateReportDoc(
|
agg_doc = _AggregateReportDoc(
|
||||||
xml_schema=aggregate_report["xml_schema"],
|
xml_schema=aggregate_report["xml_schema"],
|
||||||
org_name=metadata["org_name"],
|
org_name=metadata["org_name"],
|
||||||
@@ -457,8 +477,9 @@ def save_aggregate_report_to_elasticsearch(
|
|||||||
org_extra_contact_info=metadata["org_extra_contact_info"],
|
org_extra_contact_info=metadata["org_extra_contact_info"],
|
||||||
report_id=metadata["report_id"],
|
report_id=metadata["report_id"],
|
||||||
date_range=date_range,
|
date_range=date_range,
|
||||||
date_begin=aggregate_report["begin_date"],
|
date_begin=begin_date,
|
||||||
date_end=aggregate_report["end_date"],
|
date_end=end_date,
|
||||||
|
normalized_timespan=normalized_timespan,
|
||||||
errors=metadata["errors"],
|
errors=metadata["errors"],
|
||||||
published_policy=published_policy,
|
published_policy=published_policy,
|
||||||
source_ip_address=record["source"]["ip_address"],
|
source_ip_address=record["source"]["ip_address"],
|
||||||
@@ -517,18 +538,18 @@ def save_aggregate_report_to_elasticsearch(
|
|||||||
|
|
||||||
|
|
||||||
def save_forensic_report_to_elasticsearch(
|
def save_forensic_report_to_elasticsearch(
|
||||||
forensic_report,
|
forensic_report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[Any] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: int = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: int = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed DMARC forensic report to Elasticsearch
|
Saves a parsed DMARC forensic report to Elasticsearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
forensic_report (OrderedDict): A parsed forensic report
|
forensic_report (dict): A parsed forensic report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily
|
monthly_indexes (bool): Use monthly indexes instead of daily
|
||||||
@@ -548,7 +569,7 @@ def save_forensic_report_to_elasticsearch(
|
|||||||
sample_date = forensic_report["parsed_sample"]["date"]
|
sample_date = forensic_report["parsed_sample"]["date"]
|
||||||
sample_date = human_timestamp_to_datetime(sample_date)
|
sample_date = human_timestamp_to_datetime(sample_date)
|
||||||
original_headers = forensic_report["parsed_sample"]["headers"]
|
original_headers = forensic_report["parsed_sample"]["headers"]
|
||||||
headers = OrderedDict()
|
headers = dict()
|
||||||
for original_header in original_headers:
|
for original_header in original_headers:
|
||||||
headers[original_header.lower()] = original_headers[original_header]
|
headers[original_header.lower()] = original_headers[original_header]
|
||||||
|
|
||||||
@@ -684,18 +705,18 @@ def save_forensic_report_to_elasticsearch(
|
|||||||
|
|
||||||
|
|
||||||
def save_smtp_tls_report_to_elasticsearch(
|
def save_smtp_tls_report_to_elasticsearch(
|
||||||
report,
|
report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[str] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: Optional[int] = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: Optional[int] = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed SMTP TLS report to Elasticsearch
|
Saves a parsed SMTP TLS report to Elasticsearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
report (OrderedDict): A parsed SMTP TLS report
|
report (dict): A parsed SMTP TLS report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
||||||
@@ -781,7 +802,7 @@ def save_smtp_tls_report_to_elasticsearch(
|
|||||||
policy_doc = _SMTPTLSPolicyDoc(
|
policy_doc = _SMTPTLSPolicyDoc(
|
||||||
policy_domain=policy["policy_domain"],
|
policy_domain=policy["policy_domain"],
|
||||||
policy_type=policy["policy_type"],
|
policy_type=policy["policy_type"],
|
||||||
succesful_session_count=policy["successful_session_count"],
|
successful_session_count=policy["successful_session_count"],
|
||||||
failed_session_count=policy["failed_session_count"],
|
failed_session_count=policy["failed_session_count"],
|
||||||
policy_string=policy_strings,
|
policy_string=policy_strings,
|
||||||
mx_host_patterns=mx_host_patterns,
|
mx_host_patterns=mx_host_patterns,
|
||||||
|
|||||||
@@ -1,5 +1,9 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import logging.handlers
|
import logging.handlers
|
||||||
import json
|
import json
|
||||||
@@ -48,7 +52,9 @@ class GelfClient(object):
|
|||||||
)
|
)
|
||||||
self.logger.addHandler(self.handler)
|
self.logger.addHandler(self.handler)
|
||||||
|
|
||||||
def save_aggregate_report_to_gelf(self, aggregate_reports):
|
def save_aggregate_report_to_gelf(
|
||||||
|
self, aggregate_reports: list[dict[str, Any]]
|
||||||
|
):
|
||||||
rows = parsed_aggregate_reports_to_csv_rows(aggregate_reports)
|
rows = parsed_aggregate_reports_to_csv_rows(aggregate_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
log_context_data.parsedmarc = row
|
log_context_data.parsedmarc = row
|
||||||
@@ -56,12 +62,14 @@ class GelfClient(object):
|
|||||||
|
|
||||||
log_context_data.parsedmarc = None
|
log_context_data.parsedmarc = None
|
||||||
|
|
||||||
def save_forensic_report_to_gelf(self, forensic_reports):
|
def save_forensic_report_to_gelf(
|
||||||
|
self, forensic_reports: list[dict[str, Any]]
|
||||||
|
):
|
||||||
rows = parsed_forensic_reports_to_csv_rows(forensic_reports)
|
rows = parsed_forensic_reports_to_csv_rows(forensic_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
self.logger.info(json.dumps(row))
|
self.logger.info(json.dumps(row))
|
||||||
|
|
||||||
def save_smtp_tls_report_to_gelf(self, smtp_tls_reports):
|
def save_smtp_tls_report_to_gelf(self, smtp_tls_reports: dict[str, Any]):
|
||||||
rows = parsed_smtp_tls_reports_to_csv_rows(smtp_tls_reports)
|
rows = parsed_smtp_tls_reports_to_csv_rows(smtp_tls_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
self.logger.info(json.dumps(row))
|
self.logger.info(json.dumps(row))
|
||||||
|
|||||||
@@ -1,11 +1,15 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any, Optional, Union
|
||||||
|
from ssl import SSLContext
|
||||||
|
|
||||||
import json
|
import json
|
||||||
from ssl import create_default_context
|
from ssl import create_default_context
|
||||||
|
|
||||||
from kafka import KafkaProducer
|
from kafka import KafkaProducer
|
||||||
from kafka.errors import NoBrokersAvailable, UnknownTopicOrPartitionError
|
from kafka.errors import NoBrokersAvailable, UnknownTopicOrPartitionError
|
||||||
from collections import OrderedDict
|
|
||||||
from parsedmarc.utils import human_timestamp_to_datetime
|
from parsedmarc.utils import human_timestamp_to_datetime
|
||||||
|
|
||||||
from parsedmarc import __version__
|
from parsedmarc import __version__
|
||||||
@@ -18,7 +22,13 @@ class KafkaError(RuntimeError):
|
|||||||
|
|
||||||
class KafkaClient(object):
|
class KafkaClient(object):
|
||||||
def __init__(
|
def __init__(
|
||||||
self, kafka_hosts, ssl=False, username=None, password=None, ssl_context=None
|
self,
|
||||||
|
kafka_hosts: list[str],
|
||||||
|
*,
|
||||||
|
ssl: Optional[bool] = False,
|
||||||
|
username: Optional[str] = None,
|
||||||
|
password: Optional[str] = None,
|
||||||
|
ssl_context: Optional[SSLContext] = None,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Initializes the Kafka client
|
Initializes the Kafka client
|
||||||
@@ -28,7 +38,7 @@ class KafkaClient(object):
|
|||||||
ssl (bool): Use a SSL/TLS connection
|
ssl (bool): Use a SSL/TLS connection
|
||||||
username (str): An optional username
|
username (str): An optional username
|
||||||
password (str): An optional password
|
password (str): An optional password
|
||||||
ssl_context: SSL context options
|
ssl_context (SSLContext): SSL context options
|
||||||
|
|
||||||
Notes:
|
Notes:
|
||||||
``use_ssl=True`` is implied when a username or password are
|
``use_ssl=True`` is implied when a username or password are
|
||||||
@@ -55,7 +65,7 @@ class KafkaClient(object):
|
|||||||
raise KafkaError("No Kafka brokers available")
|
raise KafkaError("No Kafka brokers available")
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def strip_metadata(report):
|
def strip_metadata(report: dict[str, Any]):
|
||||||
"""
|
"""
|
||||||
Duplicates org_name, org_email and report_id into JSON root
|
Duplicates org_name, org_email and report_id into JSON root
|
||||||
and removes report_metadata key to bring it more inline
|
and removes report_metadata key to bring it more inline
|
||||||
@@ -69,7 +79,7 @@ class KafkaClient(object):
|
|||||||
return report
|
return report
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def generate_daterange(report):
|
def generate_date_range(report: dict[str, Any]):
|
||||||
"""
|
"""
|
||||||
Creates a date_range timestamp with format YYYY-MM-DD-T-HH:MM:SS
|
Creates a date_range timestamp with format YYYY-MM-DD-T-HH:MM:SS
|
||||||
based on begin and end dates for easier parsing in Kibana.
|
based on begin and end dates for easier parsing in Kibana.
|
||||||
@@ -86,7 +96,11 @@ class KafkaClient(object):
|
|||||||
logger.debug("date_range is {}".format(date_range))
|
logger.debug("date_range is {}".format(date_range))
|
||||||
return date_range
|
return date_range
|
||||||
|
|
||||||
def save_aggregate_reports_to_kafka(self, aggregate_reports, aggregate_topic):
|
def save_aggregate_reports_to_kafka(
|
||||||
|
self,
|
||||||
|
aggregate_reports: Union[dict[str, Any], list[dict[str, Any]]],
|
||||||
|
aggregate_topic: str,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves aggregate DMARC reports to Kafka
|
Saves aggregate DMARC reports to Kafka
|
||||||
|
|
||||||
@@ -96,16 +110,14 @@ class KafkaClient(object):
|
|||||||
aggregate_topic (str): The name of the Kafka topic
|
aggregate_topic (str): The name of the Kafka topic
|
||||||
|
|
||||||
"""
|
"""
|
||||||
if isinstance(aggregate_reports, dict) or isinstance(
|
if isinstance(aggregate_reports, dict):
|
||||||
aggregate_reports, OrderedDict
|
|
||||||
):
|
|
||||||
aggregate_reports = [aggregate_reports]
|
aggregate_reports = [aggregate_reports]
|
||||||
|
|
||||||
if len(aggregate_reports) < 1:
|
if len(aggregate_reports) < 1:
|
||||||
return
|
return
|
||||||
|
|
||||||
for report in aggregate_reports:
|
for report in aggregate_reports:
|
||||||
report["date_range"] = self.generate_daterange(report)
|
report["date_range"] = self.generate_date_range(report)
|
||||||
report = self.strip_metadata(report)
|
report = self.strip_metadata(report)
|
||||||
|
|
||||||
for slice in report["records"]:
|
for slice in report["records"]:
|
||||||
@@ -129,7 +141,11 @@ class KafkaClient(object):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise KafkaError("Kafka error: {0}".format(e.__str__()))
|
raise KafkaError("Kafka error: {0}".format(e.__str__()))
|
||||||
|
|
||||||
def save_forensic_reports_to_kafka(self, forensic_reports, forensic_topic):
|
def save_forensic_reports_to_kafka(
|
||||||
|
self,
|
||||||
|
forensic_reports: Union[dict[str, Any], list[dict[str, Any]]],
|
||||||
|
forensic_topic: str,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves forensic DMARC reports to Kafka, sends individual
|
Saves forensic DMARC reports to Kafka, sends individual
|
||||||
records (slices) since Kafka requires messages to be <= 1MB
|
records (slices) since Kafka requires messages to be <= 1MB
|
||||||
@@ -159,7 +175,11 @@ class KafkaClient(object):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise KafkaError("Kafka error: {0}".format(e.__str__()))
|
raise KafkaError("Kafka error: {0}".format(e.__str__()))
|
||||||
|
|
||||||
def save_smtp_tls_reports_to_kafka(self, smtp_tls_reports, smtp_tls_topic):
|
def save_smtp_tls_reports_to_kafka(
|
||||||
|
self,
|
||||||
|
smtp_tls_reports: Union[list[dict[str, Any]], dict[str, Any]],
|
||||||
|
smtp_tls_topic: str,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves SMTP TLS reports to Kafka, sends individual
|
Saves SMTP TLS reports to Kafka, sends individual
|
||||||
records (slices) since Kafka requires messages to be <= 1MB
|
records (slices) since Kafka requires messages to be <= 1MB
|
||||||
|
|||||||
@@ -1,4 +1,9 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any
|
||||||
|
|
||||||
from parsedmarc.log import logger
|
from parsedmarc.log import logger
|
||||||
from azure.core.exceptions import HttpResponseError
|
from azure.core.exceptions import HttpResponseError
|
||||||
from azure.identity import ClientSecretCredential
|
from azure.identity import ClientSecretCredential
|
||||||
@@ -102,7 +107,12 @@ class LogAnalyticsClient(object):
|
|||||||
"Invalid configuration. " + "One or more required settings are missing."
|
"Invalid configuration. " + "One or more required settings are missing."
|
||||||
)
|
)
|
||||||
|
|
||||||
def publish_json(self, results, logs_client: LogsIngestionClient, dcr_stream: str):
|
def publish_json(
|
||||||
|
self,
|
||||||
|
results,
|
||||||
|
logs_client: LogsIngestionClient,
|
||||||
|
dcr_stream: str,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Background function to publish given
|
Background function to publish given
|
||||||
DMARC report to specific Data Collection Rule.
|
DMARC report to specific Data Collection Rule.
|
||||||
@@ -121,7 +131,11 @@ class LogAnalyticsClient(object):
|
|||||||
raise LogAnalyticsException("Upload failed: {error}".format(error=e))
|
raise LogAnalyticsException("Upload failed: {error}".format(error=e))
|
||||||
|
|
||||||
def publish_results(
|
def publish_results(
|
||||||
self, results, save_aggregate: bool, save_forensic: bool, save_smtp_tls: bool
|
self,
|
||||||
|
results: dict[str, dict[str, Any]],
|
||||||
|
save_aggregate: bool,
|
||||||
|
save_forensic: bool,
|
||||||
|
save_smtp_tls: bool,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Function to publish DMARC and/or SMTP TLS reports to Log Analytics
|
Function to publish DMARC and/or SMTP TLS reports to Log Analytics
|
||||||
|
|||||||
@@ -1,3 +1,7 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
from base64 import urlsafe_b64decode
|
from base64 import urlsafe_b64decode
|
||||||
from functools import lru_cache
|
from functools import lru_cache
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
@@ -152,3 +156,4 @@ class GmailConnection(MailboxConnection):
|
|||||||
for label in labels:
|
for label in labels:
|
||||||
if label_name == label["id"] or label_name == label["name"]:
|
if label_name == label["id"] or label_name == label["name"]:
|
||||||
return label["id"]
|
return label["id"]
|
||||||
|
return ""
|
||||||
|
|||||||
@@ -1,3 +1,7 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
from enum import Enum
|
from enum import Enum
|
||||||
from functools import lru_cache
|
from functools import lru_cache
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|||||||
@@ -1,3 +1,9 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Optional
|
||||||
|
|
||||||
from time import sleep
|
from time import sleep
|
||||||
|
|
||||||
from imapclient.exceptions import IMAPClientError
|
from imapclient.exceptions import IMAPClientError
|
||||||
@@ -11,14 +17,15 @@ from parsedmarc.mail.mailbox_connection import MailboxConnection
|
|||||||
class IMAPConnection(MailboxConnection):
|
class IMAPConnection(MailboxConnection):
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
host=None,
|
host: Optional[str] = None,
|
||||||
user=None,
|
*,
|
||||||
password=None,
|
user: Optional[str] = None,
|
||||||
port=None,
|
password: Optional[str] = None,
|
||||||
ssl=True,
|
port: Optional[str] = None,
|
||||||
verify=True,
|
ssl: Optional[bool] = True,
|
||||||
timeout=30,
|
verify: Optional[bool] = True,
|
||||||
max_retries=4,
|
timeout: Optional[int] = 30,
|
||||||
|
max_retries: Optional[int] = 4,
|
||||||
):
|
):
|
||||||
self._username = user
|
self._username = user
|
||||||
self._password = password
|
self._password = password
|
||||||
@@ -45,13 +52,13 @@ class IMAPConnection(MailboxConnection):
|
|||||||
else:
|
else:
|
||||||
return self._client.search()
|
return self._client.search()
|
||||||
|
|
||||||
def fetch_message(self, message_id):
|
def fetch_message(self, message_id: int):
|
||||||
return self._client.fetch_message(message_id, parse=False)
|
return self._client.fetch_message(message_id, parse=False)
|
||||||
|
|
||||||
def delete_message(self, message_id: str):
|
def delete_message(self, message_id: int):
|
||||||
self._client.delete_messages([message_id])
|
self._client.delete_messages([message_id])
|
||||||
|
|
||||||
def move_message(self, message_id: str, folder_name: str):
|
def move_message(self, message_id: int, folder_name: str):
|
||||||
self._client.move_messages([message_id], folder_name)
|
self._client.move_messages([message_id], folder_name)
|
||||||
|
|
||||||
def keepalive(self):
|
def keepalive(self):
|
||||||
|
|||||||
@@ -1,5 +1,8 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
from abc import ABC
|
from abc import ABC
|
||||||
from typing import List
|
|
||||||
|
|
||||||
|
|
||||||
class MailboxConnection(ABC):
|
class MailboxConnection(ABC):
|
||||||
@@ -10,7 +13,7 @@ class MailboxConnection(ABC):
|
|||||||
def create_folder(self, folder_name: str):
|
def create_folder(self, folder_name: str):
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
def fetch_messages(self, reports_folder: str, **kwargs) -> List[str]:
|
def fetch_messages(self, reports_folder: str, **kwargs) -> list[str]:
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
def fetch_message(self, message_id) -> str:
|
def fetch_message(self, message_id) -> str:
|
||||||
|
|||||||
@@ -1,3 +1,9 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Optional
|
||||||
|
|
||||||
from time import sleep
|
from time import sleep
|
||||||
|
|
||||||
from parsedmarc.log import logger
|
from parsedmarc.log import logger
|
||||||
@@ -9,8 +15,8 @@ import os
|
|||||||
class MaildirConnection(MailboxConnection):
|
class MaildirConnection(MailboxConnection):
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
maildir_path=None,
|
maildir_path: Optional[bool] = None,
|
||||||
maildir_create=False,
|
maildir_create: Optional[bool] = False,
|
||||||
):
|
):
|
||||||
self._maildir_path = maildir_path
|
self._maildir_path = maildir_path
|
||||||
self._maildir_create = maildir_create
|
self._maildir_create = maildir_create
|
||||||
@@ -36,7 +42,7 @@ class MaildirConnection(MailboxConnection):
|
|||||||
def fetch_messages(self, reports_folder: str, **kwargs):
|
def fetch_messages(self, reports_folder: str, **kwargs):
|
||||||
return self._client.keys()
|
return self._client.keys()
|
||||||
|
|
||||||
def fetch_message(self, message_id):
|
def fetch_message(self, message_id: str):
|
||||||
return self._client.get(message_id).as_string()
|
return self._client.get(message_id).as_string()
|
||||||
|
|
||||||
def delete_message(self, message_id: str):
|
def delete_message(self, message_id: str):
|
||||||
|
|||||||
@@ -1,6 +1,9 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
from collections import OrderedDict
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Optional, Union, Any
|
||||||
|
|
||||||
|
|
||||||
from opensearchpy import (
|
from opensearchpy import (
|
||||||
Q,
|
Q,
|
||||||
@@ -67,6 +70,8 @@ class _AggregateReportDoc(Document):
|
|||||||
date_range = Date()
|
date_range = Date()
|
||||||
date_begin = Date()
|
date_begin = Date()
|
||||||
date_end = Date()
|
date_end = Date()
|
||||||
|
normalized_timespan = Boolean()
|
||||||
|
original_timespan_seconds = Integer
|
||||||
errors = Text()
|
errors = Text()
|
||||||
published_policy = Object(_PublishedPolicy)
|
published_policy = Object(_PublishedPolicy)
|
||||||
source_ip_address = Ip()
|
source_ip_address = Ip()
|
||||||
@@ -87,15 +92,15 @@ class _AggregateReportDoc(Document):
|
|||||||
dkim_results = Nested(_DKIMResult)
|
dkim_results = Nested(_DKIMResult)
|
||||||
spf_results = Nested(_SPFResult)
|
spf_results = Nested(_SPFResult)
|
||||||
|
|
||||||
def add_policy_override(self, type_, comment):
|
def add_policy_override(self, type_: str, comment: str):
|
||||||
self.policy_overrides.append(_PolicyOverride(type=type_, comment=comment))
|
self.policy_overrides.append(_PolicyOverride(type=type_, comment=comment))
|
||||||
|
|
||||||
def add_dkim_result(self, domain, selector, result):
|
def add_dkim_result(self, domain: str, selector: str, result: _DKIMResult):
|
||||||
self.dkim_results.append(
|
self.dkim_results.append(
|
||||||
_DKIMResult(domain=domain, selector=selector, result=result)
|
_DKIMResult(domain=domain, selector=selector, result=result)
|
||||||
)
|
)
|
||||||
|
|
||||||
def add_spf_result(self, domain, scope, result):
|
def add_spf_result(self, domain: str, scope: str, result: _SPFResult):
|
||||||
self.spf_results.append(_SPFResult(domain=domain, scope=scope, result=result))
|
self.spf_results.append(_SPFResult(domain=domain, scope=scope, result=result))
|
||||||
|
|
||||||
def save(self, **kwargs):
|
def save(self, **kwargs):
|
||||||
@@ -131,21 +136,21 @@ class _ForensicSampleDoc(InnerDoc):
|
|||||||
body = Text()
|
body = Text()
|
||||||
attachments = Nested(_EmailAttachmentDoc)
|
attachments = Nested(_EmailAttachmentDoc)
|
||||||
|
|
||||||
def add_to(self, display_name, address):
|
def add_to(self, display_name: str, address: str):
|
||||||
self.to.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.to.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_reply_to(self, display_name, address):
|
def add_reply_to(self, display_name: str, address: str):
|
||||||
self.reply_to.append(
|
self.reply_to.append(
|
||||||
_EmailAddressDoc(display_name=display_name, address=address)
|
_EmailAddressDoc(display_name=display_name, address=address)
|
||||||
)
|
)
|
||||||
|
|
||||||
def add_cc(self, display_name, address):
|
def add_cc(self, display_name: str, address: str):
|
||||||
self.cc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.cc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_bcc(self, display_name, address):
|
def add_bcc(self, display_name: str, address: str):
|
||||||
self.bcc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
self.bcc.append(_EmailAddressDoc(display_name=display_name, address=address))
|
||||||
|
|
||||||
def add_attachment(self, filename, content_type, sha256):
|
def add_attachment(self, filename: str, content_type: str, sha256: str):
|
||||||
self.attachments.append(
|
self.attachments.append(
|
||||||
_EmailAttachmentDoc(
|
_EmailAttachmentDoc(
|
||||||
filename=filename, content_type=content_type, sha256=sha256
|
filename=filename, content_type=content_type, sha256=sha256
|
||||||
@@ -197,15 +202,15 @@ class _SMTPTLSPolicyDoc(InnerDoc):
|
|||||||
|
|
||||||
def add_failure_details(
|
def add_failure_details(
|
||||||
self,
|
self,
|
||||||
result_type,
|
result_type: Optional[str] = None,
|
||||||
ip_address,
|
ip_address: Optional[str] = None,
|
||||||
receiving_ip,
|
receiving_ip: Optional[str] = None,
|
||||||
receiving_mx_helo,
|
receiving_mx_helo: Optional[str] = None,
|
||||||
failed_session_count,
|
failed_session_count: Optional[int] = None,
|
||||||
sending_mta_ip=None,
|
sending_mta_ip: Optional[str] = None,
|
||||||
receiving_mx_hostname=None,
|
receiving_mx_hostname: Optional[str] = None,
|
||||||
additional_information_uri=None,
|
additional_information_uri: Optional[str] = None,
|
||||||
failure_reason_code=None,
|
failure_reason_code: Union[str, int, None] = None,
|
||||||
):
|
):
|
||||||
_details = _SMTPTLSFailureDetailsDoc(
|
_details = _SMTPTLSFailureDetailsDoc(
|
||||||
result_type=result_type,
|
result_type=result_type,
|
||||||
@@ -235,13 +240,14 @@ class _SMTPTLSReportDoc(Document):
|
|||||||
|
|
||||||
def add_policy(
|
def add_policy(
|
||||||
self,
|
self,
|
||||||
policy_type,
|
policy_type: str,
|
||||||
policy_domain,
|
policy_domain: str,
|
||||||
successful_session_count,
|
successful_session_count: int,
|
||||||
failed_session_count,
|
failed_session_count: int,
|
||||||
policy_string=None,
|
*,
|
||||||
mx_host_patterns=None,
|
policy_string: Optional[str] = None,
|
||||||
failure_details=None,
|
mx_host_patterns: Optional[list[str]] = None,
|
||||||
|
failure_details: Optional[str] = None,
|
||||||
):
|
):
|
||||||
self.policies.append(
|
self.policies.append(
|
||||||
policy_type=policy_type,
|
policy_type=policy_type,
|
||||||
@@ -259,24 +265,25 @@ class AlreadySaved(ValueError):
|
|||||||
|
|
||||||
|
|
||||||
def set_hosts(
|
def set_hosts(
|
||||||
hosts,
|
hosts: Union[str, list[str]],
|
||||||
use_ssl=False,
|
*,
|
||||||
ssl_cert_path=None,
|
use_ssl: Optional[bool] = False,
|
||||||
username=None,
|
ssl_cert_path: Optional[str] = None,
|
||||||
password=None,
|
username: Optional[str] = None,
|
||||||
apiKey=None,
|
password: Optional[str] = None,
|
||||||
timeout=60.0,
|
api_key: Optional[str] = None,
|
||||||
|
timeout: Optional[float] = 60.0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Sets the OpenSearch hosts to use
|
Sets the OpenSearch hosts to use
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
hosts (str|list): A hostname or URL, or list of hostnames or URLs
|
hosts (str|list[str]): A single hostname or URL, or list of hostnames or URLs
|
||||||
use_ssl (bool): Use an HTTPS connection to the server
|
use_ssl (bool): Use an HTTPS connection to the server
|
||||||
ssl_cert_path (str): Path to the certificate chain
|
ssl_cert_path (str): Path to the certificate chain
|
||||||
username (str): The username to use for authentication
|
username (str): The username to use for authentication
|
||||||
password (str): The password to use for authentication
|
password (str): The password to use for authentication
|
||||||
apiKey (str): The Base64 encoded API key to use for authentication
|
api_key (str): The Base64 encoded API key to use for authentication
|
||||||
timeout (float): Timeout in seconds
|
timeout (float): Timeout in seconds
|
||||||
"""
|
"""
|
||||||
if not isinstance(hosts, list):
|
if not isinstance(hosts, list):
|
||||||
@@ -289,14 +296,14 @@ def set_hosts(
|
|||||||
conn_params["ca_certs"] = ssl_cert_path
|
conn_params["ca_certs"] = ssl_cert_path
|
||||||
else:
|
else:
|
||||||
conn_params["verify_certs"] = False
|
conn_params["verify_certs"] = False
|
||||||
if username:
|
if username and password:
|
||||||
conn_params["http_auth"] = username + ":" + password
|
conn_params["http_auth"] = username + ":" + password
|
||||||
if apiKey:
|
if api_key:
|
||||||
conn_params["api_key"] = apiKey
|
conn_params["api_key"] = api_key
|
||||||
connections.create_connection(**conn_params)
|
connections.create_connection(**conn_params)
|
||||||
|
|
||||||
|
|
||||||
def create_indexes(names, settings=None):
|
def create_indexes(names: list[str], settings: Optional[dict[str, Any]] = None):
|
||||||
"""
|
"""
|
||||||
Create OpenSearch indexes
|
Create OpenSearch indexes
|
||||||
|
|
||||||
@@ -319,7 +326,10 @@ def create_indexes(names, settings=None):
|
|||||||
raise OpenSearchError("OpenSearch error: {0}".format(e.__str__()))
|
raise OpenSearchError("OpenSearch error: {0}".format(e.__str__()))
|
||||||
|
|
||||||
|
|
||||||
def migrate_indexes(aggregate_indexes=None, forensic_indexes=None):
|
def migrate_indexes(
|
||||||
|
aggregate_indexes: Optional[list[str]] = None,
|
||||||
|
forensic_indexes: Optional[list[str]] = None,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Updates index mappings
|
Updates index mappings
|
||||||
|
|
||||||
@@ -366,18 +376,18 @@ def migrate_indexes(aggregate_indexes=None, forensic_indexes=None):
|
|||||||
|
|
||||||
|
|
||||||
def save_aggregate_report_to_opensearch(
|
def save_aggregate_report_to_opensearch(
|
||||||
aggregate_report,
|
aggregate_report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[str] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: Optional[int] = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: Optional[int] = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed DMARC aggregate report to OpenSearch
|
Saves a parsed DMARC aggregate report to OpenSearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
aggregate_report (OrderedDict): A parsed forensic report
|
aggregate_report (dict): A parsed forensic report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
||||||
@@ -395,15 +405,11 @@ def save_aggregate_report_to_opensearch(
|
|||||||
domain = aggregate_report["policy_published"]["domain"]
|
domain = aggregate_report["policy_published"]["domain"]
|
||||||
begin_date = human_timestamp_to_datetime(metadata["begin_date"], to_utc=True)
|
begin_date = human_timestamp_to_datetime(metadata["begin_date"], to_utc=True)
|
||||||
end_date = human_timestamp_to_datetime(metadata["end_date"], to_utc=True)
|
end_date = human_timestamp_to_datetime(metadata["end_date"], to_utc=True)
|
||||||
begin_date_human = begin_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
|
||||||
end_date_human = end_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
|
||||||
if monthly_indexes:
|
if monthly_indexes:
|
||||||
index_date = begin_date.strftime("%Y-%m")
|
index_date = begin_date.strftime("%Y-%m")
|
||||||
else:
|
else:
|
||||||
index_date = begin_date.strftime("%Y-%m-%d")
|
index_date = begin_date.strftime("%Y-%m-%d")
|
||||||
aggregate_report["begin_date"] = begin_date
|
|
||||||
aggregate_report["end_date"] = end_date
|
|
||||||
date_range = [aggregate_report["begin_date"], aggregate_report["end_date"]]
|
|
||||||
|
|
||||||
org_name_query = Q(dict(match_phrase=dict(org_name=org_name)))
|
org_name_query = Q(dict(match_phrase=dict(org_name=org_name)))
|
||||||
report_id_query = Q(dict(match_phrase=dict(report_id=report_id)))
|
report_id_query = Q(dict(match_phrase=dict(report_id=report_id)))
|
||||||
@@ -425,6 +431,9 @@ def save_aggregate_report_to_opensearch(
|
|||||||
try:
|
try:
|
||||||
existing = search.execute()
|
existing = search.execute()
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
|
begin_date_human = begin_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
||||||
|
end_date_human = end_date.strftime("%Y-%m-%d %H:%M:%SZ")
|
||||||
|
|
||||||
raise OpenSearchError(
|
raise OpenSearchError(
|
||||||
"OpenSearch's search for existing report \
|
"OpenSearch's search for existing report \
|
||||||
error: {}".format(error_.__str__())
|
error: {}".format(error_.__str__())
|
||||||
@@ -450,6 +459,17 @@ def save_aggregate_report_to_opensearch(
|
|||||||
)
|
)
|
||||||
|
|
||||||
for record in aggregate_report["records"]:
|
for record in aggregate_report["records"]:
|
||||||
|
begin_date = human_timestamp_to_datetime(record["interval_begin"], to_utc=True)
|
||||||
|
end_date = human_timestamp_to_datetime(record["interval_end"], to_utc=True)
|
||||||
|
normalized_timespan = record["normalized_timespan"]
|
||||||
|
|
||||||
|
if monthly_indexes:
|
||||||
|
index_date = begin_date.strftime("%Y-%m")
|
||||||
|
else:
|
||||||
|
index_date = begin_date.strftime("%Y-%m-%d")
|
||||||
|
aggregate_report["begin_date"] = begin_date
|
||||||
|
aggregate_report["end_date"] = end_date
|
||||||
|
date_range = [aggregate_report["begin_date"], aggregate_report["end_date"]]
|
||||||
agg_doc = _AggregateReportDoc(
|
agg_doc = _AggregateReportDoc(
|
||||||
xml_schema=aggregate_report["xml_schema"],
|
xml_schema=aggregate_report["xml_schema"],
|
||||||
org_name=metadata["org_name"],
|
org_name=metadata["org_name"],
|
||||||
@@ -457,8 +477,9 @@ def save_aggregate_report_to_opensearch(
|
|||||||
org_extra_contact_info=metadata["org_extra_contact_info"],
|
org_extra_contact_info=metadata["org_extra_contact_info"],
|
||||||
report_id=metadata["report_id"],
|
report_id=metadata["report_id"],
|
||||||
date_range=date_range,
|
date_range=date_range,
|
||||||
date_begin=aggregate_report["begin_date"],
|
date_begin=begin_date,
|
||||||
date_end=aggregate_report["end_date"],
|
date_end=end_date,
|
||||||
|
normalized_timespan=normalized_timespan,
|
||||||
errors=metadata["errors"],
|
errors=metadata["errors"],
|
||||||
published_policy=published_policy,
|
published_policy=published_policy,
|
||||||
source_ip_address=record["source"]["ip_address"],
|
source_ip_address=record["source"]["ip_address"],
|
||||||
@@ -517,18 +538,18 @@ def save_aggregate_report_to_opensearch(
|
|||||||
|
|
||||||
|
|
||||||
def save_forensic_report_to_opensearch(
|
def save_forensic_report_to_opensearch(
|
||||||
forensic_report,
|
forensic_report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[str] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: int = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: int = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed DMARC forensic report to OpenSearch
|
Saves a parsed DMARC forensic report to OpenSearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
forensic_report (OrderedDict): A parsed forensic report
|
forensic_report (dict): A parsed forensic report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily
|
monthly_indexes (bool): Use monthly indexes instead of daily
|
||||||
@@ -548,7 +569,7 @@ def save_forensic_report_to_opensearch(
|
|||||||
sample_date = forensic_report["parsed_sample"]["date"]
|
sample_date = forensic_report["parsed_sample"]["date"]
|
||||||
sample_date = human_timestamp_to_datetime(sample_date)
|
sample_date = human_timestamp_to_datetime(sample_date)
|
||||||
original_headers = forensic_report["parsed_sample"]["headers"]
|
original_headers = forensic_report["parsed_sample"]["headers"]
|
||||||
headers = OrderedDict()
|
headers = dict()
|
||||||
for original_header in original_headers:
|
for original_header in original_headers:
|
||||||
headers[original_header.lower()] = original_headers[original_header]
|
headers[original_header.lower()] = original_headers[original_header]
|
||||||
|
|
||||||
@@ -684,18 +705,18 @@ def save_forensic_report_to_opensearch(
|
|||||||
|
|
||||||
|
|
||||||
def save_smtp_tls_report_to_opensearch(
|
def save_smtp_tls_report_to_opensearch(
|
||||||
report,
|
report: dict[str, Any],
|
||||||
index_suffix=None,
|
index_suffix: Optional[str] = None,
|
||||||
index_prefix=None,
|
index_prefix: Optional[str] = None,
|
||||||
monthly_indexes=False,
|
monthly_indexes: Optional[bool] = False,
|
||||||
number_of_shards=1,
|
number_of_shards: Optional[int] = 1,
|
||||||
number_of_replicas=0,
|
number_of_replicas: Optional[int] = 0,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Saves a parsed SMTP TLS report to OpenSearch
|
Saves a parsed SMTP TLS report to OpenSearch
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
report (OrderedDict): A parsed SMTP TLS report
|
report (dict): A parsed SMTP TLS report
|
||||||
index_suffix (str): The suffix of the name of the index to save to
|
index_suffix (str): The suffix of the name of the index to save to
|
||||||
index_prefix (str): The prefix of the name of the index to save to
|
index_prefix (str): The prefix of the name of the index to save to
|
||||||
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
monthly_indexes (bool): Use monthly indexes instead of daily indexes
|
||||||
@@ -705,7 +726,7 @@ def save_smtp_tls_report_to_opensearch(
|
|||||||
Raises:
|
Raises:
|
||||||
AlreadySaved
|
AlreadySaved
|
||||||
"""
|
"""
|
||||||
logger.info("Saving aggregate report to OpenSearch")
|
logger.info("Saving SMTP TLS report to OpenSearch")
|
||||||
org_name = report["organization_name"]
|
org_name = report["organization_name"]
|
||||||
report_id = report["report_id"]
|
report_id = report["report_id"]
|
||||||
begin_date = human_timestamp_to_datetime(report["begin_date"], to_utc=True)
|
begin_date = human_timestamp_to_datetime(report["begin_date"], to_utc=True)
|
||||||
@@ -781,7 +802,7 @@ def save_smtp_tls_report_to_opensearch(
|
|||||||
policy_doc = _SMTPTLSPolicyDoc(
|
policy_doc = _SMTPTLSPolicyDoc(
|
||||||
policy_domain=policy["policy_domain"],
|
policy_domain=policy["policy_domain"],
|
||||||
policy_type=policy["policy_type"],
|
policy_type=policy["policy_type"],
|
||||||
succesful_session_count=policy["successful_session_count"],
|
successful_session_count=policy["successful_session_count"],
|
||||||
failed_session_count=policy["failed_session_count"],
|
failed_session_count=policy["failed_session_count"],
|
||||||
policy_string=policy_strings,
|
policy_string=policy_strings,
|
||||||
mx_host_patterns=mx_host_patterns,
|
mx_host_patterns=mx_host_patterns,
|
||||||
|
|||||||
@@ -132,6 +132,7 @@ asu-vei.ru,ASU-VEI,Industrial
|
|||||||
atextelecom.com.br,ATEX Telecom,ISP
|
atextelecom.com.br,ATEX Telecom,ISP
|
||||||
atmailcloud.com,atmail,Email Provider
|
atmailcloud.com,atmail,Email Provider
|
||||||
ats.ca,ATS Healthcare,Healthcare
|
ats.ca,ATS Healthcare,Healthcare
|
||||||
|
att.net,AT&T,ISP
|
||||||
atw.ne.jp,ATW,Web Host
|
atw.ne.jp,ATW,Web Host
|
||||||
au-net.ne.jp,KDDI,ISP
|
au-net.ne.jp,KDDI,ISP
|
||||||
au.com,au,ISP
|
au.com,au,ISP
|
||||||
@@ -242,6 +243,7 @@ carandainet.com.br,CN Internet,ISP
|
|||||||
cardhealth.com,Cardinal Health,Healthcare
|
cardhealth.com,Cardinal Health,Healthcare
|
||||||
cardinal.com,Cardinal Health,Healthcare
|
cardinal.com,Cardinal Health,Healthcare
|
||||||
cardinalhealth.com,Cardinal Health,Healthcare
|
cardinalhealth.com,Cardinal Health,Healthcare
|
||||||
|
cardinalscriptnet.com,Cardinal Health,Healthcare
|
||||||
carecentrix.com,CareCentrix,Healthcare
|
carecentrix.com,CareCentrix,Healthcare
|
||||||
carleton.edu,Carlton College,Education
|
carleton.edu,Carlton College,Education
|
||||||
carrierzone.com,carrierzone,Email Security
|
carrierzone.com,carrierzone,Email Security
|
||||||
@@ -697,6 +699,7 @@ hdsupply-email.com,HD Supply,Retail
|
|||||||
healthall.com,UC Health,Healthcare
|
healthall.com,UC Health,Healthcare
|
||||||
healthcaresupplypros.com,Healthcare Supply Pros,Healthcare
|
healthcaresupplypros.com,Healthcare Supply Pros,Healthcare
|
||||||
healthproductsforyou.com,Health Products For You,Healthcare
|
healthproductsforyou.com,Health Products For You,Healthcare
|
||||||
|
healthtouch.com,Cardinal Health,Healthcare
|
||||||
helloserver6.com,1st Source Web,Marketing
|
helloserver6.com,1st Source Web,Marketing
|
||||||
helpforcb.com,InterServer,Web Host
|
helpforcb.com,InterServer,Web Host
|
||||||
helpscout.net,Help Scout,SaaS
|
helpscout.net,Help Scout,SaaS
|
||||||
@@ -753,6 +756,8 @@ hostwindsdns.com,Hostwinds,Web Host
|
|||||||
hotnet.net.il,Hot Net Internet Services,ISP
|
hotnet.net.il,Hot Net Internet Services,ISP
|
||||||
hp.com,HP,Technology
|
hp.com,HP,Technology
|
||||||
hringdu.is,Hringdu,ISP
|
hringdu.is,Hringdu,ISP
|
||||||
|
hslda.net,Home School Legal Defense Association (HSLDA),Education
|
||||||
|
hslda.org,Home School Legal Defense Association (HSLDA),Education
|
||||||
hspherefilter.com,"DynamicNet, Inc. (DNI)",Web Host
|
hspherefilter.com,"DynamicNet, Inc. (DNI)",Web Host
|
||||||
htc.net,HTC,ISP
|
htc.net,HTC,ISP
|
||||||
htmlservices.it,HTMLServices.it,MSP
|
htmlservices.it,HTMLServices.it,MSP
|
||||||
@@ -763,6 +768,7 @@ hughston.com,Hughston Clinic,Healthcare
|
|||||||
hvvc.us,Hivelocity,Web Host
|
hvvc.us,Hivelocity,Web Host
|
||||||
i2ts.ne.jp,i2ts,Web Host
|
i2ts.ne.jp,i2ts,Web Host
|
||||||
i4i.com,i4i,Technology
|
i4i.com,i4i,Technology
|
||||||
|
ibindley.com,Cardinal Health,Healthcare
|
||||||
ice.co.cr,Grupo ICE,Industrial
|
ice.co.cr,Grupo ICE,Industrial
|
||||||
icehosting.nl,IceHosting,Web Host
|
icehosting.nl,IceHosting,Web Host
|
||||||
icewarpcloud.in,IceWrap,Email Provider
|
icewarpcloud.in,IceWrap,Email Provider
|
||||||
@@ -832,6 +838,7 @@ ip-5-196-151.eu,OVH,Web Host
|
|||||||
ip-51-161-36.net,OVH,Web Host
|
ip-51-161-36.net,OVH,Web Host
|
||||||
ip-51-195-53.eu,OVH,Web Host
|
ip-51-195-53.eu,OVH,Web Host
|
||||||
ip-51-254-53.eu,OVH,Web Host
|
ip-51-254-53.eu,OVH,Web Host
|
||||||
|
ip-51-38-67.eu,OVH,Web Host
|
||||||
ip-51-77-42.eu,OVH,Web Host
|
ip-51-77-42.eu,OVH,Web Host
|
||||||
ip-51-83-140.eu,OVH,Web Host
|
ip-51-83-140.eu,OVH,Web Host
|
||||||
ip-51-89-240.eu,OVH,Web Host
|
ip-51-89-240.eu,OVH,Web Host
|
||||||
@@ -1217,6 +1224,7 @@ nettoday.co.th,Net Today,Web Host
|
|||||||
netventure.pl,Netventure,MSP
|
netventure.pl,Netventure,MSP
|
||||||
netvigator.com,HKT,ISP
|
netvigator.com,HKT,ISP
|
||||||
netvision.net.il,013 Netvision,ISP
|
netvision.net.il,013 Netvision,ISP
|
||||||
|
network-tech.com,Network Technologies International (NTI),SaaS
|
||||||
network.kz,network.kz,ISP
|
network.kz,network.kz,ISP
|
||||||
network80.com,Network80,Web Host
|
network80.com,Network80,Web Host
|
||||||
neubox.net,Neubox,Web Host
|
neubox.net,Neubox,Web Host
|
||||||
|
|||||||
|
@@ -13,8 +13,6 @@ def _main():
|
|||||||
|
|
||||||
csv_headers = ["source_name", "message_count"]
|
csv_headers = ["source_name", "message_count"]
|
||||||
|
|
||||||
output_rows = []
|
|
||||||
|
|
||||||
known_unknown_domains = []
|
known_unknown_domains = []
|
||||||
psl_overrides = []
|
psl_overrides = []
|
||||||
known_domains = []
|
known_domains = []
|
||||||
|
|||||||
@@ -1,5 +1,9 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import boto3
|
import boto3
|
||||||
|
|
||||||
@@ -8,16 +12,16 @@ from parsedmarc.utils import human_timestamp_to_datetime
|
|||||||
|
|
||||||
|
|
||||||
class S3Client(object):
|
class S3Client(object):
|
||||||
"""A client for a Amazon S3"""
|
"""A client for interacting with Amazon S3"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
bucket_name,
|
bucket_name: str,
|
||||||
bucket_path,
|
bucket_path: str,
|
||||||
region_name,
|
region_name: str,
|
||||||
endpoint_url,
|
endpoint_url: str,
|
||||||
access_key_id,
|
access_key_id: str,
|
||||||
secret_access_key,
|
secret_access_key: str,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Initializes the S3Client
|
Initializes the S3Client
|
||||||
@@ -47,18 +51,18 @@ class S3Client(object):
|
|||||||
aws_access_key_id=access_key_id,
|
aws_access_key_id=access_key_id,
|
||||||
aws_secret_access_key=secret_access_key,
|
aws_secret_access_key=secret_access_key,
|
||||||
)
|
)
|
||||||
self.bucket = self.s3.Bucket(self.bucket_name)
|
self.bucket: Any = self.s3.Bucket(self.bucket_name)
|
||||||
|
|
||||||
def save_aggregate_report_to_s3(self, report):
|
def save_aggregate_report_to_s3(self, report: dict[str, Any]):
|
||||||
self.save_report_to_s3(report, "aggregate")
|
self.save_report_to_s3(report, "aggregate")
|
||||||
|
|
||||||
def save_forensic_report_to_s3(self, report):
|
def save_forensic_report_to_s3(self, report: dict[str, Any]):
|
||||||
self.save_report_to_s3(report, "forensic")
|
self.save_report_to_s3(report, "forensic")
|
||||||
|
|
||||||
def save_smtp_tls_report_to_s3(self, report):
|
def save_smtp_tls_report_to_s3(self, report: dict[str, Any]):
|
||||||
self.save_report_to_s3(report, "smtp_tls")
|
self.save_report_to_s3(report, "smtp_tls")
|
||||||
|
|
||||||
def save_report_to_s3(self, report, report_type):
|
def save_report_to_s3(self, report: dict[str, Any], report_type: str):
|
||||||
if report_type == "smtp_tls":
|
if report_type == "smtp_tls":
|
||||||
report_date = report["begin_date"]
|
report_date = report["begin_date"]
|
||||||
report_id = report["report_id"]
|
report_id = report["report_id"]
|
||||||
|
|||||||
@@ -1,3 +1,10 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any, Union
|
||||||
|
|
||||||
|
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
import socket
|
import socket
|
||||||
import json
|
import json
|
||||||
@@ -23,7 +30,13 @@ class HECClient(object):
|
|||||||
# http://docs.splunk.com/Documentation/Splunk/latest/RESTREF/RESTinput#services.2Fcollector
|
# http://docs.splunk.com/Documentation/Splunk/latest/RESTREF/RESTinput#services.2Fcollector
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self, url, access_token, index, source="parsedmarc", verify=True, timeout=60
|
self,
|
||||||
|
url: str,
|
||||||
|
access_token: str,
|
||||||
|
index: str,
|
||||||
|
source: str = "parsedmarc",
|
||||||
|
verify=True,
|
||||||
|
timeout=60,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Initializes the HECClient
|
Initializes the HECClient
|
||||||
@@ -37,9 +50,9 @@ class HECClient(object):
|
|||||||
timeout (float): Number of seconds to wait for the server to send
|
timeout (float): Number of seconds to wait for the server to send
|
||||||
data before giving up
|
data before giving up
|
||||||
"""
|
"""
|
||||||
url = urlparse(url)
|
parsed_url = urlparse(url)
|
||||||
self.url = "{0}://{1}/services/collector/event/1.0".format(
|
self.url = "{0}://{1}/services/collector/event/1.0".format(
|
||||||
url.scheme, url.netloc
|
parsed_url.scheme, parsed_url.netloc
|
||||||
)
|
)
|
||||||
self.access_token = access_token.lstrip("Splunk ")
|
self.access_token = access_token.lstrip("Splunk ")
|
||||||
self.index = index
|
self.index = index
|
||||||
@@ -48,14 +61,19 @@ class HECClient(object):
|
|||||||
self.session = requests.Session()
|
self.session = requests.Session()
|
||||||
self.timeout = timeout
|
self.timeout = timeout
|
||||||
self.session.verify = verify
|
self.session.verify = verify
|
||||||
self._common_data = dict(host=self.host, source=self.source, index=self.index)
|
self._common_data: dict[str, Union[str, int, float, dict]] = dict(
|
||||||
|
host=self.host, source=self.source, index=self.index
|
||||||
|
)
|
||||||
|
|
||||||
self.session.headers = {
|
self.session.headers = {
|
||||||
"User-Agent": USER_AGENT,
|
"User-Agent": USER_AGENT,
|
||||||
"Authorization": "Splunk {0}".format(self.access_token),
|
"Authorization": "Splunk {0}".format(self.access_token),
|
||||||
}
|
}
|
||||||
|
|
||||||
def save_aggregate_reports_to_splunk(self, aggregate_reports):
|
def save_aggregate_reports_to_splunk(
|
||||||
|
self,
|
||||||
|
aggregate_reports: Union[list[dict[str, Any]], dict[str, Any]],
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves aggregate DMARC reports to Splunk
|
Saves aggregate DMARC reports to Splunk
|
||||||
|
|
||||||
@@ -75,9 +93,12 @@ class HECClient(object):
|
|||||||
json_str = ""
|
json_str = ""
|
||||||
for report in aggregate_reports:
|
for report in aggregate_reports:
|
||||||
for record in report["records"]:
|
for record in report["records"]:
|
||||||
new_report = dict()
|
new_report: dict[str, Union[str, int, float, dict]] = dict()
|
||||||
for metadata in report["report_metadata"]:
|
for metadata in report["report_metadata"]:
|
||||||
new_report[metadata] = report["report_metadata"][metadata]
|
new_report[metadata] = report["report_metadata"][metadata]
|
||||||
|
new_report["interval_begin"] = record["interval_begin"]
|
||||||
|
new_report["interval_end"] = record["interval_end"]
|
||||||
|
new_report["normalized_timespan"] = record["normalized_timespan"]
|
||||||
new_report["published_policy"] = report["policy_published"]
|
new_report["published_policy"] = report["policy_published"]
|
||||||
new_report["source_ip_address"] = record["source"]["ip_address"]
|
new_report["source_ip_address"] = record["source"]["ip_address"]
|
||||||
new_report["source_country"] = record["source"]["country"]
|
new_report["source_country"] = record["source"]["country"]
|
||||||
@@ -98,7 +119,9 @@ class HECClient(object):
|
|||||||
new_report["spf_results"] = record["auth_results"]["spf"]
|
new_report["spf_results"] = record["auth_results"]["spf"]
|
||||||
|
|
||||||
data["sourcetype"] = "dmarc:aggregate"
|
data["sourcetype"] = "dmarc:aggregate"
|
||||||
timestamp = human_timestamp_to_unix_timestamp(new_report["begin_date"])
|
timestamp = human_timestamp_to_unix_timestamp(
|
||||||
|
new_report["interval_begin"]
|
||||||
|
)
|
||||||
data["time"] = timestamp
|
data["time"] = timestamp
|
||||||
data["event"] = new_report.copy()
|
data["event"] = new_report.copy()
|
||||||
json_str += "{0}\n".format(json.dumps(data))
|
json_str += "{0}\n".format(json.dumps(data))
|
||||||
@@ -113,7 +136,10 @@ class HECClient(object):
|
|||||||
if response["code"] != 0:
|
if response["code"] != 0:
|
||||||
raise SplunkError(response["text"])
|
raise SplunkError(response["text"])
|
||||||
|
|
||||||
def save_forensic_reports_to_splunk(self, forensic_reports):
|
def save_forensic_reports_to_splunk(
|
||||||
|
self,
|
||||||
|
forensic_reports: Union[list[dict[str, Any]], dict[str, Any]],
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves forensic DMARC reports to Splunk
|
Saves forensic DMARC reports to Splunk
|
||||||
|
|
||||||
@@ -147,7 +173,9 @@ class HECClient(object):
|
|||||||
if response["code"] != 0:
|
if response["code"] != 0:
|
||||||
raise SplunkError(response["text"])
|
raise SplunkError(response["text"])
|
||||||
|
|
||||||
def save_smtp_tls_reports_to_splunk(self, reports):
|
def save_smtp_tls_reports_to_splunk(
|
||||||
|
self, reports: Union[list[dict[str, Any]], dict[str, Any]]
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Saves aggregate DMARC reports to Splunk
|
Saves aggregate DMARC reports to Splunk
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,14 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import logging.handlers
|
import logging.handlers
|
||||||
|
|
||||||
|
from typing import Any
|
||||||
|
|
||||||
|
|
||||||
import json
|
import json
|
||||||
|
|
||||||
from parsedmarc import (
|
from parsedmarc import (
|
||||||
@@ -14,7 +21,7 @@ from parsedmarc import (
|
|||||||
class SyslogClient(object):
|
class SyslogClient(object):
|
||||||
"""A client for Syslog"""
|
"""A client for Syslog"""
|
||||||
|
|
||||||
def __init__(self, server_name, server_port):
|
def __init__(self, server_name: str, server_port: int):
|
||||||
"""
|
"""
|
||||||
Initializes the SyslogClient
|
Initializes the SyslogClient
|
||||||
Args:
|
Args:
|
||||||
@@ -28,17 +35,23 @@ class SyslogClient(object):
|
|||||||
log_handler = logging.handlers.SysLogHandler(address=(server_name, server_port))
|
log_handler = logging.handlers.SysLogHandler(address=(server_name, server_port))
|
||||||
self.logger.addHandler(log_handler)
|
self.logger.addHandler(log_handler)
|
||||||
|
|
||||||
def save_aggregate_report_to_syslog(self, aggregate_reports):
|
def save_aggregate_report_to_syslog(
|
||||||
|
self, aggregate_reports: list[dict[str, Any]]
|
||||||
|
):
|
||||||
rows = parsed_aggregate_reports_to_csv_rows(aggregate_reports)
|
rows = parsed_aggregate_reports_to_csv_rows(aggregate_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
self.logger.info(json.dumps(row))
|
self.logger.info(json.dumps(row))
|
||||||
|
|
||||||
def save_forensic_report_to_syslog(self, forensic_reports):
|
def save_forensic_report_to_syslog(
|
||||||
|
self, forensic_reports: list[dict[str, Any]]
|
||||||
|
):
|
||||||
rows = parsed_forensic_reports_to_csv_rows(forensic_reports)
|
rows = parsed_forensic_reports_to_csv_rows(forensic_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
self.logger.info(json.dumps(row))
|
self.logger.info(json.dumps(row))
|
||||||
|
|
||||||
def save_smtp_tls_report_to_syslog(self, smtp_tls_reports):
|
def save_smtp_tls_report_to_syslog(
|
||||||
|
self, smtp_tls_reports: list[dict[str, Any]]
|
||||||
|
):
|
||||||
rows = parsed_smtp_tls_reports_to_csv_rows(smtp_tls_reports)
|
rows = parsed_smtp_tls_reports_to_csv_rows(smtp_tls_reports)
|
||||||
for row in rows:
|
for row in rows:
|
||||||
self.logger.info(json.dumps(row))
|
self.logger.info(json.dumps(row))
|
||||||
|
|||||||
@@ -1,11 +1,17 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
"""Utility functions that might be useful for other projects"""
|
"""Utility functions that might be useful for other projects"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Optional, Union, TypedDict, Any
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from datetime import timezone
|
from datetime import timezone
|
||||||
from datetime import timedelta
|
from datetime import timedelta
|
||||||
from collections import OrderedDict
|
from expiringdict import ExpiringDict
|
||||||
import tempfile
|
import tempfile
|
||||||
import subprocess
|
import subprocess
|
||||||
import shutil
|
import shutil
|
||||||
@@ -60,12 +66,20 @@ class DownloadError(RuntimeError):
|
|||||||
"""Raised when an error occurs when downloading a file"""
|
"""Raised when an error occurs when downloading a file"""
|
||||||
|
|
||||||
|
|
||||||
def decode_base64(data):
|
class EmailAddress(TypedDict):
|
||||||
|
"""Parsed email address information"""
|
||||||
|
display_name: Optional[str]
|
||||||
|
address: str
|
||||||
|
local: Optional[str]
|
||||||
|
domain: Optional[str]
|
||||||
|
|
||||||
|
|
||||||
|
def decode_base64(data: str) -> bytes:
|
||||||
"""
|
"""
|
||||||
Decodes a base64 string, with padding being optional
|
Decodes a base64 string, with padding being optional
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
data: A base64 encoded string
|
data (str): A base64 encoded string
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
bytes: The decoded bytes
|
bytes: The decoded bytes
|
||||||
@@ -78,7 +92,7 @@ def decode_base64(data):
|
|||||||
return base64.b64decode(data)
|
return base64.b64decode(data)
|
||||||
|
|
||||||
|
|
||||||
def get_base_domain(domain):
|
def get_base_domain(domain: str) -> str:
|
||||||
"""
|
"""
|
||||||
Gets the base domain name for the given domain
|
Gets the base domain name for the given domain
|
||||||
|
|
||||||
@@ -102,7 +116,14 @@ def get_base_domain(domain):
|
|||||||
return publicsuffix
|
return publicsuffix
|
||||||
|
|
||||||
|
|
||||||
def query_dns(domain, record_type, cache=None, nameservers=None, timeout=2.0):
|
def query_dns(
|
||||||
|
domain: str,
|
||||||
|
record_type: str,
|
||||||
|
*,
|
||||||
|
cache: Optional[ExpiringDict] = None,
|
||||||
|
nameservers: list[str] = None,
|
||||||
|
timeout: int = 2.0,
|
||||||
|
) -> list[str]:
|
||||||
"""
|
"""
|
||||||
Queries DNS
|
Queries DNS
|
||||||
|
|
||||||
@@ -163,7 +184,13 @@ def query_dns(domain, record_type, cache=None, nameservers=None, timeout=2.0):
|
|||||||
return records
|
return records
|
||||||
|
|
||||||
|
|
||||||
def get_reverse_dns(ip_address, cache=None, nameservers=None, timeout=2.0):
|
def get_reverse_dns(
|
||||||
|
ip_address,
|
||||||
|
*,
|
||||||
|
cache: Optional[ExpiringDict] = None,
|
||||||
|
nameservers: list[str] = None,
|
||||||
|
timeout: int = 2.0,
|
||||||
|
) -> str:
|
||||||
"""
|
"""
|
||||||
Resolves an IP address to a hostname using a reverse DNS query
|
Resolves an IP address to a hostname using a reverse DNS query
|
||||||
|
|
||||||
@@ -191,7 +218,7 @@ def get_reverse_dns(ip_address, cache=None, nameservers=None, timeout=2.0):
|
|||||||
return hostname
|
return hostname
|
||||||
|
|
||||||
|
|
||||||
def timestamp_to_datetime(timestamp):
|
def timestamp_to_datetime(timestamp: int) -> datetime:
|
||||||
"""
|
"""
|
||||||
Converts a UNIX/DMARC timestamp to a Python ``datetime`` object
|
Converts a UNIX/DMARC timestamp to a Python ``datetime`` object
|
||||||
|
|
||||||
@@ -204,7 +231,7 @@ def timestamp_to_datetime(timestamp):
|
|||||||
return datetime.fromtimestamp(int(timestamp))
|
return datetime.fromtimestamp(int(timestamp))
|
||||||
|
|
||||||
|
|
||||||
def timestamp_to_human(timestamp):
|
def timestamp_to_human(timestamp: int) -> str:
|
||||||
"""
|
"""
|
||||||
Converts a UNIX/DMARC timestamp to a human-readable string
|
Converts a UNIX/DMARC timestamp to a human-readable string
|
||||||
|
|
||||||
@@ -217,7 +244,9 @@ def timestamp_to_human(timestamp):
|
|||||||
return timestamp_to_datetime(timestamp).strftime("%Y-%m-%d %H:%M:%S")
|
return timestamp_to_datetime(timestamp).strftime("%Y-%m-%d %H:%M:%S")
|
||||||
|
|
||||||
|
|
||||||
def human_timestamp_to_datetime(human_timestamp, to_utc=False):
|
def human_timestamp_to_datetime(
|
||||||
|
human_timestamp: str, *, to_utc: Optional[bool] = False
|
||||||
|
) -> datetime:
|
||||||
"""
|
"""
|
||||||
Converts a human-readable timestamp into a Python ``datetime`` object
|
Converts a human-readable timestamp into a Python ``datetime`` object
|
||||||
|
|
||||||
@@ -236,7 +265,7 @@ def human_timestamp_to_datetime(human_timestamp, to_utc=False):
|
|||||||
return dt.astimezone(timezone.utc) if to_utc else dt
|
return dt.astimezone(timezone.utc) if to_utc else dt
|
||||||
|
|
||||||
|
|
||||||
def human_timestamp_to_unix_timestamp(human_timestamp):
|
def human_timestamp_to_unix_timestamp(human_timestamp: str) -> int:
|
||||||
"""
|
"""
|
||||||
Converts a human-readable timestamp into a UNIX timestamp
|
Converts a human-readable timestamp into a UNIX timestamp
|
||||||
|
|
||||||
@@ -250,7 +279,7 @@ def human_timestamp_to_unix_timestamp(human_timestamp):
|
|||||||
return human_timestamp_to_datetime(human_timestamp).timestamp()
|
return human_timestamp_to_datetime(human_timestamp).timestamp()
|
||||||
|
|
||||||
|
|
||||||
def get_ip_address_country(ip_address, db_path=None):
|
def get_ip_address_country(ip_address: str, *, db_path: Optional[str] = None) -> str:
|
||||||
"""
|
"""
|
||||||
Returns the ISO code for the country associated
|
Returns the ISO code for the country associated
|
||||||
with the given IPv4 or IPv6 address
|
with the given IPv4 or IPv6 address
|
||||||
@@ -277,7 +306,7 @@ def get_ip_address_country(ip_address, db_path=None):
|
|||||||
]
|
]
|
||||||
|
|
||||||
if db_path is not None:
|
if db_path is not None:
|
||||||
if os.path.isfile(db_path) is False:
|
if not os.path.isfile(db_path):
|
||||||
db_path = None
|
db_path = None
|
||||||
logger.warning(
|
logger.warning(
|
||||||
f"No file exists at {db_path}. Falling back to an "
|
f"No file exists at {db_path}. Falling back to an "
|
||||||
@@ -313,13 +342,14 @@ def get_ip_address_country(ip_address, db_path=None):
|
|||||||
|
|
||||||
|
|
||||||
def get_service_from_reverse_dns_base_domain(
|
def get_service_from_reverse_dns_base_domain(
|
||||||
base_domain,
|
base_domain: str,
|
||||||
always_use_local_file=False,
|
*,
|
||||||
local_file_path=None,
|
always_use_local_file: Optional[bool] = False,
|
||||||
url=None,
|
local_file_path: Optional[str] = None,
|
||||||
offline=False,
|
url: Optional[str] = None,
|
||||||
reverse_dns_map=None,
|
offline: Optional[bool] = False,
|
||||||
):
|
reverse_dns_map: Optional[dict[str, Any]] = None,
|
||||||
|
) -> dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
Returns the service name of a given base domain name from reverse DNS.
|
Returns the service name of a given base domain name from reverse DNS.
|
||||||
|
|
||||||
@@ -388,17 +418,18 @@ def get_service_from_reverse_dns_base_domain(
|
|||||||
|
|
||||||
|
|
||||||
def get_ip_address_info(
|
def get_ip_address_info(
|
||||||
ip_address,
|
ip_address: str,
|
||||||
ip_db_path=None,
|
*,
|
||||||
reverse_dns_map_path=None,
|
ip_db_path: Optional[str] = None,
|
||||||
always_use_local_files=False,
|
reverse_dns_map_path: Optional[str] = None,
|
||||||
reverse_dns_map_url=None,
|
always_use_local_files: Optional[bool] = False,
|
||||||
cache=None,
|
reverse_dns_map_url: Optional[str] = None,
|
||||||
reverse_dns_map=None,
|
cache: Optional[ExpiringDict] = None,
|
||||||
offline=False,
|
reverse_dns_map: Optional[dict[str, Any]] = None,
|
||||||
nameservers=None,
|
offline: Optional[bool] = False,
|
||||||
timeout=2.0,
|
nameservers: Optional[list[str]] = None,
|
||||||
):
|
timeout: Optional[float] = 2.0,
|
||||||
|
) -> dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
Returns reverse DNS and country information for the given IP address
|
Returns reverse DNS and country information for the given IP address
|
||||||
|
|
||||||
@@ -416,7 +447,7 @@ def get_ip_address_info(
|
|||||||
timeout (float): Sets the DNS timeout in seconds
|
timeout (float): Sets the DNS timeout in seconds
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
OrderedDict: ``ip_address``, ``reverse_dns``
|
dict: ``ip_address``, ``reverse_dns``, ``country``
|
||||||
|
|
||||||
"""
|
"""
|
||||||
ip_address = ip_address.lower()
|
ip_address = ip_address.lower()
|
||||||
@@ -425,7 +456,7 @@ def get_ip_address_info(
|
|||||||
if info:
|
if info:
|
||||||
logger.debug(f"IP address {ip_address} was found in cache")
|
logger.debug(f"IP address {ip_address} was found in cache")
|
||||||
return info
|
return info
|
||||||
info = OrderedDict()
|
info = dict()
|
||||||
info["ip_address"] = ip_address
|
info["ip_address"] = ip_address
|
||||||
if offline:
|
if offline:
|
||||||
reverse_dns = None
|
reverse_dns = None
|
||||||
@@ -463,7 +494,7 @@ def get_ip_address_info(
|
|||||||
return info
|
return info
|
||||||
|
|
||||||
|
|
||||||
def parse_email_address(original_address):
|
def parse_email_address(original_address: str) -> EmailAddress:
|
||||||
if original_address[0] == "":
|
if original_address[0] == "":
|
||||||
display_name = None
|
display_name = None
|
||||||
else:
|
else:
|
||||||
@@ -476,17 +507,15 @@ def parse_email_address(original_address):
|
|||||||
local = address_parts[0].lower()
|
local = address_parts[0].lower()
|
||||||
domain = address_parts[-1].lower()
|
domain = address_parts[-1].lower()
|
||||||
|
|
||||||
return OrderedDict(
|
return {
|
||||||
[
|
"display_name": display_name,
|
||||||
("display_name", display_name),
|
"address": address,
|
||||||
("address", address),
|
"local": local,
|
||||||
("local", local),
|
"domain": domain,
|
||||||
("domain", domain),
|
}
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def get_filename_safe_string(string):
|
def get_filename_safe_string(string: str) -> str:
|
||||||
"""
|
"""
|
||||||
Converts a string to a string that is safe for a filename
|
Converts a string to a string that is safe for a filename
|
||||||
|
|
||||||
@@ -508,7 +537,7 @@ def get_filename_safe_string(string):
|
|||||||
return string
|
return string
|
||||||
|
|
||||||
|
|
||||||
def is_mbox(path):
|
def is_mbox(path: str) -> bool:
|
||||||
"""
|
"""
|
||||||
Checks if the given content is an MBOX mailbox file
|
Checks if the given content is an MBOX mailbox file
|
||||||
|
|
||||||
@@ -529,7 +558,7 @@ def is_mbox(path):
|
|||||||
return _is_mbox
|
return _is_mbox
|
||||||
|
|
||||||
|
|
||||||
def is_outlook_msg(content):
|
def is_outlook_msg(content: Union[bytes, Any]) -> bool:
|
||||||
"""
|
"""
|
||||||
Checks if the given content is an Outlook msg OLE/MSG file
|
Checks if the given content is an Outlook msg OLE/MSG file
|
||||||
|
|
||||||
@@ -544,7 +573,7 @@ def is_outlook_msg(content):
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def convert_outlook_msg(msg_bytes):
|
def convert_outlook_msg(msg_bytes: bytes) -> str:
|
||||||
"""
|
"""
|
||||||
Uses the ``msgconvert`` Perl utility to convert an Outlook MS file to
|
Uses the ``msgconvert`` Perl utility to convert an Outlook MS file to
|
||||||
standard RFC 822 format
|
standard RFC 822 format
|
||||||
@@ -562,13 +591,14 @@ def convert_outlook_msg(msg_bytes):
|
|||||||
os.chdir(tmp_dir)
|
os.chdir(tmp_dir)
|
||||||
with open("sample.msg", "wb") as msg_file:
|
with open("sample.msg", "wb") as msg_file:
|
||||||
msg_file.write(msg_bytes)
|
msg_file.write(msg_bytes)
|
||||||
|
rfc822_bytes: bytes
|
||||||
try:
|
try:
|
||||||
subprocess.check_call(
|
subprocess.check_call(
|
||||||
["msgconvert", "sample.msg"], stdout=null_file, stderr=null_file
|
["msgconvert", "sample.msg"], stdout=null_file, stderr=null_file
|
||||||
)
|
)
|
||||||
eml_path = "sample.eml"
|
eml_path = "sample.eml"
|
||||||
with open(eml_path, "rb") as eml_file:
|
with open(eml_path, "rb") as eml_file:
|
||||||
rfc822 = eml_file.read()
|
rfc822_bytes = eml_file.read()
|
||||||
except FileNotFoundError:
|
except FileNotFoundError:
|
||||||
raise EmailParserError(
|
raise EmailParserError(
|
||||||
"Failed to convert Outlook MSG: msgconvert utility not found"
|
"Failed to convert Outlook MSG: msgconvert utility not found"
|
||||||
@@ -577,10 +607,12 @@ def convert_outlook_msg(msg_bytes):
|
|||||||
os.chdir(orig_dir)
|
os.chdir(orig_dir)
|
||||||
shutil.rmtree(tmp_dir)
|
shutil.rmtree(tmp_dir)
|
||||||
|
|
||||||
return rfc822
|
return rfc822_bytes.decode("utf-8", errors="replace")
|
||||||
|
|
||||||
|
|
||||||
def parse_email(data, strip_attachment_payloads=False):
|
def parse_email(
|
||||||
|
data: Union[bytes, str], *, strip_attachment_payloads: Optional[bool] = False
|
||||||
|
) -> dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
A simplified email parser
|
A simplified email parser
|
||||||
|
|
||||||
@@ -595,7 +627,8 @@ def parse_email(data, strip_attachment_payloads=False):
|
|||||||
if isinstance(data, bytes):
|
if isinstance(data, bytes):
|
||||||
if is_outlook_msg(data):
|
if is_outlook_msg(data):
|
||||||
data = convert_outlook_msg(data)
|
data = convert_outlook_msg(data)
|
||||||
data = data.decode("utf-8", errors="replace")
|
else:
|
||||||
|
data = data.decode("utf-8", errors="replace")
|
||||||
parsed_email = mailparser.parse_from_string(data)
|
parsed_email = mailparser.parse_from_string(data)
|
||||||
headers = json.loads(parsed_email.headers_json).copy()
|
headers = json.loads(parsed_email.headers_json).copy()
|
||||||
parsed_email = json.loads(parsed_email.mail_json).copy()
|
parsed_email = json.loads(parsed_email.mail_json).copy()
|
||||||
|
|||||||
@@ -1,3 +1,9 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import Any, Optional, Union
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
|
||||||
from parsedmarc import logger
|
from parsedmarc import logger
|
||||||
@@ -7,7 +13,13 @@ from parsedmarc.constants import USER_AGENT
|
|||||||
class WebhookClient(object):
|
class WebhookClient(object):
|
||||||
"""A client for webhooks"""
|
"""A client for webhooks"""
|
||||||
|
|
||||||
def __init__(self, aggregate_url, forensic_url, smtp_tls_url, timeout=60):
|
def __init__(
|
||||||
|
self,
|
||||||
|
aggregate_url: str,
|
||||||
|
forensic_url: str,
|
||||||
|
smtp_tls_url: str,
|
||||||
|
timeout: Optional[int] = 60,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Initializes the WebhookClient
|
Initializes the WebhookClient
|
||||||
Args:
|
Args:
|
||||||
@@ -26,25 +38,27 @@ class WebhookClient(object):
|
|||||||
"Content-Type": "application/json",
|
"Content-Type": "application/json",
|
||||||
}
|
}
|
||||||
|
|
||||||
def save_forensic_report_to_webhook(self, report):
|
def save_forensic_report_to_webhook(self, report: str):
|
||||||
try:
|
try:
|
||||||
self._send_to_webhook(self.forensic_url, report)
|
self._send_to_webhook(self.forensic_url, report)
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
||||||
|
|
||||||
def save_smtp_tls_report_to_webhook(self, report):
|
def save_smtp_tls_report_to_webhook(self, report: str):
|
||||||
try:
|
try:
|
||||||
self._send_to_webhook(self.smtp_tls_url, report)
|
self._send_to_webhook(self.smtp_tls_url, report)
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
||||||
|
|
||||||
def save_aggregate_report_to_webhook(self, report):
|
def save_aggregate_report_to_webhook(self, report: str):
|
||||||
try:
|
try:
|
||||||
self._send_to_webhook(self.aggregate_url, report)
|
self._send_to_webhook(self.aggregate_url, report)
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
logger.error("Webhook Error: {0}".format(error_.__str__()))
|
||||||
|
|
||||||
def _send_to_webhook(self, webhook_url, payload):
|
def _send_to_webhook(
|
||||||
|
self, webhook_url: str, payload: Union[bytes, str, dict[str, Any]]
|
||||||
|
):
|
||||||
try:
|
try:
|
||||||
self.session.post(webhook_url, data=payload, timeout=self.timeout)
|
self.session.post(webhook_url, data=payload, timeout=self.timeout)
|
||||||
except Exception as error_:
|
except Exception as error_:
|
||||||
|
|||||||
@@ -2,6 +2,7 @@
|
|||||||
requires = [
|
requires = [
|
||||||
"hatchling>=1.27.0",
|
"hatchling>=1.27.0",
|
||||||
]
|
]
|
||||||
|
requires_python = ">=3.9,<3.14"
|
||||||
build-backend = "hatchling.build"
|
build-backend = "hatchling.build"
|
||||||
|
|
||||||
[project]
|
[project]
|
||||||
@@ -28,6 +29,7 @@ classifiers = [
|
|||||||
"Operating System :: OS Independent",
|
"Operating System :: OS Independent",
|
||||||
"Programming Language :: Python :: 3"
|
"Programming Language :: Python :: 3"
|
||||||
]
|
]
|
||||||
|
requires-python = ">=3.9, <3.14"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"azure-identity>=1.8.0",
|
"azure-identity>=1.8.0",
|
||||||
"azure-monitor-ingestion>=1.0.0",
|
"azure-monitor-ingestion>=1.0.0",
|
||||||
@@ -86,11 +88,11 @@ include = [
|
|||||||
|
|
||||||
[tool.hatch.build]
|
[tool.hatch.build]
|
||||||
exclude = [
|
exclude = [
|
||||||
"base_reverse_dns.csv",
|
"base_reverse_dns.csv",
|
||||||
"find_bad_utf8.py",
|
"find_bad_utf8.py",
|
||||||
"find_unknown_base_reverse_dns.py",
|
"find_unknown_base_reverse_dns.py",
|
||||||
"unknown_base_reverse_dns.csv",
|
"unknown_base_reverse_dns.csv",
|
||||||
"sortmaps.py",
|
"sortmaps.py",
|
||||||
"README.md",
|
"README.md",
|
||||||
"*.bak"
|
"*.bak"
|
||||||
]
|
]
|
||||||
|
|||||||
5
tests.py
5
tests.py
@@ -1,3 +1,6 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
|
||||||
from __future__ import absolute_import, print_function, unicode_literals
|
from __future__ import absolute_import, print_function, unicode_literals
|
||||||
|
|
||||||
import os
|
import os
|
||||||
@@ -74,7 +77,7 @@ class Test(unittest.TestCase):
|
|||||||
print()
|
print()
|
||||||
file = "samples/extract_report/nice-input.xml"
|
file = "samples/extract_report/nice-input.xml"
|
||||||
print("Testing {0}: ".format(file), end="")
|
print("Testing {0}: ".format(file), end="")
|
||||||
xmlout = parsedmarc.extract_report(file)
|
xmlout = parsedmarc.extract_report_from_file_path(file)
|
||||||
xmlin_file = open("samples/extract_report/nice-input.xml")
|
xmlin_file = open("samples/extract_report/nice-input.xml")
|
||||||
xmlin = xmlin_file.read()
|
xmlin = xmlin_file.read()
|
||||||
xmlin_file.close()
|
xmlin_file.close()
|
||||||
|
|||||||
Reference in New Issue
Block a user