Compare commits
	
		
			11 Commits
		
	
	
		
			docs-event
			...
			navbar-rev
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| eec91338ae | |||
| 6e9d8bc101 | |||
| 38e3ed4d05 | |||
| 93f54a596c | |||
| 2cfb6ce17e | |||
| 31d25eaead | |||
| 7a12fbf8b7 | |||
| c50793133b | |||
| b8e394227e | |||
| e8b1f82c3e | |||
| 92629578dd | 
| @ -1,5 +1,5 @@ | ||||
| [bumpversion] | ||||
| current_version = 2025.6.2 | ||||
| current_version = 2025.4.1 | ||||
| tag = True | ||||
| commit = True | ||||
| parse = (?P<major>\d+)\.(?P<minor>\d+)\.(?P<patch>\d+)(?:-(?P<rc_t>[a-zA-Z-]+)(?P<rc_n>[1-9]\\d*))? | ||||
| @ -21,8 +21,6 @@ optional_value = final | ||||
|  | ||||
| [bumpversion:file:package.json] | ||||
|  | ||||
| [bumpversion:file:package-lock.json] | ||||
|  | ||||
| [bumpversion:file:docker-compose.yml] | ||||
|  | ||||
| [bumpversion:file:schema.yml] | ||||
| @ -33,4 +31,6 @@ optional_value = final | ||||
|  | ||||
| [bumpversion:file:internal/constants/constants.go] | ||||
|  | ||||
| [bumpversion:file:web/src/common/constants.ts] | ||||
|  | ||||
| [bumpversion:file:lifecycle/aws/template.yaml] | ||||
|  | ||||
| @ -5,10 +5,8 @@ dist/** | ||||
| build/** | ||||
| build_docs/** | ||||
| *Dockerfile | ||||
| **/*Dockerfile | ||||
| blueprints/local | ||||
| .git | ||||
| !gen-ts-api/node_modules | ||||
| !gen-ts-api/dist/** | ||||
| !gen-go-api/ | ||||
| .venv | ||||
|  | ||||
| @ -7,9 +7,6 @@ charset = utf-8 | ||||
| trim_trailing_whitespace = true | ||||
| insert_final_newline = true | ||||
|  | ||||
| [*.toml] | ||||
| indent_size = 2 | ||||
|  | ||||
| [*.html] | ||||
| indent_size = 2 | ||||
|  | ||||
|  | ||||
							
								
								
									
										2
									
								
								.github/actions/setup/action.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.github/actions/setup/action.yml
									
									
									
									
										vendored
									
									
								
							| @ -36,7 +36,7 @@ runs: | ||||
|       with: | ||||
|         go-version-file: "go.mod" | ||||
|     - name: Setup docker cache | ||||
|       uses: AndreKurait/docker-cache@0fe76702a40db986d9663c24954fc14c6a6031b7 | ||||
|       uses: ScribeMD/docker-cache@0.5.0 | ||||
|       with: | ||||
|         key: docker-images-${{ runner.os }}-${{ hashFiles('.github/actions/setup/docker-compose.yml', 'Makefile') }}-${{ inputs.postgresql_version }} | ||||
|     - name: Setup dependencies | ||||
|  | ||||
							
								
								
									
										21
									
								
								.github/dependabot.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										21
									
								
								.github/dependabot.yml
									
									
									
									
										vendored
									
									
								
							| @ -23,13 +23,7 @@ updates: | ||||
|   - package-ecosystem: npm | ||||
|     directories: | ||||
|       - "/web" | ||||
|       - "/web/packages/sfe" | ||||
|       - "/web/packages/core" | ||||
|       - "/web/packages/esbuild-plugin-live-reload" | ||||
|       - "/packages/prettier-config" | ||||
|       - "/packages/tsconfig" | ||||
|       - "/packages/docusaurus-config" | ||||
|       - "/packages/eslint-config" | ||||
|       - "/web/sfe" | ||||
|     schedule: | ||||
|       interval: daily | ||||
|       time: "04:00" | ||||
| @ -74,9 +68,6 @@ updates: | ||||
|       wdio: | ||||
|         patterns: | ||||
|           - "@wdio/*" | ||||
|       goauthentik: | ||||
|         patterns: | ||||
|           - "@goauthentik/*" | ||||
|   - package-ecosystem: npm | ||||
|     directory: "/website" | ||||
|     schedule: | ||||
| @ -97,16 +88,6 @@ updates: | ||||
|           - "swc-*" | ||||
|           - "lightningcss*" | ||||
|           - "@rspack/binding*" | ||||
|       goauthentik: | ||||
|         patterns: | ||||
|           - "@goauthentik/*" | ||||
|       eslint: | ||||
|         patterns: | ||||
|           - "@eslint/*" | ||||
|           - "@typescript-eslint/*" | ||||
|           - "eslint-*" | ||||
|           - "eslint" | ||||
|           - "typescript-eslint" | ||||
|   - package-ecosystem: npm | ||||
|     directory: "/lifecycle/aws" | ||||
|     schedule: | ||||
|  | ||||
							
								
								
									
										4
									
								
								.github/workflows/ci-main.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										4
									
								
								.github/workflows/ci-main.yml
									
									
									
									
										vendored
									
									
								
							| @ -62,7 +62,6 @@ jobs: | ||||
|         psql: | ||||
|           - 15-alpine | ||||
|           - 16-alpine | ||||
|           - 17-alpine | ||||
|         run_id: [1, 2, 3, 4, 5] | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
| @ -117,7 +116,6 @@ jobs: | ||||
|         psql: | ||||
|           - 15-alpine | ||||
|           - 16-alpine | ||||
|           - 17-alpine | ||||
|         run_id: [1, 2, 3, 4, 5] | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
| @ -202,7 +200,7 @@ jobs: | ||||
|         uses: actions/cache@v4 | ||||
|         with: | ||||
|           path: web/dist | ||||
|           key: ${{ runner.os }}-web-${{ hashFiles('web/package-lock.json', 'package-lock.json', 'web/src/**', 'web/packages/sfe/src/**') }}-b | ||||
|           key: ${{ runner.os }}-web-${{ hashFiles('web/package-lock.json', 'web/src/**', 'web/packages/sfe/src/**') }}-b | ||||
|       - name: prepare web ui | ||||
|         if: steps.cache-web.outputs.cache-hit != 'true' | ||||
|         working-directory: web | ||||
|  | ||||
							
								
								
									
										50
									
								
								.github/workflows/ci-website.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										50
									
								
								.github/workflows/ci-website.yml
									
									
									
									
										vendored
									
									
								
							| @ -49,7 +49,6 @@ jobs: | ||||
|       matrix: | ||||
|         job: | ||||
|           - build | ||||
|           - build:integrations | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - uses: actions/setup-node@v4 | ||||
| @ -62,61 +61,12 @@ jobs: | ||||
|       - name: build | ||||
|         working-directory: website/ | ||||
|         run: npm run ${{ matrix.job }} | ||||
|   build-container: | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
|       # Needed to upload container images to ghcr.io | ||||
|       packages: write | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|         with: | ||||
|           ref: ${{ github.event.pull_request.head.sha }} | ||||
|       - name: Set up QEMU | ||||
|         uses: docker/setup-qemu-action@v3.6.0 | ||||
|       - name: Set up Docker Buildx | ||||
|         uses: docker/setup-buildx-action@v3 | ||||
|       - name: prepare variables | ||||
|         uses: ./.github/actions/docker-push-variables | ||||
|         id: ev | ||||
|         env: | ||||
|           DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }} | ||||
|         with: | ||||
|           image-name: ghcr.io/goauthentik/dev-docs | ||||
|       - name: Login to Container Registry | ||||
|         if: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|         uses: docker/login-action@v3 | ||||
|         with: | ||||
|           registry: ghcr.io | ||||
|           username: ${{ github.repository_owner }} | ||||
|           password: ${{ secrets.GITHUB_TOKEN }} | ||||
|       - name: Build Docker Image | ||||
|         id: push | ||||
|         uses: docker/build-push-action@v6 | ||||
|         with: | ||||
|           tags: ${{ steps.ev.outputs.imageTags }} | ||||
|           file: website/Dockerfile | ||||
|           push: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|           platforms: linux/amd64,linux/arm64 | ||||
|           context: . | ||||
|           cache-from: type=registry,ref=ghcr.io/goauthentik/dev-docs:buildcache | ||||
|           cache-to: ${{ steps.ev.outputs.shouldPush == 'true' && 'type=registry,ref=ghcr.io/goauthentik/dev-docs:buildcache,mode=max' || '' }} | ||||
|       - uses: actions/attest-build-provenance@v2 | ||||
|         id: attest | ||||
|         if: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|         with: | ||||
|           subject-name: ${{ steps.ev.outputs.attestImageNames }} | ||||
|           subject-digest: ${{ steps.push.outputs.digest }} | ||||
|           push-to-registry: true | ||||
|   ci-website-mark: | ||||
|     if: always() | ||||
|     needs: | ||||
|       - lint | ||||
|       - test | ||||
|       - build | ||||
|       - build-container | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|       - uses: re-actors/alls-green@release/v1 | ||||
|  | ||||
							
								
								
									
										16
									
								
								.github/workflows/packages-npm-publish.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										16
									
								
								.github/workflows/packages-npm-publish.yml
									
									
									
									
										vendored
									
									
								
							| @ -7,7 +7,6 @@ on: | ||||
|       - packages/eslint-config/** | ||||
|       - packages/prettier-config/** | ||||
|       - packages/tsconfig/** | ||||
|       - web/packages/esbuild-plugin-live-reload/** | ||||
|   workflow_dispatch: | ||||
| jobs: | ||||
|   publish: | ||||
| @ -17,28 +16,27 @@ jobs: | ||||
|       fail-fast: false | ||||
|       matrix: | ||||
|         package: | ||||
|           - packages/docusaurus-config | ||||
|           - packages/eslint-config | ||||
|           - packages/prettier-config | ||||
|           - packages/tsconfig | ||||
|           - web/packages/esbuild-plugin-live-reload | ||||
|           - docusaurus-config | ||||
|           - eslint-config | ||||
|           - prettier-config | ||||
|           - tsconfig | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|         with: | ||||
|           fetch-depth: 2 | ||||
|       - uses: actions/setup-node@v4 | ||||
|         with: | ||||
|           node-version-file: ${{ matrix.package }}/package.json | ||||
|           node-version-file: packages/${{ matrix.package }}/package.json | ||||
|           registry-url: "https://registry.npmjs.org" | ||||
|       - name: Get changed files | ||||
|         id: changed-files | ||||
|         uses: tj-actions/changed-files@ed68ef82c095e0d48ec87eccea555d944a631a4c | ||||
|         with: | ||||
|           files: | | ||||
|             ${{ matrix.package }}/package.json | ||||
|             packages/${{ matrix.package }}/package.json | ||||
|       - name: Publish package | ||||
|         if: steps.changed-files.outputs.any_changed == 'true' | ||||
|         working-directory: ${{ matrix.package }} | ||||
|         working-directory: packages/${{ matrix.package}} | ||||
|         run: | | ||||
|           npm ci | ||||
|           npm run build | ||||
|  | ||||
							
								
								
									
										45
									
								
								.github/workflows/release-publish.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										45
									
								
								.github/workflows/release-publish.yml
									
									
									
									
										vendored
									
									
								
							| @ -20,49 +20,6 @@ jobs: | ||||
|       release: true | ||||
|       registry_dockerhub: true | ||||
|       registry_ghcr: true | ||||
|   build-docs: | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
|       # Needed to upload container images to ghcr.io | ||||
|       packages: write | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - name: Set up QEMU | ||||
|         uses: docker/setup-qemu-action@v3.6.0 | ||||
|       - name: Set up Docker Buildx | ||||
|         uses: docker/setup-buildx-action@v3 | ||||
|       - name: prepare variables | ||||
|         uses: ./.github/actions/docker-push-variables | ||||
|         id: ev | ||||
|         env: | ||||
|           DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }} | ||||
|         with: | ||||
|           image-name: ghcr.io/goauthentik/docs | ||||
|       - name: Login to GitHub Container Registry | ||||
|         uses: docker/login-action@v3 | ||||
|         with: | ||||
|           registry: ghcr.io | ||||
|           username: ${{ github.repository_owner }} | ||||
|           password: ${{ secrets.GITHUB_TOKEN }} | ||||
|       - name: Build Docker Image | ||||
|         id: push | ||||
|         uses: docker/build-push-action@v6 | ||||
|         with: | ||||
|           tags: ${{ steps.ev.outputs.imageTags }} | ||||
|           file: website/Dockerfile | ||||
|           push: true | ||||
|           platforms: linux/amd64,linux/arm64 | ||||
|           context: . | ||||
|       - uses: actions/attest-build-provenance@v2 | ||||
|         id: attest | ||||
|         if: true | ||||
|         with: | ||||
|           subject-name: ${{ steps.ev.outputs.attestImageNames }} | ||||
|           subject-digest: ${{ steps.push.outputs.digest }} | ||||
|           push-to-registry: true | ||||
|   build-outpost: | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
| @ -236,6 +193,6 @@ jobs: | ||||
|           SENTRY_ORG: authentik-security-inc | ||||
|           SENTRY_PROJECT: authentik | ||||
|         with: | ||||
|           release: authentik@${{ steps.ev.outputs.version }} | ||||
|           version: authentik@${{ steps.ev.outputs.version }} | ||||
|           sourcemaps: "./web/dist" | ||||
|           url_prefix: "~/static/dist" | ||||
|  | ||||
							
								
								
									
										3
									
								
								.github/workflows/translation-rename.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										3
									
								
								.github/workflows/translation-rename.yml
									
									
									
									
										vendored
									
									
								
							| @ -15,7 +15,6 @@ jobs: | ||||
|     runs-on: ubuntu-latest | ||||
|     if: ${{ github.event.pull_request.user.login == 'transifex-integration[bot]'}} | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - id: generate_token | ||||
|         uses: tibdex/github-app-token@v2 | ||||
|         with: | ||||
| @ -32,7 +31,7 @@ jobs: | ||||
|         env: | ||||
|           GH_TOKEN: ${{ steps.generate_token.outputs.token }} | ||||
|         run: | | ||||
|           gh pr edit ${{ github.event.pull_request.number }} -t "translate: ${{ steps.title.outputs.title }}" --add-label dependencies | ||||
|           gh pr edit -t "translate: ${{ steps.title.outputs.title }}" --add-label dependencies | ||||
|       - uses: peter-evans/enable-pull-request-automerge@v3 | ||||
|         with: | ||||
|           token: ${{ steps.generate_token.outputs.token }} | ||||
|  | ||||
							
								
								
									
										50
									
								
								Dockerfile
									
									
									
									
									
								
							
							
						
						
									
										50
									
								
								Dockerfile
									
									
									
									
									
								
							| @ -1,7 +1,26 @@ | ||||
| # syntax=docker/dockerfile:1 | ||||
|  | ||||
| # Stage 1: Build webui | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:24-slim AS node-builder | ||||
| # Stage 1: Build website | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:22 AS website-builder | ||||
|  | ||||
| ENV NODE_ENV=production | ||||
|  | ||||
| WORKDIR /work/website | ||||
|  | ||||
| RUN --mount=type=bind,target=/work/website/package.json,src=./website/package.json \ | ||||
|     --mount=type=bind,target=/work/website/package-lock.json,src=./website/package-lock.json \ | ||||
|     --mount=type=cache,id=npm-website,sharing=shared,target=/root/.npm \ | ||||
|     npm ci --include=dev | ||||
|  | ||||
| COPY ./website /work/website/ | ||||
| COPY ./blueprints /work/blueprints/ | ||||
| COPY ./schema.yml /work/ | ||||
| COPY ./SECURITY.md /work/ | ||||
|  | ||||
| RUN npm run build-bundled | ||||
|  | ||||
| # Stage 2: Build webui | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:22 AS web-builder | ||||
|  | ||||
| ARG GIT_BUILD_HASH | ||||
| ENV GIT_BUILD_HASH=$GIT_BUILD_HASH | ||||
| @ -13,7 +32,7 @@ RUN --mount=type=bind,target=/work/web/package.json,src=./web/package.json \ | ||||
|     --mount=type=bind,target=/work/web/package-lock.json,src=./web/package-lock.json \ | ||||
|     --mount=type=bind,target=/work/web/packages/sfe/package.json,src=./web/packages/sfe/package.json \ | ||||
|     --mount=type=bind,target=/work/web/scripts,src=./web/scripts \ | ||||
|     --mount=type=cache,id=npm-ak,sharing=shared,target=/root/.npm \ | ||||
|     --mount=type=cache,id=npm-web,sharing=shared,target=/root/.npm \ | ||||
|     npm ci --include=dev | ||||
|  | ||||
| COPY ./package.json /work | ||||
| @ -24,7 +43,7 @@ COPY ./gen-ts-api /work/web/node_modules/@goauthentik/api | ||||
| RUN npm run build && \ | ||||
|     npm run build:sfe | ||||
|  | ||||
| # Stage 2: Build go proxy | ||||
| # Stage 3: Build go proxy | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/golang:1.24-bookworm AS go-builder | ||||
|  | ||||
| ARG TARGETOS | ||||
| @ -49,8 +68,8 @@ RUN --mount=type=bind,target=/go/src/goauthentik.io/go.mod,src=./go.mod \ | ||||
| COPY ./cmd /go/src/goauthentik.io/cmd | ||||
| COPY ./authentik/lib /go/src/goauthentik.io/authentik/lib | ||||
| COPY ./web/static.go /go/src/goauthentik.io/web/static.go | ||||
| COPY --from=node-builder /work/web/robots.txt /go/src/goauthentik.io/web/robots.txt | ||||
| COPY --from=node-builder /work/web/security.txt /go/src/goauthentik.io/web/security.txt | ||||
| COPY --from=web-builder /work/web/robots.txt /go/src/goauthentik.io/web/robots.txt | ||||
| COPY --from=web-builder /work/web/security.txt /go/src/goauthentik.io/web/security.txt | ||||
| COPY ./internal /go/src/goauthentik.io/internal | ||||
| COPY ./go.mod /go/src/goauthentik.io/go.mod | ||||
| COPY ./go.sum /go/src/goauthentik.io/go.sum | ||||
| @ -61,7 +80,7 @@ RUN --mount=type=cache,sharing=locked,target=/go/pkg/mod \ | ||||
|     CGO_ENABLED=1 GOFIPS140=latest GOARM="${TARGETVARIANT#v}" \ | ||||
|     go build -o /go/authentik ./cmd/server | ||||
|  | ||||
| # Stage 3: MaxMind GeoIP | ||||
| # Stage 4: MaxMind GeoIP | ||||
| FROM --platform=${BUILDPLATFORM} ghcr.io/maxmind/geoipupdate:v7.1.0 AS geoip | ||||
|  | ||||
| ENV GEOIPUPDATE_EDITION_IDS="GeoLite2-City GeoLite2-ASN" | ||||
| @ -74,10 +93,10 @@ RUN --mount=type=secret,id=GEOIPUPDATE_ACCOUNT_ID \ | ||||
|     mkdir -p /usr/share/GeoIP && \ | ||||
|     /bin/sh -c "GEOIPUPDATE_LICENSE_KEY_FILE=/run/secrets/GEOIPUPDATE_LICENSE_KEY /usr/bin/entry.sh || echo 'Failed to get GeoIP database, disabling'; exit 0" | ||||
|  | ||||
| # Stage 4: Download uv | ||||
| FROM ghcr.io/astral-sh/uv:0.7.13 AS uv | ||||
| # Stage 5: Base python image | ||||
| FROM ghcr.io/goauthentik/fips-python:3.13.5-slim-bookworm-fips AS python-base | ||||
| # Stage 5: Download uv | ||||
| FROM ghcr.io/astral-sh/uv:0.7.5 AS uv | ||||
| # Stage 6: Base python image | ||||
| FROM ghcr.io/goauthentik/fips-python:3.13.3-slim-bookworm-fips AS python-base | ||||
|  | ||||
| ENV VENV_PATH="/ak-root/.venv" \ | ||||
|     PATH="/lifecycle:/ak-root/.venv/bin:$PATH" \ | ||||
| @ -90,7 +109,7 @@ WORKDIR /ak-root/ | ||||
|  | ||||
| COPY --from=uv /uv /uvx /bin/ | ||||
|  | ||||
| # Stage 6: Python dependencies | ||||
| # Stage 7: Python dependencies | ||||
| FROM python-base AS python-deps | ||||
|  | ||||
| ARG TARGETARCH | ||||
| @ -125,7 +144,7 @@ RUN --mount=type=bind,target=pyproject.toml,src=pyproject.toml \ | ||||
|     --mount=type=cache,target=/root/.cache/uv \ | ||||
|     uv sync --frozen --no-install-project --no-dev | ||||
|  | ||||
| # Stage 7: Run | ||||
| # Stage 8: Run | ||||
| FROM python-base AS final-image | ||||
|  | ||||
| ARG VERSION | ||||
| @ -168,8 +187,9 @@ COPY ./lifecycle/ /lifecycle | ||||
| COPY ./authentik/sources/kerberos/krb5.conf /etc/krb5.conf | ||||
| COPY --from=go-builder /go/authentik /bin/authentik | ||||
| COPY --from=python-deps /ak-root/.venv /ak-root/.venv | ||||
| COPY --from=node-builder /work/web/dist/ /web/dist/ | ||||
| COPY --from=node-builder /work/web/authentik/ /web/authentik/ | ||||
| COPY --from=web-builder /work/web/dist/ /web/dist/ | ||||
| COPY --from=web-builder /work/web/authentik/ /web/authentik/ | ||||
| COPY --from=website-builder /work/website/build/ /website/help/ | ||||
| COPY --from=geoip /usr/share/GeoIP /geoip | ||||
|  | ||||
| USER 1000 | ||||
|  | ||||
							
								
								
									
										8
									
								
								Makefile
									
									
									
									
									
								
							
							
						
						
									
										8
									
								
								Makefile
									
									
									
									
									
								
							| @ -1,6 +1,6 @@ | ||||
| .PHONY: gen dev-reset all clean test web website | ||||
|  | ||||
| SHELL := /usr/bin/env bash | ||||
| SHELL := /bin/bash | ||||
| .SHELLFLAGS += ${SHELLFLAGS} -e -o pipefail | ||||
| PWD = $(shell pwd) | ||||
| UID = $(shell id -u) | ||||
| @ -86,10 +86,6 @@ dev-create-db: | ||||
|  | ||||
| dev-reset: dev-drop-db dev-create-db migrate  ## Drop and restore the Authentik PostgreSQL instance to a "fresh install" state. | ||||
|  | ||||
| update-test-mmdb:  ## Update test GeoIP and ASN Databases | ||||
| 	curl -L https://raw.githubusercontent.com/maxmind/MaxMind-DB/refs/heads/main/test-data/GeoLite2-ASN-Test.mmdb -o ${PWD}/tests/GeoLite2-ASN-Test.mmdb | ||||
| 	curl -L https://raw.githubusercontent.com/maxmind/MaxMind-DB/refs/heads/main/test-data/GeoLite2-City-Test.mmdb -o ${PWD}/tests/GeoLite2-City-Test.mmdb | ||||
|  | ||||
| ######################### | ||||
| ## API Schema | ||||
| ######################### | ||||
| @ -98,7 +94,7 @@ gen-build:  ## Extract the schema from the database | ||||
| 	AUTHENTIK_DEBUG=true \ | ||||
| 		AUTHENTIK_TENANTS__ENABLED=true \ | ||||
| 		AUTHENTIK_OUTPOSTS__DISABLE_EMBEDDED_OUTPOST=true \ | ||||
| 		uv run ak make_blueprint_schema --file blueprints/schema.json | ||||
| 		uv run ak make_blueprint_schema > blueprints/schema.json | ||||
| 	AUTHENTIK_DEBUG=true \ | ||||
| 		AUTHENTIK_TENANTS__ENABLED=true \ | ||||
| 		AUTHENTIK_OUTPOSTS__DISABLE_EMBEDDED_OUTPOST=true \ | ||||
|  | ||||
| @ -20,8 +20,8 @@ Even if the issue is not a CVE, we still greatly appreciate your help in hardeni | ||||
|  | ||||
| | Version   | Supported | | ||||
| | --------- | --------- | | ||||
| | 2025.2.x  | ✅        | | ||||
| | 2025.4.x  | ✅        | | ||||
| | 2025.6.x  | ✅        | | ||||
|  | ||||
| ## Reporting a Vulnerability | ||||
|  | ||||
|  | ||||
| @ -2,7 +2,7 @@ | ||||
|  | ||||
| from os import environ | ||||
|  | ||||
| __version__ = "2025.6.2" | ||||
| __version__ = "2025.4.1" | ||||
| ENV_GIT_HASH_KEY = "GIT_BUILD_HASH" | ||||
|  | ||||
|  | ||||
|  | ||||
							
								
								
									
										79
									
								
								authentik/admin/api/metrics.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										79
									
								
								authentik/admin/api/metrics.py
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,79 @@ | ||||
| """authentik administration metrics""" | ||||
|  | ||||
| from datetime import timedelta | ||||
|  | ||||
| from django.db.models.functions import ExtractHour | ||||
| from drf_spectacular.utils import extend_schema, extend_schema_field | ||||
| from guardian.shortcuts import get_objects_for_user | ||||
| from rest_framework.fields import IntegerField, SerializerMethodField | ||||
| from rest_framework.permissions import IsAuthenticated | ||||
| from rest_framework.request import Request | ||||
| from rest_framework.response import Response | ||||
| from rest_framework.views import APIView | ||||
|  | ||||
| from authentik.core.api.utils import PassiveSerializer | ||||
| from authentik.events.models import EventAction | ||||
|  | ||||
|  | ||||
| class CoordinateSerializer(PassiveSerializer): | ||||
|     """Coordinates for diagrams""" | ||||
|  | ||||
|     x_cord = IntegerField(read_only=True) | ||||
|     y_cord = IntegerField(read_only=True) | ||||
|  | ||||
|  | ||||
| class LoginMetricsSerializer(PassiveSerializer): | ||||
|     """Login Metrics per 1h""" | ||||
|  | ||||
|     logins = SerializerMethodField() | ||||
|     logins_failed = SerializerMethodField() | ||||
|     authorizations = SerializerMethodField() | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins(self, _): | ||||
|         """Get successful logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins_failed(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN_FAILED | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_authorizations(self, _): | ||||
|         """Get successful authorizations per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|  | ||||
| class AdministrationMetricsViewSet(APIView): | ||||
|     """Login Metrics per 1h""" | ||||
|  | ||||
|     permission_classes = [IsAuthenticated] | ||||
|  | ||||
|     @extend_schema(responses={200: LoginMetricsSerializer(many=False)}) | ||||
|     def get(self, request: Request) -> Response: | ||||
|         """Login Metrics per 1h""" | ||||
|         serializer = LoginMetricsSerializer(True) | ||||
|         serializer.context["user"] = request.user | ||||
|         return Response(serializer.data) | ||||
| @ -1,7 +1,6 @@ | ||||
| """authentik administration overview""" | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django_tenants.utils import get_public_schema_name | ||||
| from drf_spectacular.utils import extend_schema | ||||
| from packaging.version import parse | ||||
| from rest_framework.fields import SerializerMethodField | ||||
| @ -14,7 +13,6 @@ from authentik import __version__, get_build_hash | ||||
| from authentik.admin.tasks import VERSION_CACHE_KEY, VERSION_NULL, update_latest_version | ||||
| from authentik.core.api.utils import PassiveSerializer | ||||
| from authentik.outposts.models import Outpost | ||||
| from authentik.tenants.utils import get_current_tenant | ||||
|  | ||||
|  | ||||
| class VersionSerializer(PassiveSerializer): | ||||
| @ -37,8 +35,6 @@ class VersionSerializer(PassiveSerializer): | ||||
|  | ||||
|     def get_version_latest(self, _) -> str: | ||||
|         """Get latest version from cache""" | ||||
|         if get_current_tenant().schema_name == get_public_schema_name(): | ||||
|             return __version__ | ||||
|         version_in_cache = cache.get(VERSION_CACHE_KEY) | ||||
|         if not version_in_cache:  # pragma: no cover | ||||
|             update_latest_version.delay() | ||||
|  | ||||
| @ -14,19 +14,3 @@ class AuthentikAdminConfig(ManagedAppConfig): | ||||
|     label = "authentik_admin" | ||||
|     verbose_name = "authentik Admin" | ||||
|     default = True | ||||
|  | ||||
|     @ManagedAppConfig.reconcile_global | ||||
|     def clear_update_notifications(self): | ||||
|         """Clear update notifications on startup if the notification was for the version | ||||
|         we're running now.""" | ||||
|         from packaging.version import parse | ||||
|  | ||||
|         from authentik.admin.tasks import LOCAL_VERSION | ||||
|         from authentik.events.models import EventAction, Notification | ||||
|  | ||||
|         for notification in Notification.objects.filter(event__action=EventAction.UPDATE_AVAILABLE): | ||||
|             if "new_version" not in notification.event.context: | ||||
|                 continue | ||||
|             notification_version = notification.event.context["new_version"] | ||||
|             if LOCAL_VERSION >= parse(notification_version): | ||||
|                 notification.delete() | ||||
|  | ||||
| @ -1,7 +1,6 @@ | ||||
| """authentik admin settings""" | ||||
|  | ||||
| from celery.schedules import crontab | ||||
| from django_tenants.utils import get_public_schema_name | ||||
|  | ||||
| from authentik.lib.utils.time import fqdn_rand | ||||
|  | ||||
| @ -9,7 +8,6 @@ CELERY_BEAT_SCHEDULE = { | ||||
|     "admin_latest_version": { | ||||
|         "task": "authentik.admin.tasks.update_latest_version", | ||||
|         "schedule": crontab(minute=fqdn_rand("admin_latest_version"), hour="*"), | ||||
|         "tenant_schemas": [get_public_schema_name()], | ||||
|         "options": {"queue": "authentik_scheduled"}, | ||||
|     } | ||||
| } | ||||
|  | ||||
| @ -1,6 +1,7 @@ | ||||
| """authentik admin tasks""" | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django.db import DatabaseError, InternalError, ProgrammingError | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
| from packaging.version import parse | ||||
| from requests import RequestException | ||||
| @ -8,7 +9,7 @@ from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik import __version__, get_build_hash | ||||
| from authentik.admin.apps import PROM_INFO | ||||
| from authentik.events.models import Event, EventAction | ||||
| from authentik.events.models import Event, EventAction, Notification | ||||
| from authentik.events.system_tasks import SystemTask, TaskStatus, prefill_task | ||||
| from authentik.lib.config import CONFIG | ||||
| from authentik.lib.utils.http import get_http_session | ||||
| @ -32,6 +33,20 @@ def _set_prom_info(): | ||||
|     ) | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task( | ||||
|     throws=(DatabaseError, ProgrammingError, InternalError), | ||||
| ) | ||||
| def clear_update_notifications(): | ||||
|     """Clear update notifications on startup if the notification was for the version | ||||
|     we're running now.""" | ||||
|     for notification in Notification.objects.filter(event__action=EventAction.UPDATE_AVAILABLE): | ||||
|         if "new_version" not in notification.event.context: | ||||
|             continue | ||||
|         notification_version = notification.event.context["new_version"] | ||||
|         if LOCAL_VERSION >= parse(notification_version): | ||||
|             notification.delete() | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task(bind=True, base=SystemTask) | ||||
| @prefill_task | ||||
| def update_latest_version(self: SystemTask): | ||||
|  | ||||
| @ -36,6 +36,11 @@ class TestAdminAPI(TestCase): | ||||
|         body = loads(response.content) | ||||
|         self.assertEqual(len(body), 0) | ||||
|  | ||||
|     def test_metrics(self): | ||||
|         """Test metrics API""" | ||||
|         response = self.client.get(reverse("authentik_api:admin_metrics")) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_apps(self): | ||||
|         """Test apps API""" | ||||
|         response = self.client.get(reverse("authentik_api:apps-list")) | ||||
|  | ||||
| @ -1,12 +1,12 @@ | ||||
| """test admin tasks""" | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.core.cache import cache | ||||
| from django.test import TestCase | ||||
| from requests_mock import Mocker | ||||
|  | ||||
| from authentik.admin.tasks import ( | ||||
|     VERSION_CACHE_KEY, | ||||
|     clear_update_notifications, | ||||
|     update_latest_version, | ||||
| ) | ||||
| from authentik.events.models import Event, EventAction | ||||
| @ -72,13 +72,12 @@ class TestAdminTasks(TestCase): | ||||
|  | ||||
|     def test_clear_update_notifications(self): | ||||
|         """Test clear of previous notification""" | ||||
|         admin_config = apps.get_app_config("authentik_admin") | ||||
|         Event.objects.create( | ||||
|             action=EventAction.UPDATE_AVAILABLE, context={"new_version": "99999999.9999999.9999999"} | ||||
|         ) | ||||
|         Event.objects.create(action=EventAction.UPDATE_AVAILABLE, context={"new_version": "1.1.1"}) | ||||
|         Event.objects.create(action=EventAction.UPDATE_AVAILABLE, context={}) | ||||
|         admin_config.clear_update_notifications() | ||||
|         clear_update_notifications() | ||||
|         self.assertFalse( | ||||
|             Event.objects.filter( | ||||
|                 action=EventAction.UPDATE_AVAILABLE, context__new_version="1.1" | ||||
|  | ||||
| @ -3,6 +3,7 @@ | ||||
| from django.urls import path | ||||
|  | ||||
| from authentik.admin.api.meta import AppsViewSet, ModelViewSet | ||||
| from authentik.admin.api.metrics import AdministrationMetricsViewSet | ||||
| from authentik.admin.api.system import SystemView | ||||
| from authentik.admin.api.version import VersionView | ||||
| from authentik.admin.api.version_history import VersionHistoryViewSet | ||||
| @ -11,6 +12,11 @@ from authentik.admin.api.workers import WorkerView | ||||
| api_urlpatterns = [ | ||||
|     ("admin/apps", AppsViewSet, "apps"), | ||||
|     ("admin/models", ModelViewSet, "models"), | ||||
|     path( | ||||
|         "admin/metrics/", | ||||
|         AdministrationMetricsViewSet.as_view(), | ||||
|         name="admin_metrics", | ||||
|     ), | ||||
|     path("admin/version/", VersionView.as_view(), name="admin_version"), | ||||
|     ("admin/version/history", VersionHistoryViewSet, "version_history"), | ||||
|     path("admin/workers/", WorkerView.as_view(), name="admin_workers"), | ||||
|  | ||||
| @ -1,13 +1,12 @@ | ||||
| """authentik API AppConfig""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikAPIConfig(ManagedAppConfig): | ||||
| class AuthentikAPIConfig(AppConfig): | ||||
|     """authentik API Config""" | ||||
|  | ||||
|     name = "authentik.api" | ||||
|     label = "authentik_api" | ||||
|     mountpoint = "api/" | ||||
|     verbose_name = "authentik API" | ||||
|     default = True | ||||
|  | ||||
| @ -72,33 +72,20 @@ class Command(BaseCommand): | ||||
|                     "additionalProperties": True, | ||||
|                 }, | ||||
|                 "entries": { | ||||
|                     "anyOf": [ | ||||
|                         { | ||||
|                     "type": "array", | ||||
|                             "items": {"$ref": "#/$defs/blueprint_entry"}, | ||||
|                         }, | ||||
|                         { | ||||
|                             "type": "object", | ||||
|                             "additionalProperties": { | ||||
|                                 "type": "array", | ||||
|                                 "items": {"$ref": "#/$defs/blueprint_entry"}, | ||||
|                     "items": { | ||||
|                         "oneOf": [], | ||||
|                     }, | ||||
|                 }, | ||||
|                     ], | ||||
|             }, | ||||
|             }, | ||||
|             "$defs": {"blueprint_entry": {"oneOf": []}}, | ||||
|             "$defs": {}, | ||||
|         } | ||||
|  | ||||
|     def add_arguments(self, parser): | ||||
|         parser.add_argument("--file", type=str) | ||||
|  | ||||
|     @no_translations | ||||
|     def handle(self, *args, file: str, **options): | ||||
|     def handle(self, *args, **options): | ||||
|         """Generate JSON Schema for blueprints""" | ||||
|         self.build() | ||||
|         with open(file, "w") as _schema: | ||||
|             _schema.write(dumps(self.schema, indent=4, default=Command.json_default)) | ||||
|         self.stdout.write(dumps(self.schema, indent=4, default=Command.json_default)) | ||||
|  | ||||
|     @staticmethod | ||||
|     def json_default(value: Any) -> Any: | ||||
| @ -125,7 +112,7 @@ class Command(BaseCommand): | ||||
|                 } | ||||
|             ) | ||||
|             model_path = f"{model._meta.app_label}.{model._meta.model_name}" | ||||
|             self.schema["$defs"]["blueprint_entry"]["oneOf"].append( | ||||
|             self.schema["properties"]["entries"]["items"]["oneOf"].append( | ||||
|                 self.template_entry(model_path, model, serializer) | ||||
|             ) | ||||
|  | ||||
| @ -147,7 +134,7 @@ class Command(BaseCommand): | ||||
|                 "id": {"type": "string"}, | ||||
|                 "state": { | ||||
|                     "type": "string", | ||||
|                     "enum": sorted([s.value for s in BlueprintEntryDesiredState]), | ||||
|                     "enum": [s.value for s in BlueprintEntryDesiredState], | ||||
|                     "default": "present", | ||||
|                 }, | ||||
|                 "conditions": {"type": "array", "items": {"type": "boolean"}}, | ||||
| @ -218,7 +205,7 @@ class Command(BaseCommand): | ||||
|                 "type": "object", | ||||
|                 "required": ["permission"], | ||||
|                 "properties": { | ||||
|                     "permission": {"type": "string", "enum": sorted(perms)}, | ||||
|                     "permission": {"type": "string", "enum": perms}, | ||||
|                     "user": {"type": "integer"}, | ||||
|                     "role": {"type": "string"}, | ||||
|                 }, | ||||
|  | ||||
| @ -1,6 +1,5 @@ | ||||
| version: 1 | ||||
| entries: | ||||
|   foo: | ||||
|     - identifiers: | ||||
|           name: "%(id)s" | ||||
|           slug: "%(id)s" | ||||
|  | ||||
| @ -1,14 +0,0 @@ | ||||
| from django.test import TestCase | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from authentik.enterprise.apps import EnterpriseConfig | ||||
| from authentik.lib.utils.reflection import get_apps | ||||
|  | ||||
|  | ||||
| class TestManagedAppConfig(TestCase): | ||||
|     def test_apps_use_managed_app_config(self): | ||||
|         for app in get_apps(): | ||||
|             if app.name.startswith("authentik.enterprise"): | ||||
|                 self.assertIn(EnterpriseConfig, app.__class__.__bases__) | ||||
|             else: | ||||
|                 self.assertIn(ManagedAppConfig, app.__class__.__bases__) | ||||
| @ -191,18 +191,11 @@ class Blueprint: | ||||
|     """Dataclass used for a full export""" | ||||
|  | ||||
|     version: int = field(default=1) | ||||
|     entries: list[BlueprintEntry] | dict[str, list[BlueprintEntry]] = field(default_factory=list) | ||||
|     entries: list[BlueprintEntry] = field(default_factory=list) | ||||
|     context: dict = field(default_factory=dict) | ||||
|  | ||||
|     metadata: BlueprintMetadata | None = field(default=None) | ||||
|  | ||||
|     def iter_entries(self) -> Iterable[BlueprintEntry]: | ||||
|         if isinstance(self.entries, dict): | ||||
|             for _section, entries in self.entries.items(): | ||||
|                 yield from entries | ||||
|         else: | ||||
|             yield from self.entries | ||||
|  | ||||
|  | ||||
| class YAMLTag: | ||||
|     """Base class for all YAML Tags""" | ||||
| @ -233,7 +226,7 @@ class KeyOf(YAMLTag): | ||||
|         self.id_from = node.value | ||||
|  | ||||
|     def resolve(self, entry: BlueprintEntry, blueprint: Blueprint) -> Any: | ||||
|         for _entry in blueprint.iter_entries(): | ||||
|         for _entry in blueprint.entries: | ||||
|             if _entry.id == self.id_from and _entry._state.instance: | ||||
|                 # Special handling for PolicyBindingModels, as they'll have a different PK | ||||
|                 # which is used when creating policy bindings | ||||
|  | ||||
| @ -384,7 +384,7 @@ class Importer: | ||||
|     def _apply_models(self, raise_errors=False) -> bool: | ||||
|         """Apply (create/update) models yaml""" | ||||
|         self.__pk_map = {} | ||||
|         for entry in self._import.iter_entries(): | ||||
|         for entry in self._import.entries: | ||||
|             model_app_label, model_name = entry.get_model(self._import).split(".") | ||||
|             try: | ||||
|                 model: type[SerializerModel] = registry.get_model(model_app_label, model_name) | ||||
|  | ||||
| @ -47,7 +47,7 @@ class MetaModelRegistry: | ||||
|         models = apps.get_models() | ||||
|         for _, value in self.models.items(): | ||||
|             models.append(value) | ||||
|         return sorted(models, key=str) | ||||
|         return models | ||||
|  | ||||
|     def get_model(self, app_label: str, model_id: str) -> type[Model]: | ||||
|         """Get model checks if any virtual models are registered, and falls back | ||||
|  | ||||
| @ -1,9 +1,9 @@ | ||||
| """authentik brands app""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikBrandsConfig(ManagedAppConfig): | ||||
| class AuthentikBrandsConfig(AppConfig): | ||||
|     """authentik Brand app""" | ||||
|  | ||||
|     name = "authentik.brands" | ||||
| @ -12,4 +12,3 @@ class AuthentikBrandsConfig(ManagedAppConfig): | ||||
|     mountpoints = { | ||||
|         "authentik.brands.urls_root": "", | ||||
|     } | ||||
|     default = True | ||||
|  | ||||
| @ -148,14 +148,3 @@ class TestBrands(APITestCase): | ||||
|                 "default_locale": "", | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
|     def test_custom_css(self): | ||||
|         """Test custom_css""" | ||||
|         brand = create_test_brand() | ||||
|         brand.branding_custom_css = """* { | ||||
|             font-family: "Foo bar"; | ||||
|         }""" | ||||
|         brand.save() | ||||
|         res = self.client.get(reverse("authentik_core:if-user")) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         self.assertIn(brand.branding_custom_css, res.content.decode()) | ||||
|  | ||||
| @ -5,8 +5,6 @@ from typing import Any | ||||
| from django.db.models import F, Q | ||||
| from django.db.models import Value as V | ||||
| from django.http.request import HttpRequest | ||||
| from django.utils.html import _json_script_escapes | ||||
| from django.utils.safestring import mark_safe | ||||
|  | ||||
| from authentik import get_full_version | ||||
| from authentik.brands.models import Brand | ||||
| @ -34,13 +32,8 @@ def context_processor(request: HttpRequest) -> dict[str, Any]: | ||||
|     """Context Processor that injects brand object into every template""" | ||||
|     brand = getattr(request, "brand", DEFAULT_BRAND) | ||||
|     tenant = getattr(request, "tenant", Tenant()) | ||||
|     # similarly to `json_script` we escape everything HTML-related, however django | ||||
|     # only directly exposes this as a function that also wraps it in a <script> tag | ||||
|     # which we dont want for CSS | ||||
|     brand_css = mark_safe(str(brand.branding_custom_css).translate(_json_script_escapes))  # nosec | ||||
|     return { | ||||
|         "brand": brand, | ||||
|         "brand_css": brand_css, | ||||
|         "footer_links": tenant.footer_links, | ||||
|         "html_meta": {**get_http_meta()}, | ||||
|         "version": get_full_version(), | ||||
|  | ||||
| @ -2,9 +2,11 @@ | ||||
|  | ||||
| from collections.abc import Iterator | ||||
| from copy import copy | ||||
| from datetime import timedelta | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django.db.models import QuerySet | ||||
| from django.db.models.functions import ExtractHour | ||||
| from django.shortcuts import get_object_or_404 | ||||
| from drf_spectacular.types import OpenApiTypes | ||||
| from drf_spectacular.utils import OpenApiParameter, OpenApiResponse, extend_schema | ||||
| @ -18,6 +20,7 @@ from rest_framework.response import Response | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.admin.api.metrics import CoordinateSerializer | ||||
| from authentik.api.pagination import Pagination | ||||
| from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT | ||||
| from authentik.core.api.providers import ProviderSerializer | ||||
| @ -25,6 +28,7 @@ from authentik.core.api.used_by import UsedByMixin | ||||
| from authentik.core.api.utils import ModelSerializer | ||||
| from authentik.core.models import Application, User | ||||
| from authentik.events.logs import LogEventSerializer, capture_logs | ||||
| from authentik.events.models import EventAction | ||||
| from authentik.lib.utils.file import ( | ||||
|     FilePathSerializer, | ||||
|     FileUploadSerializer, | ||||
| @ -317,3 +321,18 @@ class ApplicationViewSet(UsedByMixin, ModelViewSet): | ||||
|         """Set application icon (as URL)""" | ||||
|         app: Application = self.get_object() | ||||
|         return set_file_url(request, app, "meta_icon") | ||||
|  | ||||
|     @permission_required("authentik_core.view_application", ["authentik_events.view_event"]) | ||||
|     @extend_schema(responses={200: CoordinateSerializer(many=True)}) | ||||
|     @action(detail=True, pagination_class=None, filter_backends=[]) | ||||
|     def metrics(self, request: Request, slug: str): | ||||
|         """Metrics for application logins""" | ||||
|         app = self.get_object() | ||||
|         return Response( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION, | ||||
|                 context__authorized_application__pk=app.pk.hex, | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
| @ -6,6 +6,7 @@ from typing import Any | ||||
|  | ||||
| from django.contrib.auth import update_session_auth_hash | ||||
| from django.contrib.auth.models import Permission | ||||
| from django.db.models.functions import ExtractHour | ||||
| from django.db.transaction import atomic | ||||
| from django.db.utils import IntegrityError | ||||
| from django.urls import reverse_lazy | ||||
| @ -51,6 +52,7 @@ from rest_framework.validators import UniqueValidator | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.admin.api.metrics import CoordinateSerializer | ||||
| from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT | ||||
| from authentik.brands.models import Brand | ||||
| from authentik.core.api.used_by import UsedByMixin | ||||
| @ -82,7 +84,6 @@ from authentik.flows.views.executor import QS_KEY_TOKEN | ||||
| from authentik.lib.avatars import get_avatar | ||||
| from authentik.rbac.decorators import permission_required | ||||
| from authentik.rbac.models import get_permission_choices | ||||
| from authentik.stages.email.flow import pickle_flow_token_for_email | ||||
| from authentik.stages.email.models import EmailStage | ||||
| from authentik.stages.email.tasks import send_mails | ||||
| from authentik.stages.email.utils import TemplateEmailMessage | ||||
| @ -315,6 +316,53 @@ class SessionUserSerializer(PassiveSerializer): | ||||
|     original = UserSelfSerializer(required=False) | ||||
|  | ||||
|  | ||||
| class UserMetricsSerializer(PassiveSerializer): | ||||
|     """User Metrics""" | ||||
|  | ||||
|     logins = SerializerMethodField() | ||||
|     logins_failed = SerializerMethodField() | ||||
|     authorizations = SerializerMethodField() | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins(self, _): | ||||
|         """Get successful logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN, user__pk=user.pk | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins_failed(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN_FAILED, context__username=user.username | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_authorizations(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION, user__pk=user.pk | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|  | ||||
| class UsersFilter(FilterSet): | ||||
|     """Filter for users""" | ||||
|  | ||||
| @ -386,23 +434,8 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|     queryset = User.objects.none() | ||||
|     ordering = ["username"] | ||||
|     serializer_class = UserSerializer | ||||
|     filterset_class = UsersFilter | ||||
|     search_fields = ["username", "name", "is_active", "email", "uuid", "attributes"] | ||||
|  | ||||
|     def get_ql_fields(self): | ||||
|         from djangoql.schema import BoolField, StrField | ||||
|  | ||||
|         from authentik.enterprise.search.fields import ChoiceSearchField, JSONSearchField | ||||
|  | ||||
|         return [ | ||||
|             StrField(User, "username"), | ||||
|             StrField(User, "name"), | ||||
|             StrField(User, "email"), | ||||
|             StrField(User, "path"), | ||||
|             BoolField(User, "is_active", nullable=True), | ||||
|             ChoiceSearchField(User, "type"), | ||||
|             JSONSearchField(User, "attributes"), | ||||
|         ] | ||||
|     filterset_class = UsersFilter | ||||
|  | ||||
|     def get_queryset(self): | ||||
|         base_qs = User.objects.all().exclude_anonymous() | ||||
| @ -418,7 +451,7 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|     def list(self, request, *args, **kwargs): | ||||
|         return super().list(request, *args, **kwargs) | ||||
|  | ||||
|     def _create_recovery_link(self, for_email=False) -> tuple[str, Token]: | ||||
|     def _create_recovery_link(self) -> tuple[str, Token]: | ||||
|         """Create a recovery link (when the current brand has a recovery flow set), | ||||
|         that can either be shown to an admin or sent to the user directly""" | ||||
|         brand: Brand = self.request._request.brand | ||||
| @ -440,16 +473,12 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|             raise ValidationError( | ||||
|                 {"non_field_errors": "Recovery flow not applicable to user"} | ||||
|             ) from None | ||||
|         _plan = FlowToken.pickle(plan) | ||||
|         if for_email: | ||||
|             _plan = pickle_flow_token_for_email(plan) | ||||
|         token, __ = FlowToken.objects.update_or_create( | ||||
|             identifier=f"{user.uid}-password-reset", | ||||
|             defaults={ | ||||
|                 "user": user, | ||||
|                 "flow": flow, | ||||
|                 "_plan": _plan, | ||||
|                 "revoke_on_execution": not for_email, | ||||
|                 "_plan": FlowToken.pickle(plan), | ||||
|             }, | ||||
|         ) | ||||
|         querystring = urlencode({QS_KEY_TOKEN: token.key}) | ||||
| @ -573,6 +602,17 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|             update_session_auth_hash(self.request, user) | ||||
|         return Response(status=204) | ||||
|  | ||||
|     @permission_required("authentik_core.view_user", ["authentik_events.view_event"]) | ||||
|     @extend_schema(responses={200: UserMetricsSerializer(many=False)}) | ||||
|     @action(detail=True, pagination_class=None, filter_backends=[]) | ||||
|     def metrics(self, request: Request, pk: int) -> Response: | ||||
|         """User metrics per 1h""" | ||||
|         user: User = self.get_object() | ||||
|         serializer = UserMetricsSerializer(instance={}) | ||||
|         serializer.context["user"] = user | ||||
|         serializer.context["request"] = request | ||||
|         return Response(serializer.data) | ||||
|  | ||||
|     @permission_required("authentik_core.reset_user_password") | ||||
|     @extend_schema( | ||||
|         responses={ | ||||
| @ -608,7 +648,7 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|         if for_user.email == "": | ||||
|             LOGGER.debug("User doesn't have an email address") | ||||
|             raise ValidationError({"non_field_errors": "User does not have an email address set."}) | ||||
|         link, token = self._create_recovery_link(for_email=True) | ||||
|         link, token = self._create_recovery_link() | ||||
|         # Lookup the email stage to assure the current user can access it | ||||
|         stages = get_objects_for_user( | ||||
|             request.user, "authentik_stages_email.view_emailstage" | ||||
|  | ||||
| @ -79,7 +79,6 @@ def _migrate_session( | ||||
|         AuthenticatedSession.objects.using(db_alias).create( | ||||
|             session=session, | ||||
|             user=old_auth_session.user, | ||||
|             uuid=old_auth_session.uuid, | ||||
|         ) | ||||
|  | ||||
|  | ||||
|  | ||||
| @ -1,81 +1,10 @@ | ||||
| # Generated by Django 5.1.9 on 2025-05-14 11:15 | ||||
|  | ||||
| from django.apps.registry import Apps, apps as global_apps | ||||
| from django.apps.registry import Apps | ||||
| from django.db import migrations | ||||
| from django.contrib.contenttypes.management import create_contenttypes | ||||
| from django.contrib.auth.management import create_permissions | ||||
| from django.db.backends.base.schema import BaseDatabaseSchemaEditor | ||||
|  | ||||
|  | ||||
| def migrate_authenticated_session_permissions(apps: Apps, schema_editor: BaseDatabaseSchemaEditor): | ||||
|     """Migrate permissions from OldAuthenticatedSession to AuthenticatedSession""" | ||||
|     db_alias = schema_editor.connection.alias | ||||
|  | ||||
|     # `apps` here is just an instance of `django.db.migrations.state.AppConfigStub`, we need the | ||||
|     # real config for creating permissions and content types | ||||
|     authentik_core_config = global_apps.get_app_config("authentik_core") | ||||
|     # These are only ran by django after all migrations, but we need them right now. | ||||
|     # `global_apps` is needed, | ||||
|     create_permissions(authentik_core_config, using=db_alias, verbosity=1) | ||||
|     create_contenttypes(authentik_core_config, using=db_alias, verbosity=1) | ||||
|  | ||||
|     # But from now on, this is just a regular migration, so use `apps` | ||||
|     Permission = apps.get_model("auth", "Permission") | ||||
|     ContentType = apps.get_model("contenttypes", "ContentType") | ||||
|  | ||||
|     try: | ||||
|         old_ct = ContentType.objects.using(db_alias).get( | ||||
|             app_label="authentik_core", model="oldauthenticatedsession" | ||||
|         ) | ||||
|         new_ct = ContentType.objects.using(db_alias).get( | ||||
|             app_label="authentik_core", model="authenticatedsession" | ||||
|         ) | ||||
|     except ContentType.DoesNotExist: | ||||
|         # This should exist at this point, but if not, let's cut our losses | ||||
|         return | ||||
|  | ||||
|     # Get all permissions for the old content type | ||||
|     old_perms = Permission.objects.using(db_alias).filter(content_type=old_ct) | ||||
|  | ||||
|     # Create equivalent permissions for the new content type | ||||
|     for old_perm in old_perms: | ||||
|         new_perm = ( | ||||
|             Permission.objects.using(db_alias) | ||||
|             .filter( | ||||
|                 content_type=new_ct, | ||||
|                 codename=old_perm.codename, | ||||
|             ) | ||||
|             .first() | ||||
|         ) | ||||
|         if not new_perm: | ||||
|             # This should exist at this point, but if not, let's cut our losses | ||||
|             continue | ||||
|  | ||||
|         # Global user permissions | ||||
|         User = apps.get_model("authentik_core", "User") | ||||
|         User.user_permissions.through.objects.using(db_alias).filter( | ||||
|             permission=old_perm | ||||
|         ).all().update(permission=new_perm) | ||||
|  | ||||
|         # Global role permissions | ||||
|         DjangoGroup = apps.get_model("auth", "Group") | ||||
|         DjangoGroup.permissions.through.objects.using(db_alias).filter( | ||||
|             permission=old_perm | ||||
|         ).all().update(permission=new_perm) | ||||
|  | ||||
|         # Object user permissions | ||||
|         UserObjectPermission = apps.get_model("guardian", "UserObjectPermission") | ||||
|         UserObjectPermission.objects.using(db_alias).filter(permission=old_perm).all().update( | ||||
|             permission=new_perm, content_type=new_ct | ||||
|         ) | ||||
|  | ||||
|         # Object role permissions | ||||
|         GroupObjectPermission = apps.get_model("guardian", "GroupObjectPermission") | ||||
|         GroupObjectPermission.objects.using(db_alias).filter(permission=old_perm).all().update( | ||||
|             permission=new_perm, content_type=new_ct | ||||
|         ) | ||||
|  | ||||
|  | ||||
| def remove_old_authenticated_session_content_type( | ||||
|     apps: Apps, schema_editor: BaseDatabaseSchemaEditor | ||||
| ): | ||||
| @ -92,12 +21,7 @@ class Migration(migrations.Migration): | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.RunPython( | ||||
|             code=migrate_authenticated_session_permissions, | ||||
|             reverse_code=migrations.RunPython.noop, | ||||
|         ), | ||||
|         migrations.RunPython( | ||||
|             code=remove_old_authenticated_session_content_type, | ||||
|             reverse_code=migrations.RunPython.noop, | ||||
|         ), | ||||
|     ] | ||||
|  | ||||
| @ -18,7 +18,7 @@ from django.http import HttpRequest | ||||
| from django.utils.functional import SimpleLazyObject, cached_property | ||||
| from django.utils.timezone import now | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
| from django_cte import CTE, with_cte | ||||
| from django_cte import CTEQuerySet, With | ||||
| from guardian.conf import settings | ||||
| from guardian.mixins import GuardianUserMixin | ||||
| from model_utils.managers import InheritanceManager | ||||
| @ -136,7 +136,7 @@ class AttributesMixin(models.Model): | ||||
|         return instance, False | ||||
|  | ||||
|  | ||||
| class GroupQuerySet(QuerySet): | ||||
| class GroupQuerySet(CTEQuerySet): | ||||
|     def with_children_recursive(self): | ||||
|         """Recursively get all groups that have the current queryset as parents | ||||
|         or are indirectly related.""" | ||||
| @ -165,9 +165,9 @@ class GroupQuerySet(QuerySet): | ||||
|             ) | ||||
|  | ||||
|         # Build the recursive query, see above | ||||
|         cte = CTE.recursive(make_cte) | ||||
|         cte = With.recursive(make_cte) | ||||
|         # Return the result, as a usable queryset for Group. | ||||
|         return with_cte(cte, select=cte.join(Group, group_uuid=cte.col.group_uuid)) | ||||
|         return cte.join(Group, group_uuid=cte.col.group_uuid).with_cte(cte) | ||||
|  | ||||
|  | ||||
| class Group(SerializerModel, AttributesMixin): | ||||
|  | ||||
| @ -16,7 +16,7 @@ | ||||
|         {% block head_before %} | ||||
|         {% endblock %} | ||||
|         <link rel="stylesheet" type="text/css" href="{% static 'dist/authentik.css' %}"> | ||||
|         <style>{{ brand_css }}</style> | ||||
|         <style>{{ brand.branding_custom_css }}</style> | ||||
|         <script src="{% versioned_script 'dist/poly-%v.js' %}" type="module"></script> | ||||
|         <script src="{% versioned_script 'dist/standalone/loading/index-%v.js' %}" type="module"></script> | ||||
|         {% block head %} | ||||
|  | ||||
| @ -10,7 +10,7 @@ | ||||
| {% endblock %} | ||||
|  | ||||
| {% block body %} | ||||
| <ak-message-container alignment="bottom"></ak-message-container> | ||||
| <ak-message-container></ak-message-container> | ||||
| <ak-interface-admin> | ||||
|     <ak-loading></ak-loading> | ||||
| </ak-interface-admin> | ||||
|  | ||||
| @ -114,7 +114,6 @@ class TestApplicationsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
| @ -168,7 +167,6 @@ class TestApplicationsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -81,6 +81,22 @@ class TestUsersAPI(APITestCase): | ||||
|         response = self.client.get(reverse("authentik_api:user-list"), {"include_groups": "true"}) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_metrics(self): | ||||
|         """Test user's metrics""" | ||||
|         self.client.force_login(self.admin) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:user-metrics", kwargs={"pk": self.user.pk}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_metrics_denied(self): | ||||
|         """Test user's metrics (non-superuser)""" | ||||
|         self.client.force_login(self.user) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:user-metrics", kwargs={"pk": self.user.pk}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 403) | ||||
|  | ||||
|     def test_recovery_no_flow(self): | ||||
|         """Test user recovery link (no recovery flow set)""" | ||||
|         self.client.force_login(self.admin) | ||||
|  | ||||
| @ -119,17 +119,17 @@ class TestTrimPasswordHistory(TestCase): | ||||
|             [ | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter1",  # nosec | ||||
|                     old_password="hunter1",  # nosec B106 | ||||
|                     created_at=_now - timedelta(days=3), | ||||
|                 ), | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter2",  # nosec | ||||
|                     old_password="hunter2",  # nosec B106 | ||||
|                     created_at=_now - timedelta(days=2), | ||||
|                 ), | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter3",  # nosec | ||||
|                     old_password="hunter3",  # nosec B106 | ||||
|                     created_at=_now, | ||||
|                 ), | ||||
|             ] | ||||
|  | ||||
| @ -1,12 +0,0 @@ | ||||
| """Enterprise app config""" | ||||
|  | ||||
| from authentik.enterprise.apps import EnterpriseConfig | ||||
|  | ||||
|  | ||||
| class AuthentikEnterpriseSearchConfig(EnterpriseConfig): | ||||
|     """Enterprise app config""" | ||||
|  | ||||
|     name = "authentik.enterprise.search" | ||||
|     label = "authentik_search" | ||||
|     verbose_name = "authentik Enterprise.Search" | ||||
|     default = True | ||||
| @ -1,128 +0,0 @@ | ||||
| """DjangoQL search""" | ||||
|  | ||||
| from collections import OrderedDict, defaultdict | ||||
| from collections.abc import Generator | ||||
|  | ||||
| from django.db import connection | ||||
| from django.db.models import Model, Q | ||||
| from djangoql.compat import text_type | ||||
| from djangoql.schema import StrField | ||||
|  | ||||
|  | ||||
| class JSONSearchField(StrField): | ||||
|     """JSON field for DjangoQL""" | ||||
|  | ||||
|     model: Model | ||||
|  | ||||
|     def __init__(self, model=None, name=None, nullable=None, suggest_nested=True): | ||||
|         # Set this in the constructor to not clobber the type variable | ||||
|         self.type = "relation" | ||||
|         self.suggest_nested = suggest_nested | ||||
|         super().__init__(model, name, nullable) | ||||
|  | ||||
|     def get_lookup(self, path, operator, value): | ||||
|         search = "__".join(path) | ||||
|         op, invert = self.get_operator(operator) | ||||
|         q = Q(**{f"{search}{op}": self.get_lookup_value(value)}) | ||||
|         return ~q if invert else q | ||||
|  | ||||
|     def json_field_keys(self) -> Generator[tuple[str]]: | ||||
|         with connection.cursor() as cursor: | ||||
|             cursor.execute( | ||||
|                 f""" | ||||
|                 WITH RECURSIVE "{self.name}_keys" AS ( | ||||
|                     SELECT | ||||
|                         ARRAY[jsonb_object_keys("{self.name}")] AS key_path_array, | ||||
|                         "{self.name}" -> jsonb_object_keys("{self.name}") AS value | ||||
|                     FROM {self.model._meta.db_table} | ||||
|                     WHERE "{self.name}" IS NOT NULL | ||||
|                         AND jsonb_typeof("{self.name}") = 'object' | ||||
|  | ||||
|                     UNION ALL | ||||
|  | ||||
|                     SELECT | ||||
|                         ck.key_path_array || jsonb_object_keys(ck.value), | ||||
|                         ck.value -> jsonb_object_keys(ck.value) AS value | ||||
|                     FROM "{self.name}_keys" ck | ||||
|                     WHERE jsonb_typeof(ck.value) = 'object' | ||||
|                 ), | ||||
|  | ||||
|                 unique_paths AS ( | ||||
|                     SELECT DISTINCT key_path_array | ||||
|                     FROM "{self.name}_keys" | ||||
|                 ) | ||||
|  | ||||
|                 SELECT key_path_array FROM unique_paths; | ||||
|             """  # nosec | ||||
|             ) | ||||
|             return (x[0] for x in cursor.fetchall()) | ||||
|  | ||||
|     def get_nested_options(self) -> OrderedDict: | ||||
|         """Get keys of all nested objects to show autocomplete""" | ||||
|         if not self.suggest_nested: | ||||
|             return OrderedDict() | ||||
|         base_model_name = f"{self.model._meta.app_label}.{self.model._meta.model_name}_{self.name}" | ||||
|  | ||||
|         def recursive_function(parts: list[str], parent_parts: list[str] | None = None): | ||||
|             if not parent_parts: | ||||
|                 parent_parts = [] | ||||
|             path = parts.pop(0) | ||||
|             parent_parts.append(path) | ||||
|             relation_key = "_".join(parent_parts) | ||||
|             if len(parts) > 1: | ||||
|                 out_dict = { | ||||
|                     relation_key: { | ||||
|                         parts[0]: { | ||||
|                             "type": "relation", | ||||
|                             "relation": f"{relation_key}_{parts[0]}", | ||||
|                         } | ||||
|                     } | ||||
|                 } | ||||
|                 child_paths = recursive_function(parts.copy(), parent_parts.copy()) | ||||
|                 child_paths.update(out_dict) | ||||
|                 return child_paths | ||||
|             else: | ||||
|                 return {relation_key: {parts[0]: {}}} | ||||
|  | ||||
|         relation_structure = defaultdict(dict) | ||||
|  | ||||
|         for relations in self.json_field_keys(): | ||||
|             result = recursive_function([base_model_name] + relations) | ||||
|             for relation_key, value in result.items(): | ||||
|                 for sub_relation_key, sub_value in value.items(): | ||||
|                     if not relation_structure[relation_key].get(sub_relation_key, None): | ||||
|                         relation_structure[relation_key][sub_relation_key] = sub_value | ||||
|                     else: | ||||
|                         relation_structure[relation_key][sub_relation_key].update(sub_value) | ||||
|  | ||||
|         final_dict = defaultdict(dict) | ||||
|  | ||||
|         for key, value in relation_structure.items(): | ||||
|             for sub_key, sub_value in value.items(): | ||||
|                 if not sub_value: | ||||
|                     final_dict[key][sub_key] = { | ||||
|                         "type": "str", | ||||
|                         "nullable": True, | ||||
|                     } | ||||
|                 else: | ||||
|                     final_dict[key][sub_key] = sub_value | ||||
|         return OrderedDict(final_dict) | ||||
|  | ||||
|     def relation(self) -> str: | ||||
|         return f"{self.model._meta.app_label}.{self.model._meta.model_name}_{self.name}" | ||||
|  | ||||
|  | ||||
| class ChoiceSearchField(StrField): | ||||
|     def __init__(self, model=None, name=None, nullable=None): | ||||
|         super().__init__(model, name, nullable, suggest_options=True) | ||||
|  | ||||
|     def get_options(self, search): | ||||
|         result = [] | ||||
|         choices = self._field_choices() | ||||
|         if choices: | ||||
|             search = search.lower() | ||||
|             for c in choices: | ||||
|                 choice = text_type(c[0]) | ||||
|                 if search in choice.lower(): | ||||
|                     result.append(choice) | ||||
|         return result | ||||
| @ -1,53 +0,0 @@ | ||||
| from rest_framework.response import Response | ||||
|  | ||||
| from authentik.api.pagination import Pagination | ||||
| from authentik.enterprise.search.ql import AUTOCOMPLETE_COMPONENT_NAME, QLSearch | ||||
|  | ||||
|  | ||||
| class AutocompletePagination(Pagination): | ||||
|  | ||||
|     def paginate_queryset(self, queryset, request, view=None): | ||||
|         self.view = view | ||||
|         return super().paginate_queryset(queryset, request, view) | ||||
|  | ||||
|     def get_autocomplete(self): | ||||
|         schema = QLSearch().get_schema(self.request, self.view) | ||||
|         introspections = {} | ||||
|         if hasattr(self.view, "get_ql_fields"): | ||||
|             from authentik.enterprise.search.schema import AKQLSchemaSerializer | ||||
|  | ||||
|             introspections = AKQLSchemaSerializer().serialize( | ||||
|                 schema(self.page.paginator.object_list.model) | ||||
|             ) | ||||
|         return introspections | ||||
|  | ||||
|     def get_paginated_response(self, data): | ||||
|         previous_page_number = 0 | ||||
|         if self.page.has_previous(): | ||||
|             previous_page_number = self.page.previous_page_number() | ||||
|         next_page_number = 0 | ||||
|         if self.page.has_next(): | ||||
|             next_page_number = self.page.next_page_number() | ||||
|         return Response( | ||||
|             { | ||||
|                 "pagination": { | ||||
|                     "next": next_page_number, | ||||
|                     "previous": previous_page_number, | ||||
|                     "count": self.page.paginator.count, | ||||
|                     "current": self.page.number, | ||||
|                     "total_pages": self.page.paginator.num_pages, | ||||
|                     "start_index": self.page.start_index(), | ||||
|                     "end_index": self.page.end_index(), | ||||
|                 }, | ||||
|                 "results": data, | ||||
|                 "autocomplete": self.get_autocomplete(), | ||||
|             } | ||||
|         ) | ||||
|  | ||||
|     def get_paginated_response_schema(self, schema): | ||||
|         final_schema = super().get_paginated_response_schema(schema) | ||||
|         final_schema["properties"]["autocomplete"] = { | ||||
|             "$ref": f"#/components/schemas/{AUTOCOMPLETE_COMPONENT_NAME}" | ||||
|         } | ||||
|         final_schema["required"].append("autocomplete") | ||||
|         return final_schema | ||||
| @ -1,78 +0,0 @@ | ||||
| """DjangoQL search""" | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.db.models import QuerySet | ||||
| from djangoql.ast import Name | ||||
| from djangoql.exceptions import DjangoQLError | ||||
| from djangoql.queryset import apply_search | ||||
| from djangoql.schema import DjangoQLSchema | ||||
| from rest_framework.filters import SearchFilter | ||||
| from rest_framework.request import Request | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.enterprise.search.fields import JSONSearchField | ||||
|  | ||||
| LOGGER = get_logger() | ||||
| AUTOCOMPLETE_COMPONENT_NAME = "Autocomplete" | ||||
| AUTOCOMPLETE_SCHEMA = { | ||||
|     "type": "object", | ||||
|     "additionalProperties": {}, | ||||
| } | ||||
|  | ||||
|  | ||||
| class BaseSchema(DjangoQLSchema): | ||||
|     """Base Schema which deals with JSON Fields""" | ||||
|  | ||||
|     def resolve_name(self, name: Name): | ||||
|         model = self.model_label(self.current_model) | ||||
|         root_field = name.parts[0] | ||||
|         field = self.models[model].get(root_field) | ||||
|         # If the query goes into a JSON field, return the root | ||||
|         # field as the JSON field will do the rest | ||||
|         if isinstance(field, JSONSearchField): | ||||
|             # This is a workaround; build_filter will remove the right-most | ||||
|             # entry in the path as that is intended to be the same as the field | ||||
|             # however for JSON that is not the case | ||||
|             if name.parts[-1] != root_field: | ||||
|                 name.parts.append(root_field) | ||||
|             return field | ||||
|         return super().resolve_name(name) | ||||
|  | ||||
|  | ||||
| class QLSearch(SearchFilter): | ||||
|     """rest_framework search filter which uses DjangoQL""" | ||||
|  | ||||
|     @property | ||||
|     def enabled(self): | ||||
|         return apps.get_app_config("authentik_enterprise").enabled() | ||||
|  | ||||
|     def get_search_terms(self, request) -> str: | ||||
|         """ | ||||
|         Search terms are set by a ?search=... query parameter, | ||||
|         and may be comma and/or whitespace delimited. | ||||
|         """ | ||||
|         params = request.query_params.get(self.search_param, "") | ||||
|         params = params.replace("\x00", "")  # strip null characters | ||||
|         return params | ||||
|  | ||||
|     def get_schema(self, request: Request, view) -> BaseSchema: | ||||
|         ql_fields = [] | ||||
|         if hasattr(view, "get_ql_fields"): | ||||
|             ql_fields = view.get_ql_fields() | ||||
|  | ||||
|         class InlineSchema(BaseSchema): | ||||
|             def get_fields(self, model): | ||||
|                 return ql_fields or [] | ||||
|  | ||||
|         return InlineSchema | ||||
|  | ||||
|     def filter_queryset(self, request: Request, queryset: QuerySet, view) -> QuerySet: | ||||
|         search_query = self.get_search_terms(request) | ||||
|         schema = self.get_schema(request, view) | ||||
|         if len(search_query) == 0 or not self.enabled: | ||||
|             return super().filter_queryset(request, queryset, view) | ||||
|         try: | ||||
|             return apply_search(queryset, search_query, schema=schema) | ||||
|         except DjangoQLError as exc: | ||||
|             LOGGER.debug("Failed to parse search expression", exc=exc) | ||||
|             return super().filter_queryset(request, queryset, view) | ||||
| @ -1,29 +0,0 @@ | ||||
| from djangoql.serializers import DjangoQLSchemaSerializer | ||||
| from drf_spectacular.generators import SchemaGenerator | ||||
|  | ||||
| from authentik.api.schema import create_component | ||||
| from authentik.enterprise.search.fields import JSONSearchField | ||||
| from authentik.enterprise.search.ql import AUTOCOMPLETE_COMPONENT_NAME, AUTOCOMPLETE_SCHEMA | ||||
|  | ||||
|  | ||||
| class AKQLSchemaSerializer(DjangoQLSchemaSerializer): | ||||
|     def serialize(self, schema): | ||||
|         serialization = super().serialize(schema) | ||||
|         for _, fields in schema.models.items(): | ||||
|             for _, field in fields.items(): | ||||
|                 if not isinstance(field, JSONSearchField): | ||||
|                     continue | ||||
|                 serialization["models"].update(field.get_nested_options()) | ||||
|         return serialization | ||||
|  | ||||
|     def serialize_field(self, field): | ||||
|         result = super().serialize_field(field) | ||||
|         if isinstance(field, JSONSearchField): | ||||
|             result["relation"] = field.relation() | ||||
|         return result | ||||
|  | ||||
|  | ||||
| def postprocess_schema_search_autocomplete(result, generator: SchemaGenerator, **kwargs): | ||||
|     create_component(generator, AUTOCOMPLETE_COMPONENT_NAME, AUTOCOMPLETE_SCHEMA) | ||||
|  | ||||
|     return result | ||||
| @ -1,17 +0,0 @@ | ||||
| SPECTACULAR_SETTINGS = { | ||||
|     "POSTPROCESSING_HOOKS": [ | ||||
|         "authentik.api.schema.postprocess_schema_responses", | ||||
|         "authentik.enterprise.search.schema.postprocess_schema_search_autocomplete", | ||||
|         "drf_spectacular.hooks.postprocess_schema_enums", | ||||
|     ], | ||||
| } | ||||
|  | ||||
| REST_FRAMEWORK = { | ||||
|     "DEFAULT_PAGINATION_CLASS": "authentik.enterprise.search.pagination.AutocompletePagination", | ||||
|     "DEFAULT_FILTER_BACKENDS": [ | ||||
|         "authentik.enterprise.search.ql.QLSearch", | ||||
|         "authentik.rbac.filters.ObjectFilter", | ||||
|         "django_filters.rest_framework.DjangoFilterBackend", | ||||
|         "rest_framework.filters.OrderingFilter", | ||||
|     ], | ||||
| } | ||||
| @ -1,78 +0,0 @@ | ||||
| from json import loads | ||||
| from unittest.mock import PropertyMock, patch | ||||
| from urllib.parse import urlencode | ||||
|  | ||||
| from django.urls import reverse | ||||
| from rest_framework.test import APITestCase | ||||
|  | ||||
| from authentik.core.tests.utils import create_test_admin_user | ||||
|  | ||||
|  | ||||
| @patch( | ||||
|     "authentik.enterprise.audit.middleware.EnterpriseAuditMiddleware.enabled", | ||||
|     PropertyMock(return_value=True), | ||||
| ) | ||||
| class QLTest(APITestCase): | ||||
|  | ||||
|     def setUp(self): | ||||
|         self.user = create_test_admin_user() | ||||
|         # ensure we have more than 1 user | ||||
|         create_test_admin_user() | ||||
|  | ||||
|     def test_search(self): | ||||
|         """Test simple search query""" | ||||
|         self.client.force_login(self.user) | ||||
|         query = f'username = "{self.user.username}"' | ||||
|         res = self.client.get( | ||||
|             reverse( | ||||
|                 "authentik_api:user-list", | ||||
|             ) | ||||
|             + f"?{urlencode({"search": query})}" | ||||
|         ) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         content = loads(res.content) | ||||
|         self.assertEqual(content["pagination"]["count"], 1) | ||||
|         self.assertEqual(content["results"][0]["username"], self.user.username) | ||||
|  | ||||
|     def test_no_search(self): | ||||
|         """Ensure works with no search query""" | ||||
|         self.client.force_login(self.user) | ||||
|         res = self.client.get( | ||||
|             reverse( | ||||
|                 "authentik_api:user-list", | ||||
|             ) | ||||
|         ) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         content = loads(res.content) | ||||
|         self.assertNotEqual(content["pagination"]["count"], 1) | ||||
|  | ||||
|     def test_search_no_ql(self): | ||||
|         """Test simple search query (no QL)""" | ||||
|         self.client.force_login(self.user) | ||||
|         res = self.client.get( | ||||
|             reverse( | ||||
|                 "authentik_api:user-list", | ||||
|             ) | ||||
|             + f"?{urlencode({"search": self.user.username})}" | ||||
|         ) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         content = loads(res.content) | ||||
|         self.assertGreaterEqual(content["pagination"]["count"], 1) | ||||
|         self.assertEqual(content["results"][0]["username"], self.user.username) | ||||
|  | ||||
|     def test_search_json(self): | ||||
|         """Test search query with a JSON attribute""" | ||||
|         self.user.attributes = {"foo": {"bar": "baz"}} | ||||
|         self.user.save() | ||||
|         self.client.force_login(self.user) | ||||
|         query = 'attributes.foo.bar = "baz"' | ||||
|         res = self.client.get( | ||||
|             reverse( | ||||
|                 "authentik_api:user-list", | ||||
|             ) | ||||
|             + f"?{urlencode({"search": query})}" | ||||
|         ) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         content = loads(res.content) | ||||
|         self.assertEqual(content["pagination"]["count"], 1) | ||||
|         self.assertEqual(content["results"][0]["username"], self.user.username) | ||||
| @ -18,7 +18,6 @@ TENANT_APPS = [ | ||||
|     "authentik.enterprise.providers.google_workspace", | ||||
|     "authentik.enterprise.providers.microsoft_entra", | ||||
|     "authentik.enterprise.providers.ssf", | ||||
|     "authentik.enterprise.search", | ||||
|     "authentik.enterprise.stages.authenticator_endpoint_gdtc", | ||||
|     "authentik.enterprise.stages.mtls", | ||||
|     "authentik.enterprise.stages.source", | ||||
|  | ||||
| @ -3,14 +3,7 @@ from urllib.parse import unquote_plus | ||||
|  | ||||
| from cryptography.exceptions import InvalidSignature | ||||
| from cryptography.hazmat.primitives import hashes | ||||
| from cryptography.x509 import ( | ||||
|     Certificate, | ||||
|     NameOID, | ||||
|     ObjectIdentifier, | ||||
|     UnsupportedGeneralNameType, | ||||
|     load_pem_x509_certificate, | ||||
| ) | ||||
| from cryptography.x509.verification import PolicyBuilder, Store, VerificationError | ||||
| from cryptography.x509 import Certificate, NameOID, ObjectIdentifier, load_pem_x509_certificate | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
|  | ||||
| from authentik.brands.models import Brand | ||||
| @ -109,21 +102,15 @@ class MTLSStageView(ChallengeStageView): | ||||
|         return None | ||||
|  | ||||
|     def validate_cert(self, authorities: list[CertificateKeyPair], certs: list[Certificate]): | ||||
|         authorities_cert = [x.certificate for x in authorities] | ||||
|         for _cert in certs: | ||||
|             for ca in authorities: | ||||
|                 try: | ||||
|                 PolicyBuilder().store(Store(authorities_cert)).build_client_verifier().verify( | ||||
|                     _cert, [] | ||||
|                 ) | ||||
|                     _cert.verify_directly_issued_by(ca.certificate) | ||||
|                     return _cert | ||||
|             except ( | ||||
|                 InvalidSignature, | ||||
|                 TypeError, | ||||
|                 ValueError, | ||||
|                 VerificationError, | ||||
|                 UnsupportedGeneralNameType, | ||||
|             ) as exc: | ||||
|                 self.logger.warning("Discarding invalid certificate", cert=_cert, exc=exc) | ||||
|                 except (InvalidSignature, TypeError, ValueError) as exc: | ||||
|                     self.logger.warning( | ||||
|                         "Discarding cert not issued by authority", cert=_cert, authority=ca, exc=exc | ||||
|                     ) | ||||
|                     continue | ||||
|         return None | ||||
|  | ||||
| @ -154,9 +141,7 @@ class MTLSStageView(ChallengeStageView): | ||||
|             "subject": cert.subject.rfc4514_string(), | ||||
|             "issuer": cert.issuer.rfc4514_string(), | ||||
|             "fingerprint_sha256": hexlify(cert.fingerprint(hashes.SHA256()), ":").decode("utf-8"), | ||||
|             "fingerprint_sha1": hexlify(cert.fingerprint(hashes.SHA1()), ":").decode(  # nosec | ||||
|                 "utf-8" | ||||
|             ), | ||||
|             "fingerprint_sha1": hexlify(cert.fingerprint(hashes.SHA256()), ":").decode("utf-8"), | ||||
|         } | ||||
|  | ||||
|     def auth_user(self, user: User, cert: Certificate): | ||||
|  | ||||
| @ -1,31 +1,30 @@ | ||||
| -----BEGIN CERTIFICATE----- | ||||
| MIIFWTCCA0GgAwIBAgIUDEnKCSmIXG/akySGes7bhOGrN/8wDQYJKoZIhvcNAQEL | ||||
| MIIFOzCCAyOgAwIBAgIUbnIMy+Ewi5RvK7OBDxWMCk7wi08wDQYJKoZIhvcNAQEL | ||||
| BQAwRjEaMBgGA1UEAwwRYXV0aGVudGlrIFRlc3QgQ0ExEjAQBgNVBAoMCWF1dGhl | ||||
| bnRpazEUMBIGA1UECwwLU2VsZi1zaWduZWQwHhcNMjUwNTE5MTIzODQ2WhcNMjYw | ||||
| NTE1MTIzODQ2WjARMQ8wDQYDVQQDDAZjbGllbnQwggIiMA0GCSqGSIb3DQEBAQUA | ||||
| A4ICDwAwggIKAoICAQCkPkS1V6l0gj0ulxMznkxkgrw4p9Tjd8teSsGZt02A2Eo6 | ||||
| 7D8FbJ7pp3d5fYW/TWuEKVBLWTID6rijW5EGcdgTM5Jxf/QR+aZTEK6umQxUd4yO | ||||
| mOtp+xVS3KlcsSej2dFpeE5h5VkZizHpvh5xkoAP8W5VtQLOVF0hIeumHnJmaeLj | ||||
| +mhK9PBFpO7k9SFrYYhd/uLrYbIdANihbIO2Q74rNEJHewhFNM7oNSjjEWzRd/7S | ||||
| qNdQij9JGrVG7u8YJJscEQHqyHMYFVCEMjxmsge5BO6Vx5OWmUE3wXPzb5TbyTS4 | ||||
| +yg88g9rYTUXrzz+poCyKpaur45qBsdw35lJ8nq69VJj2xJLGQDwoTgGSXRuPciC | ||||
| 3OilQI+Ma+j8qQGJxJ8WJxISlf1cuhp+V4ZUd1lawlM5hAXyXmHRlH4pun4y+g7O | ||||
| O34+fE3pK25JjVCicMT/rC2A/sb95j/fHTzzJpbB70U0I50maTcIsOkyw6aiF//E | ||||
| 0ShTDz14x22SCMolUc6hxTDZvBB6yrcJHd7d9CCnFH2Sgo13QrtNJ/atXgm13HGh | ||||
| wBzRwK38XUGl/J4pJaxAupTVCPriStUM3m0EYHNelRRUE91pbyeGT0rvOuv00uLw | ||||
| Rj7K7hJZR8avTKWmKrVBVpq+gSojGW1DwBS0NiDNkZs0d/IjB1wkzczEgdZjXwID | ||||
| AQABo3QwcjAfBgNVHSMEGDAWgBTa+Ns6QzqlNvnTGszkouQQtZnVJDAdBgNVHSUE | ||||
| FjAUBggrBgEFBQcDAgYIKwYBBQUHAwEwEQYDVR0RBAowCIIGY2xpZW50MB0GA1Ud | ||||
| DgQWBBT1xg5sXkypRBwvCxBuyfoanaiZ5jANBgkqhkiG9w0BAQsFAAOCAgEAvUAz | ||||
| YwIjxY/0KHZDU8owdILVqKChzfLcy9OHNPyEI3TSOI8X6gNtBO+HE6r8aWGcC9vw | ||||
| zzeIsNQ3UEjvRWi2r+vUVbiPTbFdZboNDSZv6ZmGHxwd85VsjXRGoXV6koCT/9zi | ||||
| 9/lCM1DwqwYSwBphMJdRVFRUMluSYk1oHflGeA18xgGuts4eFivJwhabGm1AdVVQ | ||||
| /CYvqCuTxd/DCzWZBdyxYpDru64i/kyeJCt1pThKEFDWmpumFdBI4CxJ0OhxVSGp | ||||
| dOXzK+Y6ULepxCvi6/OpSog52jQ6PnNd1ghiYtq7yO1T4GQz65M1vtHHVvQ3gfBE | ||||
| AuKYQp6io7ypitRx+LpjsBQenyP4FFGfrq7pm90nLluOBOArfSdF0N+CP2wo/YFV | ||||
| 9BGf89OtvRi3BXCm2NXkE/Sc4We26tY8x7xNLOmNs8YOT0O3r/EQ690W9GIwRMx0 | ||||
| m0r/RXWn5V3o4Jib9r8eH9NzaDstD8g9dECcGfM4fHoM/DAGFaRrNcjMsS1APP3L | ||||
| jp7+BfBSXtrz9V6rVJ3CBLXlLK0AuSm7bqd1MJsGA9uMLpsVZIUA+KawcmPGdPU+ | ||||
| NxdpBCtzyurQSUyaTLtVqSeP35gMAwaNzUDph8Uh+vHz+kRwgXS19OQvTaud5LJu | ||||
| nQe4JNS+u5e2VDEBWUxt8NTpu6eShDN0iIEHtxA= | ||||
| bnRpazEUMBIGA1UECwwLU2VsZi1zaWduZWQwHhcNMjUwNDI3MTgzMTE3WhcNMjYw | ||||
| NDIzMTgzMTE3WjARMQ8wDQYDVQQDDAZjbGllbnQwggIiMA0GCSqGSIb3DQEBAQUA | ||||
| A4ICDwAwggIKAoICAQCdV+GEa7+7ito1i/z637OZW+0azv1kuF2aDwSzv+FJd+4L | ||||
| 6hCroRbVYTUFS3I3YwanOOZfau64xH0+pFM5Js8aREG68eqKBayx8vT27hyAOFhd | ||||
| giEVmSQJfla4ogvPie1rJ0HVOL7CiR72HDPQvz+9k1iDX3xQ/4sdAb3XurN13e+M | ||||
| Gtavhjiyqxmoo/H4WRd8BhD/BZQFWtaxWODDY8aKk5R7omw6Xf7aRv1BlHdE4Ucy | ||||
| Wozvpsj2Kz0l61rRUhiMlE0D9dpijgaRYFB+M7R2casH3CdhGQbBHTRiqBkZa6iq | ||||
| SDkTiTwNJQQJov8yPTsR+9P8OOuV6QN+DGm/FXJJFaPnsHw/HDy7EAbA1PcdbSyK | ||||
| XvJ8nVjdNhCEGbLGVSwAQLO+78hChVIN5YH+QSrP84YBSxKZYArnf4z2e9drqAN3 | ||||
| KmC26TkaUzkXnndnxOXBEIOSmyCdD4Dutg1XPE/bs8rA6rVGIR3pKXbCr29Z8hZn | ||||
| Cn9jbxwDwTX865ljR1Oc3dnIeCWa9AS/uHaSMdGlbGbDrt4Bj/nyyfu8xc034K/0 | ||||
| uPh3hF3FLWNAomRVZCvtuh/v7IEIQEgUbvQMWBhZJ8hu3HdtV8V9TIAryVKzEzGy | ||||
| Q72UHuQyK0njRDTmA/T+jn7P8GWOuf9eNdzd0gH0gcEuhCZFxPPRvUAeDuC7DQID | ||||
| AQABo1YwVDAdBgNVHSUEFjAUBggrBgEFBQcDAgYIKwYBBQUHAwEwFAYDVR0RAQH/ | ||||
| BAowCIIGY2xpZW50MB0GA1UdDgQWBBQ5KZwTD8+4CqLnbM/cBSXg8XeLXTANBgkq | ||||
| hkiG9w0BAQsFAAOCAgEABDkb3iyEOl1xKq1wxyRzf2L8qfPXAQw71FxMbgHArM+a | ||||
| e44wJGO3mZgPH0trOaJ+tuN6erB5YbZfsoX+xFacwskj9pKyb4QJLr/ENmJZRgyL | ||||
| wp5P6PB6IUJhvryvy/GxrG938YGmFtYQ+ikeJw5PWhB6218C1aZ9hsi94wZ1Zzrc | ||||
| Ry0q0D4QvIEZ0X2HL1Harc7gerE3VqhgQ7EWyImM+lCRtNDduwDQnZauwhr2r6cW | ||||
| XG4VTe1RCNsDA0xinXQE2Xf9voCd0Zf6wOOXJseQtrXpf+tG4N13cy5heF5ihed1 | ||||
| hDxSeki0KjTM+18kVVfVm4fzxf1Zg0gm54UlzWceIWh9EtnWMUV08H0D1M9YNmW8 | ||||
| hWTupk7M+jAw8Y+suHOe6/RLi0+fb9NSJpIpq4GqJ5UF2kerXHX0SvuAavoXyB0j | ||||
| CQrUXkRScEKOO2KAbVExSG56Ff7Ee8cRUAQ6rLC5pQRACq/R0sa6RcUsFPXul3Yv | ||||
| vbO2rTuArAUPkNVFknwkndheN4lOslRd1If02HunZETmsnal6p+nmuMWt2pQ2fDA | ||||
| vIguG54FyQ1T1IbF/QhfTEY62CQAebcgutnqqJHt9qe7Jr6ev57hMrJDEjotSzkY | ||||
| OhOVrcYqgLldr1nBqNVlIK/4VrDaWH8H5dNJ72gA9aMNVH4/bSTJhuO7cJkLnHw= | ||||
| -----END CERTIFICATE----- | ||||
|  | ||||
| @ -5,12 +5,7 @@ from django.urls import reverse | ||||
| from guardian.shortcuts import assign_perm | ||||
|  | ||||
| from authentik.core.models import User | ||||
| from authentik.core.tests.utils import ( | ||||
|     create_test_brand, | ||||
|     create_test_cert, | ||||
|     create_test_flow, | ||||
|     create_test_user, | ||||
| ) | ||||
| from authentik.core.tests.utils import create_test_brand, create_test_flow, create_test_user | ||||
| from authentik.crypto.models import CertificateKeyPair | ||||
| from authentik.enterprise.stages.mtls.models import ( | ||||
|     CertAttributes, | ||||
| @ -132,18 +127,6 @@ class MTLSStageTests(FlowTestCase): | ||||
|             self.assertEqual(res.status_code, 200) | ||||
|             self.assertStageResponse(res, self.flow, component="ak-stage-access-denied") | ||||
|  | ||||
|     def test_invalid_cert(self): | ||||
|         """Test invalid certificate""" | ||||
|         cert = create_test_cert() | ||||
|         with self.assertFlowFinishes() as plan: | ||||
|             res = self.client.get( | ||||
|                 reverse("authentik_api:flow-executor", kwargs={"flow_slug": self.flow.slug}), | ||||
|                 headers={"X-Forwarded-TLS-Client-Cert": quote_plus(cert.certificate_data)}, | ||||
|             ) | ||||
|             self.assertEqual(res.status_code, 200) | ||||
|             self.assertStageResponse(res, self.flow, component="ak-stage-access-denied") | ||||
|         self.assertNotIn(PLAN_CONTEXT_PENDING_USER, plan().context) | ||||
|  | ||||
|     def test_auth_no_user(self): | ||||
|         """Test auth with no user""" | ||||
|         User.objects.filter(username="client").delete() | ||||
| @ -217,12 +200,14 @@ class MTLSStageTests(FlowTestCase): | ||||
|         self.assertEqual( | ||||
|             plan().context[PLAN_CONTEXT_CERTIFICATE], | ||||
|             { | ||||
|                 "fingerprint_sha1": "52:39:ca:1e:3a:1f:78:3a:9f:26:3b:c2:84:99:48:68:99:99:81:8a", | ||||
|                 "fingerprint_sha1": ( | ||||
|                     "08:d4:a4:79:25:ca:c3:51:28:88:bb:30:c2:96:c3:44:5a:eb:18:07:84:ca:b4:75:27:74:61:19:8a:6a:af:fc" | ||||
|                 ), | ||||
|                 "fingerprint_sha256": ( | ||||
|                     "c1:07:8b:7c:e9:02:57:87:1e:92:e5:81:83:21:bc:92:c7:47:65:e3:97:fb:05:97:6f:36:9e:b5:31:77:98:b7" | ||||
|                     "08:d4:a4:79:25:ca:c3:51:28:88:bb:30:c2:96:c3:44:5a:eb:18:07:84:ca:b4:75:27:74:61:19:8a:6a:af:fc" | ||||
|                 ), | ||||
|                 "issuer": "OU=Self-signed,O=authentik,CN=authentik Test CA", | ||||
|                 "serial_number": "70153443448884702681996102271549704759327537151", | ||||
|                 "serial_number": "630532384467334865093173111400266136879266564943", | ||||
|                 "subject": "CN=client", | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
| @ -1,36 +1,28 @@ | ||||
| """Events API Views""" | ||||
|  | ||||
| from datetime import timedelta | ||||
| from json import loads | ||||
|  | ||||
| import django_filters | ||||
| from django.db.models import Count, ExpressionWrapper, F, QuerySet | ||||
| from django.db.models import DateTimeField as DjangoDateTimeField | ||||
| from django.db.models.aggregates import Count | ||||
| from django.db.models.fields.json import KeyTextTransform, KeyTransform | ||||
| from django.db.models.functions import TruncHour | ||||
| from django.db.models.functions import ExtractDay, ExtractHour | ||||
| from django.db.models.query_utils import Q | ||||
| from django.utils.timezone import now | ||||
| from drf_spectacular.types import OpenApiTypes | ||||
| from drf_spectacular.utils import OpenApiParameter, extend_schema | ||||
| from guardian.shortcuts import get_objects_for_user | ||||
| from rest_framework.decorators import action | ||||
| from rest_framework.fields import ChoiceField, DateTimeField, DictField, IntegerField | ||||
| from rest_framework.fields import DictField, IntegerField | ||||
| from rest_framework.request import Request | ||||
| from rest_framework.response import Response | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
|  | ||||
| from authentik.admin.api.metrics import CoordinateSerializer | ||||
| from authentik.core.api.object_types import TypeCreateSerializer | ||||
| from authentik.core.api.utils import ModelSerializer, PassiveSerializer | ||||
| from authentik.events.models import Event, EventAction | ||||
|  | ||||
|  | ||||
| class EventVolumeSerializer(PassiveSerializer): | ||||
|     """Count of events of action created on day""" | ||||
|  | ||||
|     action = ChoiceField(choices=EventAction.choices) | ||||
|     time = DateTimeField() | ||||
|     count = IntegerField() | ||||
|  | ||||
|  | ||||
| class EventSerializer(ModelSerializer): | ||||
|     """Event Serializer""" | ||||
|  | ||||
| @ -61,7 +53,7 @@ class EventsFilter(django_filters.FilterSet): | ||||
|     """Filter for events""" | ||||
|  | ||||
|     username = django_filters.CharFilter( | ||||
|         field_name="user", label="Username", method="filter_username" | ||||
|         field_name="user", lookup_expr="username", label="Username" | ||||
|     ) | ||||
|     context_model_pk = django_filters.CharFilter( | ||||
|         field_name="context", | ||||
| @ -86,19 +78,12 @@ class EventsFilter(django_filters.FilterSet): | ||||
|         field_name="action", | ||||
|         lookup_expr="icontains", | ||||
|     ) | ||||
|     actions = django_filters.MultipleChoiceFilter( | ||||
|         field_name="action", | ||||
|         choices=EventAction.choices, | ||||
|     ) | ||||
|     brand_name = django_filters.CharFilter( | ||||
|         field_name="brand", | ||||
|         lookup_expr="name", | ||||
|         label="Brand name", | ||||
|     ) | ||||
|  | ||||
|     def filter_username(self, queryset, name, value): | ||||
|         return queryset.filter(Q(user__username=value) | Q(context__username=value)) | ||||
|  | ||||
|     def filter_context_model_pk(self, queryset, name, value): | ||||
|         """Because we store the PK as UUID.hex, | ||||
|         we need to remove the dashes that a client may send. We can't use a | ||||
| @ -132,22 +117,6 @@ class EventViewSet(ModelViewSet): | ||||
|     ] | ||||
|     filterset_class = EventsFilter | ||||
|  | ||||
|     def get_ql_fields(self): | ||||
|         from djangoql.schema import DateTimeField, StrField | ||||
|  | ||||
|         from authentik.enterprise.search.fields import ChoiceSearchField, JSONSearchField | ||||
|  | ||||
|         return [ | ||||
|             ChoiceSearchField(Event, "action"), | ||||
|             StrField(Event, "event_uuid"), | ||||
|             StrField(Event, "app", suggest_options=True), | ||||
|             StrField(Event, "client_ip"), | ||||
|             JSONSearchField(Event, "user", suggest_nested=False), | ||||
|             JSONSearchField(Event, "brand", suggest_nested=False), | ||||
|             JSONSearchField(Event, "context", suggest_nested=False), | ||||
|             DateTimeField(Event, "created", suggest_options=True), | ||||
|         ] | ||||
|  | ||||
|     @extend_schema( | ||||
|         methods=["GET"], | ||||
|         responses={200: EventTopPerUserSerializer(many=True)}, | ||||
| @ -187,37 +156,45 @@ class EventViewSet(ModelViewSet): | ||||
|         return Response(EventTopPerUserSerializer(instance=events, many=True).data) | ||||
|  | ||||
|     @extend_schema( | ||||
|         responses={200: EventVolumeSerializer(many=True)}, | ||||
|         parameters=[ | ||||
|             OpenApiParameter( | ||||
|                 "history_days", | ||||
|                 type=OpenApiTypes.NUMBER, | ||||
|                 location=OpenApiParameter.QUERY, | ||||
|                 required=False, | ||||
|                 default=7, | ||||
|             ), | ||||
|         ], | ||||
|         responses={200: CoordinateSerializer(many=True)}, | ||||
|     ) | ||||
|     @action(detail=False, methods=["GET"], pagination_class=None) | ||||
|     def volume(self, request: Request) -> Response: | ||||
|         """Get event volume for specified filters and timeframe""" | ||||
|         queryset: QuerySet[Event] = self.filter_queryset(self.get_queryset()) | ||||
|         delta = timedelta(days=7) | ||||
|         time_delta = request.query_params.get("history_days", 7) | ||||
|         if time_delta: | ||||
|             delta = timedelta(days=min(int(time_delta), 60)) | ||||
|         queryset = self.filter_queryset(self.get_queryset()) | ||||
|         return Response(queryset.get_events_per(timedelta(days=7), ExtractHour, 7 * 3)) | ||||
|  | ||||
|     @extend_schema( | ||||
|         responses={200: CoordinateSerializer(many=True)}, | ||||
|         filters=[], | ||||
|         parameters=[ | ||||
|             OpenApiParameter( | ||||
|                 "action", | ||||
|                 type=OpenApiTypes.STR, | ||||
|                 location=OpenApiParameter.QUERY, | ||||
|                 required=False, | ||||
|             ), | ||||
|             OpenApiParameter( | ||||
|                 "query", | ||||
|                 type=OpenApiTypes.STR, | ||||
|                 location=OpenApiParameter.QUERY, | ||||
|                 required=False, | ||||
|             ), | ||||
|         ], | ||||
|     ) | ||||
|     @action(detail=False, methods=["GET"], pagination_class=None) | ||||
|     def per_month(self, request: Request): | ||||
|         """Get the count of events per month""" | ||||
|         filtered_action = request.query_params.get("action", EventAction.LOGIN) | ||||
|         try: | ||||
|             query = loads(request.query_params.get("query", "{}")) | ||||
|         except ValueError: | ||||
|             return Response(status=400) | ||||
|         return Response( | ||||
|             queryset.filter(created__gte=now() - delta) | ||||
|             .annotate(hour=TruncHour("created")) | ||||
|             .annotate( | ||||
|                 time=ExpressionWrapper( | ||||
|                     F("hour") - (F("hour__hour") % 6) * timedelta(hours=1), | ||||
|                     output_field=DjangoDateTimeField(), | ||||
|                 ) | ||||
|             ) | ||||
|             .values("time", "action") | ||||
|             .annotate(count=Count("pk")) | ||||
|             .order_by("time", "action") | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event") | ||||
|             .filter(action=filtered_action) | ||||
|             .filter(**query) | ||||
|             .get_events_per(timedelta(weeks=4), ExtractDay, 30) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema(responses={200: TypeCreateSerializer(many=True)}) | ||||
|  | ||||
| @ -11,7 +11,7 @@ from authentik.events.models import NotificationRule | ||||
| class NotificationRuleSerializer(ModelSerializer): | ||||
|     """NotificationRule Serializer""" | ||||
|  | ||||
|     destination_group_obj = GroupSerializer(read_only=True, source="destination_group") | ||||
|     group_obj = GroupSerializer(read_only=True, source="group") | ||||
|  | ||||
|     class Meta: | ||||
|         model = NotificationRule | ||||
| @ -20,9 +20,8 @@ class NotificationRuleSerializer(ModelSerializer): | ||||
|             "name", | ||||
|             "transports", | ||||
|             "severity", | ||||
|             "destination_group", | ||||
|             "destination_group_obj", | ||||
|             "destination_event_user", | ||||
|             "group", | ||||
|             "group_obj", | ||||
|         ] | ||||
|  | ||||
|  | ||||
| @ -31,6 +30,6 @@ class NotificationRuleViewSet(UsedByMixin, ModelViewSet): | ||||
|  | ||||
|     queryset = NotificationRule.objects.all() | ||||
|     serializer_class = NotificationRuleSerializer | ||||
|     filterset_fields = ["name", "severity", "destination_group__name"] | ||||
|     filterset_fields = ["name", "severity", "group__name"] | ||||
|     ordering = ["name"] | ||||
|     search_fields = ["name", "destination_group__name"] | ||||
|     search_fields = ["name", "group__name"] | ||||
|  | ||||
| @ -15,13 +15,13 @@ class MMDBContextProcessor(EventContextProcessor): | ||||
|         self.reader: Reader | None = None | ||||
|         self._last_mtime: float = 0.0 | ||||
|         self.logger = get_logger() | ||||
|         self.load() | ||||
|         self.open() | ||||
|  | ||||
|     def path(self) -> str | None: | ||||
|         """Get the path to the MMDB file to load""" | ||||
|         raise NotImplementedError | ||||
|  | ||||
|     def load(self): | ||||
|     def open(self): | ||||
|         """Get GeoIP Reader, if configured, otherwise none""" | ||||
|         path = self.path() | ||||
|         if path == "" or not path: | ||||
| @ -44,7 +44,7 @@ class MMDBContextProcessor(EventContextProcessor): | ||||
|             diff = self._last_mtime < mtime | ||||
|             if diff > 0: | ||||
|                 self.logger.info("Found new MMDB Database, reopening", diff=diff, path=path) | ||||
|                 self.load() | ||||
|                 self.open() | ||||
|         except OSError as exc: | ||||
|             self.logger.warning("Failed to check MMDB age", exc=exc) | ||||
|  | ||||
|  | ||||
| @ -1,26 +0,0 @@ | ||||
| # Generated by Django 5.1.11 on 2025-06-16 23:21 | ||||
|  | ||||
| from django.db import migrations, models | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
|     dependencies = [ | ||||
|         ("authentik_events", "0009_remove_notificationtransport_webhook_mapping_and_more"), | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.RenameField( | ||||
|             model_name="notificationrule", | ||||
|             old_name="group", | ||||
|             new_name="destination_group", | ||||
|         ), | ||||
|         migrations.AddField( | ||||
|             model_name="notificationrule", | ||||
|             name="destination_event_user", | ||||
|             field=models.BooleanField( | ||||
|                 default=False, | ||||
|                 help_text="When enabled, notification will be sent to user the user that triggered the event.When destination_group is configured, notification is sent to both.", | ||||
|             ), | ||||
|         ), | ||||
|     ] | ||||
| @ -1,16 +1,21 @@ | ||||
| """authentik events models""" | ||||
|  | ||||
| from collections.abc import Generator | ||||
| import time | ||||
| from collections import Counter | ||||
| from datetime import timedelta | ||||
| from difflib import get_close_matches | ||||
| from functools import lru_cache | ||||
| from inspect import currentframe | ||||
| from smtplib import SMTPException | ||||
| from typing import Any | ||||
| from uuid import uuid4 | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.db import connection, models | ||||
| from django.db.models import Count, ExpressionWrapper, F | ||||
| from django.db.models.fields import DurationField | ||||
| from django.db.models.functions import Extract | ||||
| from django.db.models.manager import Manager | ||||
| from django.db.models.query import QuerySet | ||||
| from django.http import HttpRequest | ||||
| from django.http.request import QueryDict | ||||
| from django.utils.timezone import now | ||||
| @ -119,6 +124,60 @@ class EventAction(models.TextChoices): | ||||
|     CUSTOM_PREFIX = "custom_" | ||||
|  | ||||
|  | ||||
| class EventQuerySet(QuerySet): | ||||
|     """Custom events query set with helper functions""" | ||||
|  | ||||
|     def get_events_per( | ||||
|         self, | ||||
|         time_since: timedelta, | ||||
|         extract: Extract, | ||||
|         data_points: int, | ||||
|     ) -> list[dict[str, int]]: | ||||
|         """Get event count by hour in the last day, fill with zeros""" | ||||
|         _now = now() | ||||
|         max_since = timedelta(days=60) | ||||
|         # Allow maximum of 60 days to limit load | ||||
|         if time_since.total_seconds() > max_since.total_seconds(): | ||||
|             time_since = max_since | ||||
|         date_from = _now - time_since | ||||
|         result = ( | ||||
|             self.filter(created__gte=date_from) | ||||
|             .annotate(age=ExpressionWrapper(_now - F("created"), output_field=DurationField())) | ||||
|             .annotate(age_interval=extract("age")) | ||||
|             .values("age_interval") | ||||
|             .annotate(count=Count("pk")) | ||||
|             .order_by("age_interval") | ||||
|         ) | ||||
|         data = Counter({int(d["age_interval"]): d["count"] for d in result}) | ||||
|         results = [] | ||||
|         interval_delta = time_since / data_points | ||||
|         for interval in range(1, -data_points, -1): | ||||
|             results.append( | ||||
|                 { | ||||
|                     "x_cord": time.mktime((_now + (interval_delta * interval)).timetuple()) * 1000, | ||||
|                     "y_cord": data[interval * -1], | ||||
|                 } | ||||
|             ) | ||||
|         return results | ||||
|  | ||||
|  | ||||
| class EventManager(Manager): | ||||
|     """Custom helper methods for Events""" | ||||
|  | ||||
|     def get_queryset(self) -> QuerySet: | ||||
|         """use custom queryset""" | ||||
|         return EventQuerySet(self.model, using=self._db) | ||||
|  | ||||
|     def get_events_per( | ||||
|         self, | ||||
|         time_since: timedelta, | ||||
|         extract: Extract, | ||||
|         data_points: int, | ||||
|     ) -> list[dict[str, int]]: | ||||
|         """Wrap method from queryset""" | ||||
|         return self.get_queryset().get_events_per(time_since, extract, data_points) | ||||
|  | ||||
|  | ||||
| class Event(SerializerModel, ExpiringModel): | ||||
|     """An individual Audit/Metrics/Notification/Error Event""" | ||||
|  | ||||
| @ -134,6 +193,8 @@ class Event(SerializerModel, ExpiringModel): | ||||
|     # Shadow the expires attribute from ExpiringModel to override the default duration | ||||
|     expires = models.DateTimeField(default=default_event_duration) | ||||
|  | ||||
|     objects = EventManager() | ||||
|  | ||||
|     @staticmethod | ||||
|     def _get_app_from_request(request: HttpRequest) -> str: | ||||
|         if not isinstance(request, HttpRequest): | ||||
| @ -549,7 +610,7 @@ class NotificationRule(SerializerModel, PolicyBindingModel): | ||||
|         default=NotificationSeverity.NOTICE, | ||||
|         help_text=_("Controls which severity level the created notifications will have."), | ||||
|     ) | ||||
|     destination_group = models.ForeignKey( | ||||
|     group = models.ForeignKey( | ||||
|         Group, | ||||
|         help_text=_( | ||||
|             "Define which group of users this notification should be sent and shown to. " | ||||
| @ -559,19 +620,6 @@ class NotificationRule(SerializerModel, PolicyBindingModel): | ||||
|         blank=True, | ||||
|         on_delete=models.SET_NULL, | ||||
|     ) | ||||
|     destination_event_user = models.BooleanField( | ||||
|         default=False, | ||||
|         help_text=_( | ||||
|             "When enabled, notification will be sent to user the user that triggered the event." | ||||
|             "When destination_group is configured, notification is sent to both." | ||||
|         ), | ||||
|     ) | ||||
|  | ||||
|     def destination_users(self, event: Event) -> Generator[User, Any]: | ||||
|         if self.destination_event_user and event.user.get("pk"): | ||||
|             yield User(pk=event.user.get("pk")) | ||||
|         if self.destination_group: | ||||
|             yield from self.destination_group.users.all() | ||||
|  | ||||
|     @property | ||||
|     def serializer(self) -> type[Serializer]: | ||||
|  | ||||
| @ -68,10 +68,14 @@ def event_trigger_handler(event_uuid: str, trigger_name: str): | ||||
|     if not result.passing: | ||||
|         return | ||||
|  | ||||
|     if not trigger.group: | ||||
|         LOGGER.debug("e(trigger): trigger has no group", trigger=trigger) | ||||
|         return | ||||
|  | ||||
|     LOGGER.debug("e(trigger): event trigger matched", trigger=trigger) | ||||
|     # Create the notification objects | ||||
|     for transport in trigger.transports.all(): | ||||
|         for user in trigger.destination_users(event): | ||||
|         for user in trigger.group.users.all(): | ||||
|             LOGGER.debug("created notification") | ||||
|             notification_transport.apply_async( | ||||
|                 args=[ | ||||
|  | ||||
| @ -6,7 +6,6 @@ from django.urls import reverse | ||||
| from rest_framework.test import APITestCase | ||||
|  | ||||
| from authentik.core.models import Group, User | ||||
| from authentik.core.tests.utils import create_test_user | ||||
| from authentik.events.models import ( | ||||
|     Event, | ||||
|     EventAction, | ||||
| @ -35,7 +34,7 @@ class TestEventsNotifications(APITestCase): | ||||
|     def test_trigger_empty(self): | ||||
|         """Test trigger without any policies attached""" | ||||
|         transport = NotificationTransport.objects.create(name=generate_id()) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_group=self.group) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), group=self.group) | ||||
|         trigger.transports.add(transport) | ||||
|         trigger.save() | ||||
|  | ||||
| @ -47,7 +46,7 @@ class TestEventsNotifications(APITestCase): | ||||
|     def test_trigger_single(self): | ||||
|         """Test simple transport triggering""" | ||||
|         transport = NotificationTransport.objects.create(name=generate_id()) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_group=self.group) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), group=self.group) | ||||
|         trigger.transports.add(transport) | ||||
|         trigger.save() | ||||
|         matcher = EventMatcherPolicy.objects.create( | ||||
| @ -60,25 +59,6 @@ class TestEventsNotifications(APITestCase): | ||||
|             Event.new(EventAction.CUSTOM_PREFIX).save() | ||||
|         self.assertEqual(execute_mock.call_count, 1) | ||||
|  | ||||
|     def test_trigger_event_user(self): | ||||
|         """Test trigger with event user""" | ||||
|         user = create_test_user() | ||||
|         transport = NotificationTransport.objects.create(name=generate_id()) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_event_user=True) | ||||
|         trigger.transports.add(transport) | ||||
|         trigger.save() | ||||
|         matcher = EventMatcherPolicy.objects.create( | ||||
|             name="matcher", action=EventAction.CUSTOM_PREFIX | ||||
|         ) | ||||
|         PolicyBinding.objects.create(target=trigger, policy=matcher, order=0) | ||||
|  | ||||
|         execute_mock = MagicMock() | ||||
|         with patch("authentik.events.models.NotificationTransport.send", execute_mock): | ||||
|             Event.new(EventAction.CUSTOM_PREFIX).set_user(user).save() | ||||
|         self.assertEqual(execute_mock.call_count, 1) | ||||
|         notification: Notification = execute_mock.call_args[0][0] | ||||
|         self.assertEqual(notification.user, user) | ||||
|  | ||||
|     def test_trigger_no_group(self): | ||||
|         """Test trigger without group""" | ||||
|         trigger = NotificationRule.objects.create(name=generate_id()) | ||||
| @ -96,7 +76,7 @@ class TestEventsNotifications(APITestCase): | ||||
|         """Test Policy error which would cause recursion""" | ||||
|         transport = NotificationTransport.objects.create(name=generate_id()) | ||||
|         NotificationRule.objects.filter(name__startswith="default").delete() | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_group=self.group) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), group=self.group) | ||||
|         trigger.transports.add(transport) | ||||
|         trigger.save() | ||||
|         matcher = EventMatcherPolicy.objects.create( | ||||
| @ -119,7 +99,7 @@ class TestEventsNotifications(APITestCase): | ||||
|  | ||||
|         transport = NotificationTransport.objects.create(name=generate_id(), send_once=True) | ||||
|         NotificationRule.objects.filter(name__startswith="default").delete() | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_group=self.group) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), group=self.group) | ||||
|         trigger.transports.add(transport) | ||||
|         trigger.save() | ||||
|         matcher = EventMatcherPolicy.objects.create( | ||||
| @ -143,7 +123,7 @@ class TestEventsNotifications(APITestCase): | ||||
|             name=generate_id(), webhook_mapping_body=mapping, mode=TransportMode.LOCAL | ||||
|         ) | ||||
|         NotificationRule.objects.filter(name__startswith="default").delete() | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), destination_group=self.group) | ||||
|         trigger = NotificationRule.objects.create(name=generate_id(), group=self.group) | ||||
|         trigger.transports.add(transport) | ||||
|         matcher = EventMatcherPolicy.objects.create( | ||||
|             name="matcher", action=EventAction.CUSTOM_PREFIX | ||||
|  | ||||
| @ -1,18 +0,0 @@ | ||||
| # Generated by Django 5.1.9 on 2025-05-27 12:52 | ||||
|  | ||||
| from django.db import migrations, models | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
|     dependencies = [ | ||||
|         ("authentik_flows", "0027_auto_20231028_1424"), | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.AddField( | ||||
|             model_name="flowtoken", | ||||
|             name="revoke_on_execution", | ||||
|             field=models.BooleanField(default=True), | ||||
|         ), | ||||
|     ] | ||||
| @ -303,10 +303,9 @@ class FlowToken(Token): | ||||
|  | ||||
|     flow = models.ForeignKey(Flow, on_delete=models.CASCADE) | ||||
|     _plan = models.TextField() | ||||
|     revoke_on_execution = models.BooleanField(default=True) | ||||
|  | ||||
|     @staticmethod | ||||
|     def pickle(plan: "FlowPlan") -> str: | ||||
|     def pickle(plan) -> str: | ||||
|         """Pickle into string""" | ||||
|         data = dumps(plan) | ||||
|         return b64encode(data).decode() | ||||
|  | ||||
| @ -99,10 +99,9 @@ class ChallengeStageView(StageView): | ||||
|             self.logger.debug("Got StageInvalidException", exc=exc) | ||||
|             return self.executor.stage_invalid() | ||||
|         if not challenge.is_valid(): | ||||
|             self.logger.error( | ||||
|             self.logger.warning( | ||||
|                 "f(ch): Invalid challenge", | ||||
|                 errors=challenge.errors, | ||||
|                 challenge=challenge.data, | ||||
|             ) | ||||
|         return HttpChallengeResponse(challenge) | ||||
|  | ||||
|  | ||||
| @ -146,7 +146,6 @@ class FlowExecutorView(APIView): | ||||
|         except (AttributeError, EOFError, ImportError, IndexError) as exc: | ||||
|             LOGGER.warning("f(exec): Failed to restore token plan", exc=exc) | ||||
|         finally: | ||||
|             if token.revoke_on_execution: | ||||
|             token.delete() | ||||
|         if not isinstance(plan, FlowPlan): | ||||
|             return None | ||||
|  | ||||
| @ -81,6 +81,7 @@ debugger: false | ||||
|  | ||||
| log_level: info | ||||
|  | ||||
| session_storage: cache | ||||
| sessions: | ||||
|   unauthenticated_age: days=1 | ||||
|  | ||||
|  | ||||
| @ -1,7 +1,6 @@ | ||||
| from collections.abc import Callable | ||||
| from dataclasses import asdict | ||||
|  | ||||
| from celery import group | ||||
| from celery.exceptions import Retry | ||||
| from celery.result import allow_join_result | ||||
| from django.core.paginator import Paginator | ||||
| @ -83,41 +82,21 @@ class SyncTasks: | ||||
|                 self.logger.debug("Failed to acquire sync lock, skipping", provider=provider.name) | ||||
|                 return | ||||
|             try: | ||||
|                 messages.append(_("Syncing users")) | ||||
|                 user_results = ( | ||||
|                     group( | ||||
|                         [ | ||||
|                             sync_objects.signature( | ||||
|                 for page in users_paginator.page_range: | ||||
|                     messages.append(_("Syncing page {page} of users".format(page=page))) | ||||
|                     for msg in sync_objects.apply_async( | ||||
|                         args=(class_to_path(User), page, provider_pk), | ||||
|                         time_limit=PAGE_TIMEOUT, | ||||
|                         soft_time_limit=PAGE_TIMEOUT, | ||||
|                             ) | ||||
|                             for page in users_paginator.page_range | ||||
|                         ] | ||||
|                     ) | ||||
|                     .apply_async() | ||||
|                     .get() | ||||
|                 ) | ||||
|                 for result in user_results: | ||||
|                     for msg in result: | ||||
|                     ).get(): | ||||
|                         messages.append(LogEvent(**msg)) | ||||
|                 messages.append(_("Syncing groups")) | ||||
|                 group_results = ( | ||||
|                     group( | ||||
|                         [ | ||||
|                             sync_objects.signature( | ||||
|                 for page in groups_paginator.page_range: | ||||
|                     messages.append(_("Syncing page {page} of groups".format(page=page))) | ||||
|                     for msg in sync_objects.apply_async( | ||||
|                         args=(class_to_path(Group), page, provider_pk), | ||||
|                         time_limit=PAGE_TIMEOUT, | ||||
|                         soft_time_limit=PAGE_TIMEOUT, | ||||
|                             ) | ||||
|                             for page in groups_paginator.page_range | ||||
|                         ] | ||||
|                     ) | ||||
|                     .apply_async() | ||||
|                     .get() | ||||
|                 ) | ||||
|                 for result in group_results: | ||||
|                     for msg in result: | ||||
|                     ).get(): | ||||
|                         messages.append(LogEvent(**msg)) | ||||
|             except TransientSyncException as exc: | ||||
|                 self.logger.warning("transient sync exception", exc=exc) | ||||
| @ -130,7 +109,7 @@ class SyncTasks: | ||||
|     def sync_objects( | ||||
|         self, object_type: str, page: int, provider_pk: int, override_dry_run=False, **filter | ||||
|     ): | ||||
|         _object_type: type[Model] = path_to_class(object_type) | ||||
|         _object_type = path_to_class(object_type) | ||||
|         self.logger = get_logger().bind( | ||||
|             provider_type=class_to_path(self._provider_model), | ||||
|             provider_pk=provider_pk, | ||||
| @ -153,19 +132,6 @@ class SyncTasks: | ||||
|             self.logger.debug("starting discover") | ||||
|             client.discover() | ||||
|         self.logger.debug("starting sync for page", page=page) | ||||
|         messages.append( | ||||
|             asdict( | ||||
|                 LogEvent( | ||||
|                     _( | ||||
|                         "Syncing page {page} of {object_type}".format( | ||||
|                             page=page, object_type=_object_type._meta.verbose_name_plural | ||||
|                         ) | ||||
|                     ), | ||||
|                     log_level="info", | ||||
|                     logger=f"{provider._meta.verbose_name}@{object_type}", | ||||
|                 ) | ||||
|             ) | ||||
|         ) | ||||
|         for obj in paginator.page(page).object_list: | ||||
|             obj: Model | ||||
|             try: | ||||
|  | ||||
| @ -37,9 +37,6 @@ class WebsocketMessageInstruction(IntEnum): | ||||
|     # Provider specific message | ||||
|     PROVIDER_SPECIFIC = 3 | ||||
|  | ||||
|     # Session ended | ||||
|     SESSION_END = 4 | ||||
|  | ||||
|  | ||||
| @dataclass(slots=True) | ||||
| class WebsocketMessage: | ||||
| @ -148,14 +145,6 @@ class OutpostConsumer(JsonWebsocketConsumer): | ||||
|             asdict(WebsocketMessage(instruction=WebsocketMessageInstruction.TRIGGER_UPDATE)) | ||||
|         ) | ||||
|  | ||||
|     def event_session_end(self, event): | ||||
|         """Event handler which is called when a session is ended""" | ||||
|         self.send_json( | ||||
|             asdict( | ||||
|                 WebsocketMessage(instruction=WebsocketMessageInstruction.SESSION_END, args=event) | ||||
|             ) | ||||
|         ) | ||||
|  | ||||
|     def event_provider_specific(self, event): | ||||
|         """Event handler which can be called by provider-specific | ||||
|         implementations to send specific messages to the outpost""" | ||||
|  | ||||
| @ -1,24 +1,17 @@ | ||||
| """authentik outpost signals""" | ||||
|  | ||||
| from django.contrib.auth.signals import user_logged_out | ||||
| from django.core.cache import cache | ||||
| from django.db.models import Model | ||||
| from django.db.models.signals import m2m_changed, post_save, pre_delete, pre_save | ||||
| from django.dispatch import receiver | ||||
| from django.http import HttpRequest | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.brands.models import Brand | ||||
| from authentik.core.models import AuthenticatedSession, Provider, User | ||||
| from authentik.core.models import Provider | ||||
| from authentik.crypto.models import CertificateKeyPair | ||||
| from authentik.lib.utils.reflection import class_to_path | ||||
| from authentik.outposts.models import Outpost, OutpostServiceConnection | ||||
| from authentik.outposts.tasks import ( | ||||
|     CACHE_KEY_OUTPOST_DOWN, | ||||
|     outpost_controller, | ||||
|     outpost_post_save, | ||||
|     outpost_session_end, | ||||
| ) | ||||
| from authentik.outposts.tasks import CACHE_KEY_OUTPOST_DOWN, outpost_controller, outpost_post_save | ||||
|  | ||||
| LOGGER = get_logger() | ||||
| UPDATE_TRIGGERING_MODELS = ( | ||||
| @ -80,17 +73,3 @@ def pre_delete_cleanup(sender, instance: Outpost, **_): | ||||
|     instance.user.delete() | ||||
|     cache.set(CACHE_KEY_OUTPOST_DOWN % instance.pk.hex, instance) | ||||
|     outpost_controller.delay(instance.pk.hex, action="down", from_cache=True) | ||||
|  | ||||
|  | ||||
| @receiver(user_logged_out) | ||||
| def logout_revoke_direct(sender: type[User], request: HttpRequest, **_): | ||||
|     """Catch logout by direct logout and forward to providers""" | ||||
|     if not request.session or not request.session.session_key: | ||||
|         return | ||||
|     outpost_session_end.delay(request.session.session_key) | ||||
|  | ||||
|  | ||||
| @receiver(pre_delete, sender=AuthenticatedSession) | ||||
| def logout_revoke(sender: type[AuthenticatedSession], instance: AuthenticatedSession, **_): | ||||
|     """Catch logout by expiring sessions being deleted""" | ||||
|     outpost_session_end.delay(instance.session.session_key) | ||||
|  | ||||
| @ -1,6 +1,5 @@ | ||||
| """outpost tasks""" | ||||
|  | ||||
| from hashlib import sha256 | ||||
| from os import R_OK, access | ||||
| from pathlib import Path | ||||
| from socket import gethostname | ||||
| @ -50,11 +49,6 @@ LOGGER = get_logger() | ||||
| CACHE_KEY_OUTPOST_DOWN = "goauthentik.io/outposts/teardown/%s" | ||||
|  | ||||
|  | ||||
| def hash_session_key(session_key: str) -> str: | ||||
|     """Hash the session key for sending session end signals""" | ||||
|     return sha256(session_key.encode("ascii")).hexdigest() | ||||
|  | ||||
|  | ||||
| def controller_for_outpost(outpost: Outpost) -> type[BaseController] | None: | ||||
|     """Get a controller for the outpost, when a service connection is defined""" | ||||
|     if not outpost.service_connection: | ||||
| @ -295,20 +289,3 @@ def outpost_connection_discovery(self: SystemTask): | ||||
|                 url=unix_socket_path, | ||||
|             ) | ||||
|     self.set_status(TaskStatus.SUCCESSFUL, *messages) | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task() | ||||
| def outpost_session_end(session_id: str): | ||||
|     """Update outpost instances connected to a single outpost""" | ||||
|     layer = get_channel_layer() | ||||
|     hashed_session_id = hash_session_key(session_id) | ||||
|     for outpost in Outpost.objects.all(): | ||||
|         LOGGER.info("Sending session end signal to outpost", outpost=outpost) | ||||
|         group = OUTPOST_GROUP % {"outpost_pk": str(outpost.pk)} | ||||
|         async_to_sync(layer.group_send)( | ||||
|             group, | ||||
|             { | ||||
|                 "type": "event.session.end", | ||||
|                 "session_id": hashed_session_id, | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
| @ -38,7 +38,6 @@ class TestOutpostWS(TransactionTestCase): | ||||
|         ) | ||||
|         connected, _ = await communicator.connect() | ||||
|         self.assertFalse(connected) | ||||
|         await communicator.disconnect() | ||||
|  | ||||
|     async def test_auth_valid(self): | ||||
|         """Test auth with token""" | ||||
| @ -49,7 +48,6 @@ class TestOutpostWS(TransactionTestCase): | ||||
|         ) | ||||
|         connected, _ = await communicator.connect() | ||||
|         self.assertTrue(connected) | ||||
|         await communicator.disconnect() | ||||
|  | ||||
|     async def test_send(self): | ||||
|         """Test sending of Hello""" | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """Authentik policy dummy app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPolicyDummyConfig(ManagedAppConfig): | ||||
| class AuthentikPolicyDummyConfig(AppConfig): | ||||
|     """Authentik policy_dummy app config""" | ||||
|  | ||||
|     name = "authentik.policies.dummy" | ||||
|     label = "authentik_policies_dummy" | ||||
|     verbose_name = "authentik Policies.Dummy" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """authentik Event Matcher policy app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPoliciesEventMatcherConfig(ManagedAppConfig): | ||||
| class AuthentikPoliciesEventMatcherConfig(AppConfig): | ||||
|     """authentik Event Matcher policy app config""" | ||||
|  | ||||
|     name = "authentik.policies.event_matcher" | ||||
|     label = "authentik_policies_event_matcher" | ||||
|     verbose_name = "authentik Policies.Event Matcher" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """Authentik policy_expiry app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPolicyExpiryConfig(ManagedAppConfig): | ||||
| class AuthentikPolicyExpiryConfig(AppConfig): | ||||
|     """Authentik policy_expiry app config""" | ||||
|  | ||||
|     name = "authentik.policies.expiry" | ||||
|     label = "authentik_policies_expiry" | ||||
|     verbose_name = "authentik Policies.Expiry" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """Authentik policy_expression app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPolicyExpressionConfig(ManagedAppConfig): | ||||
| class AuthentikPolicyExpressionConfig(AppConfig): | ||||
|     """Authentik policy_expression app config""" | ||||
|  | ||||
|     name = "authentik.policies.expression" | ||||
|     label = "authentik_policies_expression" | ||||
|     verbose_name = "authentik Policies.Expression" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """Authentik policy geoip app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPolicyGeoIPConfig(ManagedAppConfig): | ||||
| class AuthentikPolicyGeoIPConfig(AppConfig): | ||||
|     """Authentik policy_geoip app config""" | ||||
|  | ||||
|     name = "authentik.policies.geoip" | ||||
|     label = "authentik_policies_geoip" | ||||
|     verbose_name = "authentik Policies.GeoIP" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """authentik Password policy app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikPoliciesPasswordConfig(ManagedAppConfig): | ||||
| class AuthentikPoliciesPasswordConfig(AppConfig): | ||||
|     """authentik Password policy app config""" | ||||
|  | ||||
|     name = "authentik.policies.password" | ||||
|     label = "authentik_policies_password" | ||||
|     verbose_name = "authentik Policies.Password" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """authentik ldap provider app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikProviderLDAPConfig(ManagedAppConfig): | ||||
| class AuthentikProviderLDAPConfig(AppConfig): | ||||
|     """authentik ldap provider app config""" | ||||
|  | ||||
|     name = "authentik.providers.ldap" | ||||
|     label = "authentik_providers_ldap" | ||||
|     verbose_name = "authentik Providers.LDAP" | ||||
|     default = True | ||||
|  | ||||
| @ -7,8 +7,10 @@ from django.db import migrations | ||||
|  | ||||
|  | ||||
| def migrate_search_group(apps: Apps, schema_editor: BaseDatabaseSchemaEditor): | ||||
|     from authentik.core.models import User | ||||
|     from django.apps import apps as real_apps | ||||
|     from django.contrib.auth.management import create_permissions | ||||
|     from guardian.shortcuts import UserObjectPermission | ||||
|  | ||||
|     db_alias = schema_editor.connection.alias | ||||
|  | ||||
|  | ||||
| @ -50,4 +50,3 @@ AMR_PASSWORD = "pwd"  # nosec | ||||
| AMR_MFA = "mfa" | ||||
| AMR_OTP = "otp" | ||||
| AMR_WEBAUTHN = "user" | ||||
| AMR_SMART_CARD = "sc" | ||||
|  | ||||
| @ -16,7 +16,6 @@ from authentik.providers.oauth2.constants import ( | ||||
|     ACR_AUTHENTIK_DEFAULT, | ||||
|     AMR_MFA, | ||||
|     AMR_PASSWORD, | ||||
|     AMR_SMART_CARD, | ||||
|     AMR_WEBAUTHN, | ||||
| ) | ||||
| from authentik.stages.password.stage import PLAN_CONTEXT_METHOD, PLAN_CONTEXT_METHOD_ARGS | ||||
| @ -140,9 +139,8 @@ class IDToken: | ||||
|                 amr.append(AMR_PASSWORD) | ||||
|             if method == "auth_webauthn_pwl": | ||||
|                 amr.append(AMR_WEBAUTHN) | ||||
|             if "certificate" in method_args: | ||||
|                 amr.append(AMR_SMART_CARD) | ||||
|             if "mfa_devices" in method_args: | ||||
|                 if len(amr) > 0: | ||||
|                     amr.append(AMR_MFA) | ||||
|             if amr: | ||||
|                 id_token.amr = amr | ||||
|  | ||||
| @ -387,7 +387,8 @@ class TestAuthorize(OAuthTestCase): | ||||
|             self.assertEqual( | ||||
|                 response.url, | ||||
|                 ( | ||||
|                     f"http://localhost#id_token={provider.encode(token.id_token.to_dict())}" | ||||
|                     f"http://localhost#access_token={token.token}" | ||||
|                     f"&id_token={provider.encode(token.id_token.to_dict())}" | ||||
|                     f"&token_type={TOKEN_TYPE}" | ||||
|                     f"&expires_in={int(expires)}&state={state}" | ||||
|                 ), | ||||
| @ -562,6 +563,7 @@ class TestAuthorize(OAuthTestCase): | ||||
|                 "url": "http://localhost", | ||||
|                 "title": f"Redirecting to {app.name}...", | ||||
|                 "attrs": { | ||||
|                     "access_token": token.token, | ||||
|                     "id_token": provider.encode(token.id_token.to_dict()), | ||||
|                     "token_type": TOKEN_TYPE, | ||||
|                     "expires_in": "3600", | ||||
|  | ||||
| @ -150,12 +150,12 @@ class OAuthAuthorizationParams: | ||||
|         self.check_redirect_uri() | ||||
|         self.check_grant() | ||||
|         self.check_scope(github_compat) | ||||
|         self.check_nonce() | ||||
|         self.check_code_challenge() | ||||
|         if self.request: | ||||
|             raise AuthorizeError( | ||||
|                 self.redirect_uri, "request_not_supported", self.grant_type, self.state | ||||
|             ) | ||||
|         self.check_nonce() | ||||
|         self.check_code_challenge() | ||||
|  | ||||
|     def check_grant(self): | ||||
|         """Check grant""" | ||||
| @ -630,6 +630,7 @@ class OAuthFulfillmentStage(StageView): | ||||
|         if self.params.response_type in [ | ||||
|             ResponseTypes.ID_TOKEN_TOKEN, | ||||
|             ResponseTypes.CODE_ID_TOKEN_TOKEN, | ||||
|             ResponseTypes.ID_TOKEN, | ||||
|             ResponseTypes.CODE_TOKEN, | ||||
|         ]: | ||||
|             query_fragment["access_token"] = token.token | ||||
|  | ||||
| @ -10,11 +10,3 @@ class AuthentikProviderProxyConfig(ManagedAppConfig): | ||||
|     label = "authentik_providers_proxy" | ||||
|     verbose_name = "authentik Providers.Proxy" | ||||
|     default = True | ||||
|  | ||||
|     @ManagedAppConfig.reconcile_tenant | ||||
|     def proxy_set_defaults(self): | ||||
|         from authentik.providers.proxy.models import ProxyProvider | ||||
|  | ||||
|         for provider in ProxyProvider.objects.all(): | ||||
|             provider.set_oauth_defaults() | ||||
|             provider.save() | ||||
|  | ||||
| @ -47,8 +47,6 @@ class IngressReconciler(KubernetesObjectReconciler[V1Ingress]): | ||||
|     def reconcile(self, current: V1Ingress, reference: V1Ingress): | ||||
|         super().reconcile(current, reference) | ||||
|         self._check_annotations(current, reference) | ||||
|         if current.spec.ingress_class_name != reference.spec.ingress_class_name: | ||||
|             raise NeedsUpdate() | ||||
|         # Create a list of all expected host and tls hosts | ||||
|         expected_hosts = [] | ||||
|         expected_hosts_tls = [] | ||||
|  | ||||
							
								
								
									
										23
									
								
								authentik/providers/proxy/signals.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										23
									
								
								authentik/providers/proxy/signals.py
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,23 @@ | ||||
| """Proxy provider signals""" | ||||
|  | ||||
| from django.contrib.auth.signals import user_logged_out | ||||
| from django.db.models.signals import pre_delete | ||||
| from django.dispatch import receiver | ||||
| from django.http import HttpRequest | ||||
|  | ||||
| from authentik.core.models import AuthenticatedSession, User | ||||
| from authentik.providers.proxy.tasks import proxy_on_logout | ||||
|  | ||||
|  | ||||
| @receiver(user_logged_out) | ||||
| def logout_proxy_revoke_direct(sender: type[User], request: HttpRequest, **_): | ||||
|     """Catch logout by direct logout and forward to proxy providers""" | ||||
|     if not request.session or not request.session.session_key: | ||||
|         return | ||||
|     proxy_on_logout.delay(request.session.session_key) | ||||
|  | ||||
|  | ||||
| @receiver(pre_delete, sender=AuthenticatedSession) | ||||
| def logout_proxy_revoke(sender: type[AuthenticatedSession], instance: AuthenticatedSession, **_): | ||||
|     """Catch logout by expiring sessions being deleted""" | ||||
|     proxy_on_logout.delay(instance.session.session_key) | ||||
							
								
								
									
										38
									
								
								authentik/providers/proxy/tasks.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										38
									
								
								authentik/providers/proxy/tasks.py
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,38 @@ | ||||
| """proxy provider tasks""" | ||||
|  | ||||
| from asgiref.sync import async_to_sync | ||||
| from channels.layers import get_channel_layer | ||||
| from django.db import DatabaseError, InternalError, ProgrammingError | ||||
|  | ||||
| from authentik.outposts.consumer import OUTPOST_GROUP | ||||
| from authentik.outposts.models import Outpost, OutpostType | ||||
| from authentik.providers.oauth2.id_token import hash_session_key | ||||
| from authentik.providers.proxy.models import ProxyProvider | ||||
| from authentik.root.celery import CELERY_APP | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task( | ||||
|     throws=(DatabaseError, ProgrammingError, InternalError), | ||||
| ) | ||||
| def proxy_set_defaults(): | ||||
|     """Ensure correct defaults are set for all providers""" | ||||
|     for provider in ProxyProvider.objects.all(): | ||||
|         provider.set_oauth_defaults() | ||||
|         provider.save() | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task() | ||||
| def proxy_on_logout(session_id: str): | ||||
|     """Update outpost instances connected to a single outpost""" | ||||
|     layer = get_channel_layer() | ||||
|     hashed_session_id = hash_session_key(session_id) | ||||
|     for outpost in Outpost.objects.filter(type=OutpostType.PROXY): | ||||
|         group = OUTPOST_GROUP % {"outpost_pk": str(outpost.pk)} | ||||
|         async_to_sync(layer.group_send)( | ||||
|             group, | ||||
|             { | ||||
|                 "type": "event.provider.specific", | ||||
|                 "sub_type": "logout", | ||||
|                 "session_id": hashed_session_id, | ||||
|             }, | ||||
|         ) | ||||
| @ -166,6 +166,7 @@ class ConnectionToken(ExpiringModel): | ||||
|         always_merger.merge(settings, default_settings) | ||||
|         always_merger.merge(settings, self.endpoint.provider.settings) | ||||
|         always_merger.merge(settings, self.endpoint.settings) | ||||
|         always_merger.merge(settings, self.settings) | ||||
|  | ||||
|         def mapping_evaluator(mappings: QuerySet): | ||||
|             for mapping in mappings: | ||||
| @ -190,7 +191,6 @@ class ConnectionToken(ExpiringModel): | ||||
|         mapping_evaluator( | ||||
|             RACPropertyMapping.objects.filter(endpoint__in=[self.endpoint]).order_by("name") | ||||
|         ) | ||||
|         always_merger.merge(settings, self.settings) | ||||
|  | ||||
|         settings["drive-path"] = f"/tmp/connection/{self.token}"  # nosec | ||||
|         settings["create-drive-path"] = "true" | ||||
|  | ||||
| @ -49,7 +49,6 @@ class TestEndpointsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
| @ -102,7 +101,6 @@ class TestEndpointsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -90,6 +90,23 @@ class TestModels(TransactionTestCase): | ||||
|                 "resize-method": "display-update", | ||||
|             }, | ||||
|         ) | ||||
|         # Set settings in token | ||||
|         token.settings = { | ||||
|             "level": "token", | ||||
|         } | ||||
|         token.save() | ||||
|         self.assertEqual( | ||||
|             token.get_settings(), | ||||
|             { | ||||
|                 "hostname": self.endpoint.host.split(":")[0], | ||||
|                 "port": "1324", | ||||
|                 "client-name": f"authentik - {self.user}", | ||||
|                 "drive-path": path, | ||||
|                 "create-drive-path": "true", | ||||
|                 "level": "token", | ||||
|                 "resize-method": "display-update", | ||||
|             }, | ||||
|         ) | ||||
|         # Set settings in property mapping (provider) | ||||
|         mapping = RACPropertyMapping.objects.create( | ||||
|             name=generate_id(), | ||||
| @ -134,22 +151,3 @@ class TestModels(TransactionTestCase): | ||||
|                 "resize-method": "display-update", | ||||
|             }, | ||||
|         ) | ||||
|         # Set settings in token | ||||
|         token.settings = { | ||||
|             "level": "token", | ||||
|         } | ||||
|         token.save() | ||||
|         self.assertEqual( | ||||
|             token.get_settings(), | ||||
|             { | ||||
|                 "hostname": self.endpoint.host.split(":")[0], | ||||
|                 "port": "1324", | ||||
|                 "client-name": f"authentik - {self.user}", | ||||
|                 "drive-path": path, | ||||
|                 "create-drive-path": "true", | ||||
|                 "foo": "true", | ||||
|                 "bar": "6", | ||||
|                 "resize-method": "display-update", | ||||
|                 "level": "token", | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
| @ -20,9 +20,6 @@ from authentik.lib.utils.time import timedelta_from_string | ||||
| from authentik.policies.engine import PolicyEngine | ||||
| from authentik.policies.views import PolicyAccessView | ||||
| from authentik.providers.rac.models import ConnectionToken, Endpoint, RACProvider | ||||
| from authentik.stages.prompt.stage import PLAN_CONTEXT_PROMPT | ||||
|  | ||||
| PLAN_CONNECTION_SETTINGS = "connection_settings" | ||||
|  | ||||
|  | ||||
| class RACStartView(PolicyAccessView): | ||||
| @ -112,15 +109,10 @@ class RACFinalStage(RedirectStage): | ||||
|         return super().dispatch(request, *args, **kwargs) | ||||
|  | ||||
|     def get_challenge(self, *args, **kwargs) -> RedirectChallenge: | ||||
|         settings = self.executor.plan.context.get(PLAN_CONNECTION_SETTINGS) | ||||
|         if not settings: | ||||
|             settings = self.executor.plan.context.get(PLAN_CONTEXT_PROMPT, {}).get( | ||||
|                 PLAN_CONNECTION_SETTINGS | ||||
|             ) | ||||
|         token = ConnectionToken.objects.create( | ||||
|             provider=self.provider, | ||||
|             endpoint=self.endpoint, | ||||
|             settings=settings or {}, | ||||
|             settings=self.executor.plan.context.get("connection_settings", {}), | ||||
|             session=self.request.session["authenticatedsession"], | ||||
|             expires=now() + timedelta_from_string(self.provider.connection_expiry), | ||||
|             expiring=True, | ||||
|  | ||||
| @ -1,12 +1,11 @@ | ||||
| """authentik radius provider app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikProviderRadiusConfig(ManagedAppConfig): | ||||
| class AuthentikProviderRadiusConfig(AppConfig): | ||||
|     """authentik radius provider app config""" | ||||
|  | ||||
|     name = "authentik.providers.radius" | ||||
|     label = "authentik_providers_radius" | ||||
|     verbose_name = "authentik Providers.Radius" | ||||
|     default = True | ||||
|  | ||||
| @ -1,13 +1,12 @@ | ||||
| """authentik SAML IdP app config""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikProviderSAMLConfig(ManagedAppConfig): | ||||
| class AuthentikProviderSAMLConfig(AppConfig): | ||||
|     """authentik SAML IdP app config""" | ||||
|  | ||||
|     name = "authentik.providers.saml" | ||||
|     label = "authentik_providers_saml" | ||||
|     verbose_name = "authentik Providers.SAML" | ||||
|     mountpoint = "application/saml/" | ||||
|     default = True | ||||
|  | ||||
| @ -47,16 +47,15 @@ class SCIMGroupClient(SCIMClient[Group, SCIMProviderGroup, SCIMGroupSchema]): | ||||
|  | ||||
|     def to_schema(self, obj: Group, connection: SCIMProviderGroup) -> SCIMGroupSchema: | ||||
|         """Convert authentik user into SCIM""" | ||||
|         raw_scim_group = super().to_schema(obj, connection) | ||||
|         raw_scim_group = super().to_schema( | ||||
|             obj, | ||||
|             connection, | ||||
|             schemas=(SCIM_GROUP_SCHEMA,), | ||||
|         ) | ||||
|         try: | ||||
|             scim_group = SCIMGroupSchema.model_validate(delete_none_values(raw_scim_group)) | ||||
|         except ValidationError as exc: | ||||
|             raise StopSync(exc, obj) from exc | ||||
|         if SCIM_GROUP_SCHEMA not in scim_group.schemas: | ||||
|             scim_group.schemas.insert(0, SCIM_GROUP_SCHEMA) | ||||
|         # As this might be unset, we need to tell pydantic it's set so ensure the schemas | ||||
|         # are included, even if its just the defaults | ||||
|         scim_group.schemas = list(scim_group.schemas) | ||||
|         if not scim_group.externalId: | ||||
|             scim_group.externalId = str(obj.pk) | ||||
|  | ||||
|  | ||||
| @ -31,16 +31,15 @@ class SCIMUserClient(SCIMClient[User, SCIMProviderUser, SCIMUserSchema]): | ||||
|  | ||||
|     def to_schema(self, obj: User, connection: SCIMProviderUser) -> SCIMUserSchema: | ||||
|         """Convert authentik user into SCIM""" | ||||
|         raw_scim_user = super().to_schema(obj, connection) | ||||
|         raw_scim_user = super().to_schema( | ||||
|             obj, | ||||
|             connection, | ||||
|             schemas=(SCIM_USER_SCHEMA,), | ||||
|         ) | ||||
|         try: | ||||
|             scim_user = SCIMUserSchema.model_validate(delete_none_values(raw_scim_user)) | ||||
|         except ValidationError as exc: | ||||
|             raise StopSync(exc, obj) from exc | ||||
|         if SCIM_USER_SCHEMA not in scim_user.schemas: | ||||
|             scim_user.schemas.insert(0, SCIM_USER_SCHEMA) | ||||
|         # As this might be unset, we need to tell pydantic it's set so ensure the schemas | ||||
|         # are included, even if its just the defaults | ||||
|         scim_user.schemas = list(scim_user.schemas) | ||||
|         if not scim_user.externalId: | ||||
|             scim_user.externalId = str(obj.uid) | ||||
|         return scim_user | ||||
|  | ||||
| @ -91,57 +91,6 @@ class SCIMUserTests(TestCase): | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
|     @Mocker() | ||||
|     def test_user_create_custom_schema(self, mock: Mocker): | ||||
|         """Test user creation with custom schema""" | ||||
|         schema = SCIMMapping.objects.create( | ||||
|             name="custom_schema", | ||||
|             expression="""return {"schemas": ["foo"]}""", | ||||
|         ) | ||||
|         self.provider.property_mappings.add(schema) | ||||
|         scim_id = generate_id() | ||||
|         mock.get( | ||||
|             "https://localhost/ServiceProviderConfig", | ||||
|             json={}, | ||||
|         ) | ||||
|         mock.post( | ||||
|             "https://localhost/Users", | ||||
|             json={ | ||||
|                 "id": scim_id, | ||||
|             }, | ||||
|         ) | ||||
|         uid = generate_id() | ||||
|         user = User.objects.create( | ||||
|             username=uid, | ||||
|             name=f"{uid} {uid}", | ||||
|             email=f"{uid}@goauthentik.io", | ||||
|         ) | ||||
|         self.assertEqual(mock.call_count, 2) | ||||
|         self.assertEqual(mock.request_history[0].method, "GET") | ||||
|         self.assertEqual(mock.request_history[1].method, "POST") | ||||
|         self.assertJSONEqual( | ||||
|             mock.request_history[1].body, | ||||
|             { | ||||
|                 "schemas": ["urn:ietf:params:scim:schemas:core:2.0:User", "foo"], | ||||
|                 "active": True, | ||||
|                 "emails": [ | ||||
|                     { | ||||
|                         "primary": True, | ||||
|                         "type": "other", | ||||
|                         "value": f"{uid}@goauthentik.io", | ||||
|                     } | ||||
|                 ], | ||||
|                 "externalId": user.uid, | ||||
|                 "name": { | ||||
|                     "familyName": uid, | ||||
|                     "formatted": f"{uid} {uid}", | ||||
|                     "givenName": uid, | ||||
|                 }, | ||||
|                 "displayName": f"{uid} {uid}", | ||||
|                 "userName": uid, | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
|     @Mocker() | ||||
|     def test_user_create_different_provider_same_id(self, mock: Mocker): | ||||
|         """Test user creation with multiple providers that happen | ||||
| @ -435,7 +384,7 @@ class SCIMUserTests(TestCase): | ||||
|                 self.assertIn(request.method, SAFE_METHODS) | ||||
|         task = SystemTask.objects.filter(uid=slugify(self.provider.name)).first() | ||||
|         self.assertIsNotNone(task) | ||||
|         drop_msg = task.messages[3] | ||||
|         drop_msg = task.messages[2] | ||||
|         self.assertEqual(drop_msg["event"], "Dropping mutating request due to dry run") | ||||
|         self.assertIsNotNone(drop_msg["attributes"]["url"]) | ||||
|         self.assertIsNotNone(drop_msg["attributes"]["body"]) | ||||
|  | ||||
| @ -44,7 +44,6 @@ class TestRBACRoleAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             res.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -46,7 +46,6 @@ class TestRBACUserAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             res.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -38,7 +38,6 @@ class TestAPIPerms(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             res.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
| @ -74,7 +73,6 @@ class TestAPIPerms(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             res.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -1,29 +1,12 @@ | ||||
| """test decorators api""" | ||||
|  | ||||
| from django.urls import reverse | ||||
| from guardian.shortcuts import assign_perm | ||||
| from rest_framework.decorators import action | ||||
| from rest_framework.request import Request | ||||
| from rest_framework.response import Response | ||||
| from rest_framework.test import APITestCase | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
|  | ||||
| from authentik.core.models import Application | ||||
| from authentik.core.tests.utils import create_test_user | ||||
| from authentik.lib.generators import generate_id | ||||
| from authentik.lib.tests.utils import get_request | ||||
| from authentik.rbac.decorators import permission_required | ||||
|  | ||||
|  | ||||
| class MVS(ModelViewSet): | ||||
|  | ||||
|     queryset = Application.objects.all() | ||||
|     lookup_field = "slug" | ||||
|  | ||||
|     @permission_required("authentik_core.view_application", ["authentik_events.view_event"]) | ||||
|     @action(detail=True, pagination_class=None, filter_backends=[]) | ||||
|     def test(self, request: Request, slug: str): | ||||
|         self.get_object() | ||||
|         return Response(status=200) | ||||
|  | ||||
|  | ||||
| class TestAPIDecorators(APITestCase): | ||||
| @ -35,33 +18,41 @@ class TestAPIDecorators(APITestCase): | ||||
|  | ||||
|     def test_obj_perm_denied(self): | ||||
|         """Test object perm denied""" | ||||
|         request = get_request("", user=self.user) | ||||
|         self.client.force_login(self.user) | ||||
|         app = Application.objects.create(name=generate_id(), slug=generate_id()) | ||||
|         response = MVS.as_view({"get": "test"})(request, slug=app.slug) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:application-metrics", kwargs={"slug": app.slug}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 403) | ||||
|  | ||||
|     def test_obj_perm_global(self): | ||||
|         """Test object perm successful (global)""" | ||||
|         assign_perm("authentik_core.view_application", self.user) | ||||
|         assign_perm("authentik_events.view_event", self.user) | ||||
|         self.client.force_login(self.user) | ||||
|         app = Application.objects.create(name=generate_id(), slug=generate_id()) | ||||
|         request = get_request("", user=self.user) | ||||
|         response = MVS.as_view({"get": "test"})(request, slug=app.slug) | ||||
|         self.assertEqual(response.status_code, 200, response.data) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:application-metrics", kwargs={"slug": app.slug}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_obj_perm_scoped(self): | ||||
|         """Test object perm successful (scoped)""" | ||||
|         assign_perm("authentik_events.view_event", self.user) | ||||
|         app = Application.objects.create(name=generate_id(), slug=generate_id()) | ||||
|         assign_perm("authentik_core.view_application", self.user, app) | ||||
|         request = get_request("", user=self.user) | ||||
|         response = MVS.as_view({"get": "test"})(request, slug=app.slug) | ||||
|         self.client.force_login(self.user) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:application-metrics", kwargs={"slug": app.slug}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_other_perm_denied(self): | ||||
|         """Test other perm denied""" | ||||
|         self.client.force_login(self.user) | ||||
|         app = Application.objects.create(name=generate_id(), slug=generate_id()) | ||||
|         assign_perm("authentik_core.view_application", self.user, app) | ||||
|         request = get_request("", user=self.user) | ||||
|         response = MVS.as_view({"get": "test"})(request, slug=app.slug) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:application-metrics", kwargs={"slug": app.slug}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 403) | ||||
|  | ||||
Some files were not shown because too many files have changed in this diff Show More
		Reference in New Issue
	
	Block a user
	