Compare commits
	
		
			2 Commits
		
	
	
		
			website/do
			...
			policies-e
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| ad9b5e98ba | |||
| e4a21c824a | 
| @ -1,5 +1,5 @@ | ||||
| [bumpversion] | ||||
| current_version = 2025.6.3 | ||||
| current_version = 2025.2.4 | ||||
| tag = True | ||||
| commit = True | ||||
| parse = (?P<major>\d+)\.(?P<minor>\d+)\.(?P<patch>\d+)(?:-(?P<rc_t>[a-zA-Z-]+)(?P<rc_n>[1-9]\\d*))? | ||||
| @ -21,8 +21,6 @@ optional_value = final | ||||
|  | ||||
| [bumpversion:file:package.json] | ||||
|  | ||||
| [bumpversion:file:package-lock.json] | ||||
|  | ||||
| [bumpversion:file:docker-compose.yml] | ||||
|  | ||||
| [bumpversion:file:schema.yml] | ||||
| @ -33,4 +31,6 @@ optional_value = final | ||||
|  | ||||
| [bumpversion:file:internal/constants/constants.go] | ||||
|  | ||||
| [bumpversion:file:web/src/common/constants.ts] | ||||
|  | ||||
| [bumpversion:file:lifecycle/aws/template.yaml] | ||||
|  | ||||
| @ -5,10 +5,8 @@ dist/** | ||||
| build/** | ||||
| build_docs/** | ||||
| *Dockerfile | ||||
| **/*Dockerfile | ||||
| blueprints/local | ||||
| .git | ||||
| !gen-ts-api/node_modules | ||||
| !gen-ts-api/dist/** | ||||
| !gen-go-api/ | ||||
| .venv | ||||
|  | ||||
| @ -7,9 +7,6 @@ charset = utf-8 | ||||
| trim_trailing_whitespace = true | ||||
| insert_final_newline = true | ||||
|  | ||||
| [*.toml] | ||||
| indent_size = 2 | ||||
|  | ||||
| [*.html] | ||||
| indent_size = 2 | ||||
|  | ||||
|  | ||||
							
								
								
									
										2
									
								
								.github/actions/setup/action.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.github/actions/setup/action.yml
									
									
									
									
										vendored
									
									
								
							| @ -36,7 +36,7 @@ runs: | ||||
|       with: | ||||
|         go-version-file: "go.mod" | ||||
|     - name: Setup docker cache | ||||
|       uses: AndreKurait/docker-cache@0fe76702a40db986d9663c24954fc14c6a6031b7 | ||||
|       uses: ScribeMD/docker-cache@0.5.0 | ||||
|       with: | ||||
|         key: docker-images-${{ runner.os }}-${{ hashFiles('.github/actions/setup/docker-compose.yml', 'Makefile') }}-${{ inputs.postgresql_version }} | ||||
|     - name: Setup dependencies | ||||
|  | ||||
							
								
								
									
										33
									
								
								.github/dependabot.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										33
									
								
								.github/dependabot.yml
									
									
									
									
										vendored
									
									
								
							| @ -23,13 +23,7 @@ updates: | ||||
|   - package-ecosystem: npm | ||||
|     directories: | ||||
|       - "/web" | ||||
|       - "/web/packages/sfe" | ||||
|       - "/web/packages/core" | ||||
|       - "/web/packages/esbuild-plugin-live-reload" | ||||
|       - "/packages/prettier-config" | ||||
|       - "/packages/tsconfig" | ||||
|       - "/packages/docusaurus-config" | ||||
|       - "/packages/eslint-config" | ||||
|       - "/web/sfe" | ||||
|     schedule: | ||||
|       interval: daily | ||||
|       time: "04:00" | ||||
| @ -74,9 +68,6 @@ updates: | ||||
|       wdio: | ||||
|         patterns: | ||||
|           - "@wdio/*" | ||||
|       goauthentik: | ||||
|         patterns: | ||||
|           - "@goauthentik/*" | ||||
|   - package-ecosystem: npm | ||||
|     directory: "/website" | ||||
|     schedule: | ||||
| @ -97,16 +88,6 @@ updates: | ||||
|           - "swc-*" | ||||
|           - "lightningcss*" | ||||
|           - "@rspack/binding*" | ||||
|       goauthentik: | ||||
|         patterns: | ||||
|           - "@goauthentik/*" | ||||
|       eslint: | ||||
|         patterns: | ||||
|           - "@eslint/*" | ||||
|           - "@typescript-eslint/*" | ||||
|           - "eslint-*" | ||||
|           - "eslint" | ||||
|           - "typescript-eslint" | ||||
|   - package-ecosystem: npm | ||||
|     directory: "/lifecycle/aws" | ||||
|     schedule: | ||||
| @ -137,15 +118,3 @@ updates: | ||||
|       prefix: "core:" | ||||
|     labels: | ||||
|       - dependencies | ||||
|   - package-ecosystem: docker-compose | ||||
|     directories: | ||||
|       # - /scripts # Maybe | ||||
|       - /tests/e2e | ||||
|     schedule: | ||||
|       interval: daily | ||||
|       time: "04:00" | ||||
|     open-pull-requests-limit: 10 | ||||
|     commit-message: | ||||
|       prefix: "core:" | ||||
|     labels: | ||||
|       - dependencies | ||||
|  | ||||
| @ -38,8 +38,6 @@ jobs: | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|       # Needed for checkout | ||||
|       contents: read | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - uses: docker/setup-qemu-action@v3.6.0 | ||||
|  | ||||
							
								
								
									
										1
									
								
								.github/workflows/api-ts-publish.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										1
									
								
								.github/workflows/api-ts-publish.yml
									
									
									
									
										vendored
									
									
								
							| @ -53,7 +53,6 @@ jobs: | ||||
|           signoff: true | ||||
|           # ID from https://api.github.com/users/authentik-automation[bot] | ||||
|           author: authentik-automation[bot] <135050075+authentik-automation[bot]@users.noreply.github.com> | ||||
|           labels: dependencies | ||||
|       - uses: peter-evans/enable-pull-request-automerge@v3 | ||||
|         with: | ||||
|           token: ${{ steps.generate_token.outputs.token }} | ||||
|  | ||||
							
								
								
									
										3
									
								
								.github/workflows/ci-main-daily.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										3
									
								
								.github/workflows/ci-main-daily.yml
									
									
									
									
										vendored
									
									
								
							| @ -9,15 +9,14 @@ on: | ||||
|  | ||||
| jobs: | ||||
|   test-container: | ||||
|     if: ${{ github.repository != 'goauthentik/authentik-internal' }} | ||||
|     runs-on: ubuntu-latest | ||||
|     strategy: | ||||
|       fail-fast: false | ||||
|       matrix: | ||||
|         version: | ||||
|           - docs | ||||
|           - version-2025-4 | ||||
|           - version-2025-2 | ||||
|           - version-2024-12 | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - run: | | ||||
|  | ||||
							
								
								
									
										21
									
								
								.github/workflows/ci-main.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										21
									
								
								.github/workflows/ci-main.yml
									
									
									
									
										vendored
									
									
								
							| @ -62,7 +62,6 @@ jobs: | ||||
|         psql: | ||||
|           - 15-alpine | ||||
|           - 16-alpine | ||||
|           - 17-alpine | ||||
|         run_id: [1, 2, 3, 4, 5] | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
| @ -71,18 +70,22 @@ jobs: | ||||
|       - name: checkout stable | ||||
|         run: | | ||||
|           # Copy current, latest config to local | ||||
|           # Temporarly comment the .github backup while migrating to uv | ||||
|           cp authentik/lib/default.yml local.env.yml | ||||
|           cp -R .github .. | ||||
|           # cp -R .github .. | ||||
|           cp -R scripts .. | ||||
|           git checkout $(git tag --sort=version:refname | grep '^version/' | grep -vE -- '-rc[0-9]+$' | tail -n1) | ||||
|           rm -rf .github/ scripts/ | ||||
|           mv ../.github ../scripts . | ||||
|           # rm -rf .github/ scripts/ | ||||
|           # mv ../.github ../scripts . | ||||
|           rm -rf scripts/ | ||||
|           mv ../scripts . | ||||
|       - name: Setup authentik env (stable) | ||||
|         uses: ./.github/actions/setup | ||||
|         with: | ||||
|           postgresql_version: ${{ matrix.psql }} | ||||
|         continue-on-error: true | ||||
|       - name: run migrations to stable | ||||
|         run: uv run python -m lifecycle.migrate | ||||
|         run: poetry run python -m lifecycle.migrate | ||||
|       - name: checkout current code | ||||
|         run: | | ||||
|           set -x | ||||
| @ -117,7 +120,6 @@ jobs: | ||||
|         psql: | ||||
|           - 15-alpine | ||||
|           - 16-alpine | ||||
|           - 17-alpine | ||||
|         run_id: [1, 2, 3, 4, 5] | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
| @ -202,7 +204,7 @@ jobs: | ||||
|         uses: actions/cache@v4 | ||||
|         with: | ||||
|           path: web/dist | ||||
|           key: ${{ runner.os }}-web-${{ hashFiles('web/package-lock.json', 'package-lock.json', 'web/src/**', 'web/packages/sfe/src/**') }}-b | ||||
|           key: ${{ runner.os }}-web-${{ hashFiles('web/package-lock.json', 'web/src/**') }} | ||||
|       - name: prepare web ui | ||||
|         if: steps.cache-web.outputs.cache-hit != 'true' | ||||
|         working-directory: web | ||||
| @ -210,7 +212,6 @@ jobs: | ||||
|           npm ci | ||||
|           make -C .. gen-client-ts | ||||
|           npm run build | ||||
|           npm run build:sfe | ||||
|       - name: run e2e | ||||
|         run: | | ||||
|           uv run coverage run manage.py test ${{ matrix.job.glob }} | ||||
| @ -247,13 +248,11 @@ jobs: | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|       # Needed for checkout | ||||
|       contents: read | ||||
|     needs: ci-core-mark | ||||
|     uses: ./.github/workflows/_reusable-docker-build.yaml | ||||
|     secrets: inherit | ||||
|     with: | ||||
|       image_name: ${{ github.repository == 'goauthentik/authentik-internal' && 'ghcr.io/goauthentik/internal-server' || 'ghcr.io/goauthentik/dev-server' }} | ||||
|       image_name: ghcr.io/goauthentik/dev-server | ||||
|       release: false | ||||
|   pr-comment: | ||||
|     needs: | ||||
|  | ||||
							
								
								
									
										3
									
								
								.github/workflows/ci-outpost.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										3
									
								
								.github/workflows/ci-outpost.yml
									
									
									
									
										vendored
									
									
								
							| @ -29,7 +29,7 @@ jobs: | ||||
|       - name: Generate API | ||||
|         run: make gen-client-go | ||||
|       - name: golangci-lint | ||||
|         uses: golangci/golangci-lint-action@v8 | ||||
|         uses: golangci/golangci-lint-action@v7 | ||||
|         with: | ||||
|           version: latest | ||||
|           args: --timeout 5000s --verbose | ||||
| @ -59,7 +59,6 @@ jobs: | ||||
|         with: | ||||
|           jobs: ${{ toJSON(needs) }} | ||||
|   build-container: | ||||
|     if: ${{ github.repository != 'goauthentik/authentik-internal' }} | ||||
|     timeout-minutes: 120 | ||||
|     needs: | ||||
|       - ci-outpost-mark | ||||
|  | ||||
							
								
								
									
										52
									
								
								.github/workflows/ci-website.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										52
									
								
								.github/workflows/ci-website.yml
									
									
									
									
										vendored
									
									
								
							| @ -49,7 +49,6 @@ jobs: | ||||
|       matrix: | ||||
|         job: | ||||
|           - build | ||||
|           - build:integrations | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - uses: actions/setup-node@v4 | ||||
| @ -62,65 +61,14 @@ jobs: | ||||
|       - name: build | ||||
|         working-directory: website/ | ||||
|         run: npm run ${{ matrix.job }} | ||||
|   build-container: | ||||
|     if: ${{ github.repository != 'goauthentik/authentik-internal' }} | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
|       # Needed to upload container images to ghcr.io | ||||
|       packages: write | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|         with: | ||||
|           ref: ${{ github.event.pull_request.head.sha }} | ||||
|       - name: Set up QEMU | ||||
|         uses: docker/setup-qemu-action@v3.6.0 | ||||
|       - name: Set up Docker Buildx | ||||
|         uses: docker/setup-buildx-action@v3 | ||||
|       - name: prepare variables | ||||
|         uses: ./.github/actions/docker-push-variables | ||||
|         id: ev | ||||
|         env: | ||||
|           DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }} | ||||
|         with: | ||||
|           image-name: ghcr.io/goauthentik/dev-docs | ||||
|       - name: Login to Container Registry | ||||
|         if: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|         uses: docker/login-action@v3 | ||||
|         with: | ||||
|           registry: ghcr.io | ||||
|           username: ${{ github.repository_owner }} | ||||
|           password: ${{ secrets.GITHUB_TOKEN }} | ||||
|       - name: Build Docker Image | ||||
|         id: push | ||||
|         uses: docker/build-push-action@v6 | ||||
|         with: | ||||
|           tags: ${{ steps.ev.outputs.imageTags }} | ||||
|           file: website/Dockerfile | ||||
|           push: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|           platforms: linux/amd64,linux/arm64 | ||||
|           context: . | ||||
|           cache-from: type=registry,ref=ghcr.io/goauthentik/dev-docs:buildcache | ||||
|           cache-to: ${{ steps.ev.outputs.shouldPush == 'true' && 'type=registry,ref=ghcr.io/goauthentik/dev-docs:buildcache,mode=max' || '' }} | ||||
|       - uses: actions/attest-build-provenance@v2 | ||||
|         id: attest | ||||
|         if: ${{ steps.ev.outputs.shouldPush == 'true' }} | ||||
|         with: | ||||
|           subject-name: ${{ steps.ev.outputs.attestImageNames }} | ||||
|           subject-digest: ${{ steps.push.outputs.digest }} | ||||
|           push-to-registry: true | ||||
|   ci-website-mark: | ||||
|     if: always() | ||||
|     needs: | ||||
|       - lint | ||||
|       - test | ||||
|       - build | ||||
|       - build-container | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|       - uses: re-actors/alls-green@release/v1 | ||||
|         with: | ||||
|           jobs: ${{ toJSON(needs) }} | ||||
|           allowed-skips: ${{ github.repository == 'goauthentik/authentik-internal' && 'build-container' || '[]' }} | ||||
|  | ||||
							
								
								
									
										2
									
								
								.github/workflows/codeql-analysis.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.github/workflows/codeql-analysis.yml
									
									
									
									
										vendored
									
									
								
							| @ -2,7 +2,7 @@ name: "CodeQL" | ||||
|  | ||||
| on: | ||||
|   push: | ||||
|     branches: [main, next, version*] | ||||
|     branches: [main, "*", next, version*] | ||||
|   pull_request: | ||||
|     branches: [main] | ||||
|   schedule: | ||||
|  | ||||
| @ -37,7 +37,6 @@ jobs: | ||||
|           signoff: true | ||||
|           # ID from https://api.github.com/users/authentik-automation[bot] | ||||
|           author: authentik-automation[bot] <135050075+authentik-automation[bot]@users.noreply.github.com> | ||||
|           labels: dependencies | ||||
|       - uses: peter-evans/enable-pull-request-automerge@v3 | ||||
|         with: | ||||
|           token: ${{ steps.generate_token.outputs.token }} | ||||
|  | ||||
							
								
								
									
										1
									
								
								.github/workflows/image-compress.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										1
									
								
								.github/workflows/image-compress.yml
									
									
									
									
										vendored
									
									
								
							| @ -53,7 +53,6 @@ jobs: | ||||
|           body: ${{ steps.compress.outputs.markdown }} | ||||
|           delete-branch: true | ||||
|           signoff: true | ||||
|           labels: dependencies | ||||
|       - uses: peter-evans/enable-pull-request-automerge@v3 | ||||
|         if: "${{ github.event_name != 'pull_request' && steps.compress.outputs.markdown != '' }}" | ||||
|         with: | ||||
|  | ||||
							
								
								
									
										24
									
								
								.github/workflows/packages-npm-publish.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										24
									
								
								.github/workflows/packages-npm-publish.yml
									
									
									
									
										vendored
									
									
								
							| @ -3,11 +3,10 @@ on: | ||||
|   push: | ||||
|     branches: [main] | ||||
|     paths: | ||||
|       - packages/docusaurus-config/** | ||||
|       - packages/eslint-config/** | ||||
|       - packages/prettier-config/** | ||||
|       - packages/tsconfig/** | ||||
|       - web/packages/esbuild-plugin-live-reload/** | ||||
|       - packages/docusaurus-config | ||||
|       - packages/eslint-config | ||||
|       - packages/prettier-config | ||||
|       - packages/tsconfig | ||||
|   workflow_dispatch: | ||||
| jobs: | ||||
|   publish: | ||||
| @ -17,28 +16,27 @@ jobs: | ||||
|       fail-fast: false | ||||
|       matrix: | ||||
|         package: | ||||
|           - packages/docusaurus-config | ||||
|           - packages/eslint-config | ||||
|           - packages/prettier-config | ||||
|           - packages/tsconfig | ||||
|           - web/packages/esbuild-plugin-live-reload | ||||
|           - docusaurus-config | ||||
|           - eslint-config | ||||
|           - prettier-config | ||||
|           - tsconfig | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|         with: | ||||
|           fetch-depth: 2 | ||||
|       - uses: actions/setup-node@v4 | ||||
|         with: | ||||
|           node-version-file: ${{ matrix.package }}/package.json | ||||
|           node-version-file: packages/${{ matrix.package }}/package.json | ||||
|           registry-url: "https://registry.npmjs.org" | ||||
|       - name: Get changed files | ||||
|         id: changed-files | ||||
|         uses: tj-actions/changed-files@ed68ef82c095e0d48ec87eccea555d944a631a4c | ||||
|         with: | ||||
|           files: | | ||||
|             ${{ matrix.package }}/package.json | ||||
|             packages/${{ matrix.package }}/package.json | ||||
|       - name: Publish package | ||||
|         if: steps.changed-files.outputs.any_changed == 'true' | ||||
|         working-directory: ${{ matrix.package }} | ||||
|         working-directory: packages/${{ matrix.package}} | ||||
|         run: | | ||||
|           npm ci | ||||
|           npm run build | ||||
|  | ||||
							
								
								
									
										45
									
								
								.github/workflows/release-publish.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										45
									
								
								.github/workflows/release-publish.yml
									
									
									
									
										vendored
									
									
								
							| @ -20,49 +20,6 @@ jobs: | ||||
|       release: true | ||||
|       registry_dockerhub: true | ||||
|       registry_ghcr: true | ||||
|   build-docs: | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
|       # Needed to upload container images to ghcr.io | ||||
|       packages: write | ||||
|       # Needed for attestation | ||||
|       id-token: write | ||||
|       attestations: write | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - name: Set up QEMU | ||||
|         uses: docker/setup-qemu-action@v3.6.0 | ||||
|       - name: Set up Docker Buildx | ||||
|         uses: docker/setup-buildx-action@v3 | ||||
|       - name: prepare variables | ||||
|         uses: ./.github/actions/docker-push-variables | ||||
|         id: ev | ||||
|         env: | ||||
|           DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }} | ||||
|         with: | ||||
|           image-name: ghcr.io/goauthentik/docs | ||||
|       - name: Login to GitHub Container Registry | ||||
|         uses: docker/login-action@v3 | ||||
|         with: | ||||
|           registry: ghcr.io | ||||
|           username: ${{ github.repository_owner }} | ||||
|           password: ${{ secrets.GITHUB_TOKEN }} | ||||
|       - name: Build Docker Image | ||||
|         id: push | ||||
|         uses: docker/build-push-action@v6 | ||||
|         with: | ||||
|           tags: ${{ steps.ev.outputs.imageTags }} | ||||
|           file: website/Dockerfile | ||||
|           push: true | ||||
|           platforms: linux/amd64,linux/arm64 | ||||
|           context: . | ||||
|       - uses: actions/attest-build-provenance@v2 | ||||
|         id: attest | ||||
|         if: true | ||||
|         with: | ||||
|           subject-name: ${{ steps.ev.outputs.attestImageNames }} | ||||
|           subject-digest: ${{ steps.push.outputs.digest }} | ||||
|           push-to-registry: true | ||||
|   build-outpost: | ||||
|     runs-on: ubuntu-latest | ||||
|     permissions: | ||||
| @ -236,6 +193,6 @@ jobs: | ||||
|           SENTRY_ORG: authentik-security-inc | ||||
|           SENTRY_PROJECT: authentik | ||||
|         with: | ||||
|           release: authentik@${{ steps.ev.outputs.version }} | ||||
|           version: authentik@${{ steps.ev.outputs.version }} | ||||
|           sourcemaps: "./web/dist" | ||||
|           url_prefix: "~/static/dist" | ||||
|  | ||||
							
								
								
									
										21
									
								
								.github/workflows/repo-mirror-cleanup.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										21
									
								
								.github/workflows/repo-mirror-cleanup.yml
									
									
									
									
										vendored
									
									
								
							| @ -1,21 +0,0 @@ | ||||
| name: "authentik-repo-mirror-cleanup" | ||||
|  | ||||
| on: | ||||
|   workflow_dispatch: | ||||
|  | ||||
| jobs: | ||||
|   to_internal: | ||||
|     if: ${{ github.repository != 'goauthentik/authentik-internal' }} | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|         with: | ||||
|           fetch-depth: 0 | ||||
|       - if: ${{ env.MIRROR_KEY != '' }} | ||||
|         uses: BeryJu/repository-mirroring-action@5cf300935bc2e068f73ea69bcc411a8a997208eb | ||||
|         with: | ||||
|           target_repo_url: git@github.com:goauthentik/authentik-internal.git | ||||
|           ssh_private_key: ${{ secrets.GH_MIRROR_KEY }} | ||||
|           args: --tags --force --prune | ||||
|         env: | ||||
|           MIRROR_KEY: ${{ secrets.GH_MIRROR_KEY }} | ||||
							
								
								
									
										9
									
								
								.github/workflows/repo-mirror.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										9
									
								
								.github/workflows/repo-mirror.yml
									
									
									
									
										vendored
									
									
								
							| @ -11,10 +11,11 @@ jobs: | ||||
|         with: | ||||
|           fetch-depth: 0 | ||||
|       - if: ${{ env.MIRROR_KEY != '' }} | ||||
|         uses: BeryJu/repository-mirroring-action@5cf300935bc2e068f73ea69bcc411a8a997208eb | ||||
|         uses: pixta-dev/repository-mirroring-action@v1 | ||||
|         with: | ||||
|           target_repo_url: git@github.com:goauthentik/authentik-internal.git | ||||
|           ssh_private_key: ${{ secrets.GH_MIRROR_KEY }} | ||||
|           args: --tags --force | ||||
|           target_repo_url: | ||||
|             git@github.com:goauthentik/authentik-internal.git | ||||
|           ssh_private_key: | ||||
|             ${{ secrets.GH_MIRROR_KEY }} | ||||
|         env: | ||||
|           MIRROR_KEY: ${{ secrets.GH_MIRROR_KEY }} | ||||
|  | ||||
| @ -16,7 +16,6 @@ env: | ||||
|  | ||||
| jobs: | ||||
|   compile: | ||||
|     if: ${{ github.repository != 'goauthentik/authentik-internal' }} | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|       - id: generate_token | ||||
| @ -53,6 +52,3 @@ jobs: | ||||
|           body: "core, web: update translations" | ||||
|           delete-branch: true | ||||
|           signoff: true | ||||
|           labels: dependencies | ||||
|           # ID from https://api.github.com/users/authentik-automation[bot] | ||||
|           author: authentik-automation[bot] <135050075+authentik-automation[bot]@users.noreply.github.com> | ||||
|  | ||||
							
								
								
									
										15
									
								
								.github/workflows/translation-rename.yml
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										15
									
								
								.github/workflows/translation-rename.yml
									
									
									
									
										vendored
									
									
								
							| @ -15,7 +15,6 @@ jobs: | ||||
|     runs-on: ubuntu-latest | ||||
|     if: ${{ github.event.pull_request.user.login == 'transifex-integration[bot]'}} | ||||
|     steps: | ||||
|       - uses: actions/checkout@v4 | ||||
|       - id: generate_token | ||||
|         uses: tibdex/github-app-token@v2 | ||||
|         with: | ||||
| @ -26,13 +25,23 @@ jobs: | ||||
|         env: | ||||
|           GH_TOKEN: ${{ steps.generate_token.outputs.token }} | ||||
|         run: | | ||||
|           title=$(gh pr view ${{ github.event.pull_request.number }} --json  "title" -q ".title") | ||||
|           title=$(curl -q -L \ | ||||
|             -H "Accept: application/vnd.github+json" \ | ||||
|             -H "Authorization: Bearer ${GH_TOKEN}" \ | ||||
|             -H "X-GitHub-Api-Version: 2022-11-28" \ | ||||
|             https://api.github.com/repos/${GITHUB_REPOSITORY}/pulls/${{ github.event.pull_request.number }} | jq -r .title) | ||||
|           echo "title=${title}" >> "$GITHUB_OUTPUT" | ||||
|       - name: Rename | ||||
|         env: | ||||
|           GH_TOKEN: ${{ steps.generate_token.outputs.token }} | ||||
|         run: | | ||||
|           gh pr edit ${{ github.event.pull_request.number }} -t "translate: ${{ steps.title.outputs.title }}" --add-label dependencies | ||||
|           curl -L \ | ||||
|             -X PATCH \ | ||||
|             -H "Accept: application/vnd.github+json" \ | ||||
|             -H "Authorization: Bearer ${GH_TOKEN}" \ | ||||
|             -H "X-GitHub-Api-Version: 2022-11-28" \ | ||||
|             https://api.github.com/repos/${GITHUB_REPOSITORY}/pulls/${{ github.event.pull_request.number }} \ | ||||
|             -d "{\"title\":\"translate: ${{ steps.title.outputs.title }}\"}" | ||||
|       - uses: peter-evans/enable-pull-request-automerge@v3 | ||||
|         with: | ||||
|           token: ${{ steps.generate_token.outputs.token }} | ||||
|  | ||||
							
								
								
									
										10
									
								
								.vscode/settings.json
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										10
									
								
								.vscode/settings.json
									
									
									
									
										vendored
									
									
								
							| @ -6,19 +6,17 @@ | ||||
|         "!Context scalar", | ||||
|         "!Enumerate sequence", | ||||
|         "!Env scalar", | ||||
|         "!Env sequence", | ||||
|         "!Find sequence", | ||||
|         "!Format sequence", | ||||
|         "!If sequence", | ||||
|         "!Index scalar", | ||||
|         "!KeyOf scalar", | ||||
|         "!Value scalar", | ||||
|         "!AtIndex scalar", | ||||
|         "!ParseJSON scalar" | ||||
|         "!AtIndex scalar" | ||||
|     ], | ||||
|     "typescript.preferences.importModuleSpecifier": "non-relative", | ||||
|     "typescript.preferences.importModuleSpecifierEnding": "index", | ||||
|     "typescript.tsdk": "./node_modules/typescript/lib", | ||||
|     "typescript.tsdk": "./web/node_modules/typescript/lib", | ||||
|     "typescript.enablePromptUseWorkspaceTsdk": true, | ||||
|     "yaml.schemas": { | ||||
|         "./blueprints/schema.json": "blueprints/**/*.yaml" | ||||
| @ -32,5 +30,7 @@ | ||||
|         } | ||||
|     ], | ||||
|     "go.testFlags": ["-count=1"], | ||||
|     "github-actions.workflows.pinned.workflows": [".github/workflows/ci-main.yml"] | ||||
|     "github-actions.workflows.pinned.workflows": [ | ||||
|         ".github/workflows/ci-main.yml" | ||||
|     ] | ||||
| } | ||||
|  | ||||
							
								
								
									
										56
									
								
								Dockerfile
									
									
									
									
									
								
							
							
						
						
									
										56
									
								
								Dockerfile
									
									
									
									
									
								
							| @ -1,7 +1,26 @@ | ||||
| # syntax=docker/dockerfile:1 | ||||
|  | ||||
| # Stage 1: Build webui | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:24-slim AS node-builder | ||||
| # Stage 1: Build website | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:22 AS website-builder | ||||
|  | ||||
| ENV NODE_ENV=production | ||||
|  | ||||
| WORKDIR /work/website | ||||
|  | ||||
| RUN --mount=type=bind,target=/work/website/package.json,src=./website/package.json \ | ||||
|     --mount=type=bind,target=/work/website/package-lock.json,src=./website/package-lock.json \ | ||||
|     --mount=type=cache,id=npm-website,sharing=shared,target=/root/.npm \ | ||||
|     npm ci --include=dev | ||||
|  | ||||
| COPY ./website /work/website/ | ||||
| COPY ./blueprints /work/blueprints/ | ||||
| COPY ./schema.yml /work/ | ||||
| COPY ./SECURITY.md /work/ | ||||
|  | ||||
| RUN npm run build-bundled | ||||
|  | ||||
| # Stage 2: Build webui | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/node:22 AS web-builder | ||||
|  | ||||
| ARG GIT_BUILD_HASH | ||||
| ENV GIT_BUILD_HASH=$GIT_BUILD_HASH | ||||
| @ -13,7 +32,7 @@ RUN --mount=type=bind,target=/work/web/package.json,src=./web/package.json \ | ||||
|     --mount=type=bind,target=/work/web/package-lock.json,src=./web/package-lock.json \ | ||||
|     --mount=type=bind,target=/work/web/packages/sfe/package.json,src=./web/packages/sfe/package.json \ | ||||
|     --mount=type=bind,target=/work/web/scripts,src=./web/scripts \ | ||||
|     --mount=type=cache,id=npm-ak,sharing=shared,target=/root/.npm \ | ||||
|     --mount=type=cache,id=npm-web,sharing=shared,target=/root/.npm \ | ||||
|     npm ci --include=dev | ||||
|  | ||||
| COPY ./package.json /work | ||||
| @ -21,10 +40,9 @@ COPY ./web /work/web/ | ||||
| COPY ./website /work/website/ | ||||
| COPY ./gen-ts-api /work/web/node_modules/@goauthentik/api | ||||
|  | ||||
| RUN npm run build && \ | ||||
|     npm run build:sfe | ||||
| RUN npm run build | ||||
|  | ||||
| # Stage 2: Build go proxy | ||||
| # Stage 3: Build go proxy | ||||
| FROM --platform=${BUILDPLATFORM} docker.io/library/golang:1.24-bookworm AS go-builder | ||||
|  | ||||
| ARG TARGETOS | ||||
| @ -49,8 +67,8 @@ RUN --mount=type=bind,target=/go/src/goauthentik.io/go.mod,src=./go.mod \ | ||||
| COPY ./cmd /go/src/goauthentik.io/cmd | ||||
| COPY ./authentik/lib /go/src/goauthentik.io/authentik/lib | ||||
| COPY ./web/static.go /go/src/goauthentik.io/web/static.go | ||||
| COPY --from=node-builder /work/web/robots.txt /go/src/goauthentik.io/web/robots.txt | ||||
| COPY --from=node-builder /work/web/security.txt /go/src/goauthentik.io/web/security.txt | ||||
| COPY --from=web-builder /work/web/robots.txt /go/src/goauthentik.io/web/robots.txt | ||||
| COPY --from=web-builder /work/web/security.txt /go/src/goauthentik.io/web/security.txt | ||||
| COPY ./internal /go/src/goauthentik.io/internal | ||||
| COPY ./go.mod /go/src/goauthentik.io/go.mod | ||||
| COPY ./go.sum /go/src/goauthentik.io/go.sum | ||||
| @ -61,23 +79,24 @@ RUN --mount=type=cache,sharing=locked,target=/go/pkg/mod \ | ||||
|     CGO_ENABLED=1 GOFIPS140=latest GOARM="${TARGETVARIANT#v}" \ | ||||
|     go build -o /go/authentik ./cmd/server | ||||
|  | ||||
| # Stage 3: MaxMind GeoIP | ||||
| # Stage 4: MaxMind GeoIP | ||||
| FROM --platform=${BUILDPLATFORM} ghcr.io/maxmind/geoipupdate:v7.1.0 AS geoip | ||||
|  | ||||
| ENV GEOIPUPDATE_EDITION_IDS="GeoLite2-City GeoLite2-ASN" | ||||
| ENV GEOIPUPDATE_VERBOSE="1" | ||||
| ENV GEOIPUPDATE_ACCOUNT_ID_FILE="/run/secrets/GEOIPUPDATE_ACCOUNT_ID" | ||||
| ENV GEOIPUPDATE_LICENSE_KEY_FILE="/run/secrets/GEOIPUPDATE_LICENSE_KEY" | ||||
|  | ||||
| USER root | ||||
| RUN --mount=type=secret,id=GEOIPUPDATE_ACCOUNT_ID \ | ||||
|     --mount=type=secret,id=GEOIPUPDATE_LICENSE_KEY \ | ||||
|     mkdir -p /usr/share/GeoIP && \ | ||||
|     /bin/sh -c "GEOIPUPDATE_LICENSE_KEY_FILE=/run/secrets/GEOIPUPDATE_LICENSE_KEY /usr/bin/entry.sh || echo 'Failed to get GeoIP database, disabling'; exit 0" | ||||
|     /bin/sh -c "/usr/bin/entry.sh || echo 'Failed to get GeoIP database, disabling'; exit 0" | ||||
|  | ||||
| # Stage 4: Download uv | ||||
| FROM ghcr.io/astral-sh/uv:0.7.17 AS uv | ||||
| # Stage 5: Base python image | ||||
| FROM ghcr.io/goauthentik/fips-python:3.13.5-slim-bookworm-fips AS python-base | ||||
| # Stage 5: Download uv | ||||
| FROM ghcr.io/astral-sh/uv:0.6.14 AS uv | ||||
| # Stage 6: Base python image | ||||
| FROM ghcr.io/goauthentik/fips-python:3.12.10-slim-bookworm-fips AS python-base | ||||
|  | ||||
| ENV VENV_PATH="/ak-root/.venv" \ | ||||
|     PATH="/lifecycle:/ak-root/.venv/bin:$PATH" \ | ||||
| @ -90,7 +109,7 @@ WORKDIR /ak-root/ | ||||
|  | ||||
| COPY --from=uv /uv /uvx /bin/ | ||||
|  | ||||
| # Stage 6: Python dependencies | ||||
| # Stage 7: Python dependencies | ||||
| FROM python-base AS python-deps | ||||
|  | ||||
| ARG TARGETARCH | ||||
| @ -125,7 +144,7 @@ RUN --mount=type=bind,target=pyproject.toml,src=pyproject.toml \ | ||||
|     --mount=type=cache,target=/root/.cache/uv \ | ||||
|     uv sync --frozen --no-install-project --no-dev | ||||
|  | ||||
| # Stage 7: Run | ||||
| # Stage 8: Run | ||||
| FROM python-base AS final-image | ||||
|  | ||||
| ARG VERSION | ||||
| @ -168,8 +187,9 @@ COPY ./lifecycle/ /lifecycle | ||||
| COPY ./authentik/sources/kerberos/krb5.conf /etc/krb5.conf | ||||
| COPY --from=go-builder /go/authentik /bin/authentik | ||||
| COPY --from=python-deps /ak-root/.venv /ak-root/.venv | ||||
| COPY --from=node-builder /work/web/dist/ /web/dist/ | ||||
| COPY --from=node-builder /work/web/authentik/ /web/authentik/ | ||||
| COPY --from=web-builder /work/web/dist/ /web/dist/ | ||||
| COPY --from=web-builder /work/web/authentik/ /web/authentik/ | ||||
| COPY --from=website-builder /work/website/build/ /website/help/ | ||||
| COPY --from=geoip /usr/share/GeoIP /geoip | ||||
|  | ||||
| USER 1000 | ||||
|  | ||||
							
								
								
									
										59
									
								
								Makefile
									
									
									
									
									
								
							
							
						
						
									
										59
									
								
								Makefile
									
									
									
									
									
								
							| @ -1,7 +1,6 @@ | ||||
| .PHONY: gen dev-reset all clean test web website | ||||
|  | ||||
| SHELL := /usr/bin/env bash | ||||
| .SHELLFLAGS += ${SHELLFLAGS} -e -o pipefail | ||||
| .SHELLFLAGS += ${SHELLFLAGS} -e | ||||
| PWD = $(shell pwd) | ||||
| UID = $(shell id -u) | ||||
| GID = $(shell id -g) | ||||
| @ -9,9 +8,9 @@ NPM_VERSION = $(shell python -m scripts.generate_semver) | ||||
| PY_SOURCES = authentik tests scripts lifecycle .github | ||||
| DOCKER_IMAGE ?= "authentik:test" | ||||
|  | ||||
| GEN_API_TS = gen-ts-api | ||||
| GEN_API_PY = gen-py-api | ||||
| GEN_API_GO = gen-go-api | ||||
| GEN_API_TS = "gen-ts-api" | ||||
| GEN_API_PY = "gen-py-api" | ||||
| GEN_API_GO = "gen-go-api" | ||||
|  | ||||
| pg_user := $(shell uv run python -m authentik.lib.config postgresql.user 2>/dev/null) | ||||
| pg_host := $(shell uv run python -m authentik.lib.config postgresql.host 2>/dev/null) | ||||
| @ -86,10 +85,6 @@ dev-create-db: | ||||
|  | ||||
| dev-reset: dev-drop-db dev-create-db migrate  ## Drop and restore the Authentik PostgreSQL instance to a "fresh install" state. | ||||
|  | ||||
| update-test-mmdb:  ## Update test GeoIP and ASN Databases | ||||
| 	curl -L https://raw.githubusercontent.com/maxmind/MaxMind-DB/refs/heads/main/test-data/GeoLite2-ASN-Test.mmdb -o ${PWD}/tests/GeoLite2-ASN-Test.mmdb | ||||
| 	curl -L https://raw.githubusercontent.com/maxmind/MaxMind-DB/refs/heads/main/test-data/GeoLite2-City-Test.mmdb -o ${PWD}/tests/GeoLite2-City-Test.mmdb | ||||
|  | ||||
| ######################### | ||||
| ## API Schema | ||||
| ######################### | ||||
| @ -98,7 +93,7 @@ gen-build:  ## Extract the schema from the database | ||||
| 	AUTHENTIK_DEBUG=true \ | ||||
| 		AUTHENTIK_TENANTS__ENABLED=true \ | ||||
| 		AUTHENTIK_OUTPOSTS__DISABLE_EMBEDDED_OUTPOST=true \ | ||||
| 		uv run ak make_blueprint_schema --file blueprints/schema.json | ||||
| 		uv run ak make_blueprint_schema > blueprints/schema.json | ||||
| 	AUTHENTIK_DEBUG=true \ | ||||
| 		AUTHENTIK_TENANTS__ENABLED=true \ | ||||
| 		AUTHENTIK_OUTPOSTS__DISABLE_EMBEDDED_OUTPOST=true \ | ||||
| @ -122,19 +117,14 @@ gen-diff:  ## (Release) generate the changelog diff between the current schema a | ||||
| 	npx prettier --write diff.md | ||||
|  | ||||
| gen-clean-ts:  ## Remove generated API client for Typescript | ||||
| 	rm -rf ${PWD}/${GEN_API_TS}/ | ||||
| 	rm -rf ${PWD}/web/node_modules/@goauthentik/api/ | ||||
| 	rm -rf ./${GEN_API_TS}/ | ||||
| 	rm -rf ./web/node_modules/@goauthentik/api/ | ||||
|  | ||||
| gen-clean-go:  ## Remove generated API client for Go | ||||
| 	mkdir -p ${PWD}/${GEN_API_GO} | ||||
| ifneq ($(wildcard ${PWD}/${GEN_API_GO}/.*),) | ||||
| 	make -C ${PWD}/${GEN_API_GO} clean | ||||
| else | ||||
| 	rm -rf ${PWD}/${GEN_API_GO} | ||||
| endif | ||||
| 	rm -rf ./${GEN_API_GO}/ | ||||
|  | ||||
| gen-clean-py:  ## Remove generated API client for Python | ||||
| 	rm -rf ${PWD}/${GEN_API_PY}/ | ||||
| 	rm -rf ./${GEN_API_PY}/ | ||||
|  | ||||
| gen-clean: gen-clean-ts gen-clean-go gen-clean-py  ## Remove generated API clients | ||||
|  | ||||
| @ -150,9 +140,9 @@ gen-client-ts: gen-clean-ts  ## Build and install the authentik API for Typescri | ||||
| 		--additional-properties=npmVersion=${NPM_VERSION} \ | ||||
| 		--git-repo-id authentik \ | ||||
| 		--git-user-id goauthentik | ||||
|  | ||||
| 	cd ${PWD}/${GEN_API_TS} && npm link | ||||
| 	cd ${PWD}/web && npm link @goauthentik/api | ||||
| 	mkdir -p web/node_modules/@goauthentik/api | ||||
| 	cd ./${GEN_API_TS} && npm i | ||||
| 	\cp -rf ./${GEN_API_TS}/* web/node_modules/@goauthentik/api | ||||
|  | ||||
| gen-client-py: gen-clean-py ## Build and install the authentik API for Python | ||||
| 	docker run \ | ||||
| @ -166,17 +156,24 @@ gen-client-py: gen-clean-py ## Build and install the authentik API for Python | ||||
| 		--additional-properties=packageVersion=${NPM_VERSION} \ | ||||
| 		--git-repo-id authentik \ | ||||
| 		--git-user-id goauthentik | ||||
| 	pip install ./${GEN_API_PY} | ||||
|  | ||||
| gen-client-go: gen-clean-go  ## Build and install the authentik API for Golang | ||||
| 	mkdir -p ${PWD}/${GEN_API_GO} | ||||
| ifeq ($(wildcard ${PWD}/${GEN_API_GO}/.*),) | ||||
| 	git clone --depth 1 https://github.com/goauthentik/client-go.git ${PWD}/${GEN_API_GO} | ||||
| else | ||||
| 	cd ${PWD}/${GEN_API_GO} && git pull | ||||
| endif | ||||
| 	cp ${PWD}/schema.yml ${PWD}/${GEN_API_GO} | ||||
| 	make -C ${PWD}/${GEN_API_GO} build | ||||
| 	mkdir -p ./${GEN_API_GO} ./${GEN_API_GO}/templates | ||||
| 	wget https://raw.githubusercontent.com/goauthentik/client-go/main/config.yaml -O ./${GEN_API_GO}/config.yaml | ||||
| 	wget https://raw.githubusercontent.com/goauthentik/client-go/main/templates/README.mustache -O ./${GEN_API_GO}/templates/README.mustache | ||||
| 	wget https://raw.githubusercontent.com/goauthentik/client-go/main/templates/go.mod.mustache -O ./${GEN_API_GO}/templates/go.mod.mustache | ||||
| 	cp schema.yml ./${GEN_API_GO}/ | ||||
| 	docker run \ | ||||
| 		--rm -v ${PWD}/${GEN_API_GO}:/local \ | ||||
| 		--user ${UID}:${GID} \ | ||||
| 		docker.io/openapitools/openapi-generator-cli:v6.5.0 generate \ | ||||
| 		-i /local/schema.yml \ | ||||
| 		-g go \ | ||||
| 		-o /local/ \ | ||||
| 		-c /local/config.yaml | ||||
| 	go mod edit -replace goauthentik.io/api/v3=./${GEN_API_GO} | ||||
| 	rm -rf ./${GEN_API_GO}/config.yaml ./${GEN_API_GO}/templates/ | ||||
|  | ||||
| gen-dev-config:  ## Generate a local development config file | ||||
| 	uv run scripts/generate_config.py | ||||
| @ -247,7 +244,7 @@ docker:  ## Build a docker image of the current source tree | ||||
| 	DOCKER_BUILDKIT=1 docker build . --progress plain --tag ${DOCKER_IMAGE} | ||||
|  | ||||
| test-docker: | ||||
| 	BUILD=true ${PWD}/scripts/test_docker.sh | ||||
| 	BUILD=true ./scripts/test_docker.sh | ||||
|  | ||||
| ######################### | ||||
| ## CI | ||||
|  | ||||
| @ -42,4 +42,4 @@ See [SECURITY.md](SECURITY.md) | ||||
|  | ||||
| ## Adoption and Contributions | ||||
|  | ||||
| Your organization uses authentik? We'd love to add your logo to the readme and our website! Email us @ hello@goauthentik.io or open a GitHub Issue/PR! For more information on how to contribute to authentik, please refer to our [contribution guide](https://docs.goauthentik.io/docs/developer-docs?utm_source=github). | ||||
| Your organization uses authentik? We'd love to add your logo to the readme and our website! Email us @ hello@goauthentik.io or open a GitHub Issue/PR! For more information on how to contribute to authentik, please refer to our [CONTRIBUTING.md file](./CONTRIBUTING.md). | ||||
|  | ||||
| @ -20,8 +20,8 @@ Even if the issue is not a CVE, we still greatly appreciate your help in hardeni | ||||
|  | ||||
| | Version   | Supported | | ||||
| | --------- | --------- | | ||||
| | 2025.4.x  | ✅        | | ||||
| | 2025.6.x  | ✅        | | ||||
| | 2024.12.x | ✅        | | ||||
| | 2025.2.x  | ✅        | | ||||
|  | ||||
| ## Reporting a Vulnerability | ||||
|  | ||||
|  | ||||
| @ -2,7 +2,7 @@ | ||||
|  | ||||
| from os import environ | ||||
|  | ||||
| __version__ = "2025.6.3" | ||||
| __version__ = "2025.2.4" | ||||
| ENV_GIT_HASH_KEY = "GIT_BUILD_HASH" | ||||
|  | ||||
|  | ||||
|  | ||||
							
								
								
									
										79
									
								
								authentik/admin/api/metrics.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										79
									
								
								authentik/admin/api/metrics.py
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,79 @@ | ||||
| """authentik administration metrics""" | ||||
|  | ||||
| from datetime import timedelta | ||||
|  | ||||
| from django.db.models.functions import ExtractHour | ||||
| from drf_spectacular.utils import extend_schema, extend_schema_field | ||||
| from guardian.shortcuts import get_objects_for_user | ||||
| from rest_framework.fields import IntegerField, SerializerMethodField | ||||
| from rest_framework.permissions import IsAuthenticated | ||||
| from rest_framework.request import Request | ||||
| from rest_framework.response import Response | ||||
| from rest_framework.views import APIView | ||||
|  | ||||
| from authentik.core.api.utils import PassiveSerializer | ||||
| from authentik.events.models import EventAction | ||||
|  | ||||
|  | ||||
| class CoordinateSerializer(PassiveSerializer): | ||||
|     """Coordinates for diagrams""" | ||||
|  | ||||
|     x_cord = IntegerField(read_only=True) | ||||
|     y_cord = IntegerField(read_only=True) | ||||
|  | ||||
|  | ||||
| class LoginMetricsSerializer(PassiveSerializer): | ||||
|     """Login Metrics per 1h""" | ||||
|  | ||||
|     logins = SerializerMethodField() | ||||
|     logins_failed = SerializerMethodField() | ||||
|     authorizations = SerializerMethodField() | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins(self, _): | ||||
|         """Get successful logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins_failed(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN_FAILED | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_authorizations(self, _): | ||||
|         """Get successful authorizations per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         return ( | ||||
|             get_objects_for_user(user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|  | ||||
| class AdministrationMetricsViewSet(APIView): | ||||
|     """Login Metrics per 1h""" | ||||
|  | ||||
|     permission_classes = [IsAuthenticated] | ||||
|  | ||||
|     @extend_schema(responses={200: LoginMetricsSerializer(many=False)}) | ||||
|     def get(self, request: Request) -> Response: | ||||
|         """Login Metrics per 1h""" | ||||
|         serializer = LoginMetricsSerializer(True) | ||||
|         serializer.context["user"] = request.user | ||||
|         return Response(serializer.data) | ||||
| @ -1,7 +1,6 @@ | ||||
| """authentik administration overview""" | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django_tenants.utils import get_public_schema_name | ||||
| from drf_spectacular.utils import extend_schema | ||||
| from packaging.version import parse | ||||
| from rest_framework.fields import SerializerMethodField | ||||
| @ -14,7 +13,6 @@ from authentik import __version__, get_build_hash | ||||
| from authentik.admin.tasks import VERSION_CACHE_KEY, VERSION_NULL, update_latest_version | ||||
| from authentik.core.api.utils import PassiveSerializer | ||||
| from authentik.outposts.models import Outpost | ||||
| from authentik.tenants.utils import get_current_tenant | ||||
|  | ||||
|  | ||||
| class VersionSerializer(PassiveSerializer): | ||||
| @ -37,8 +35,6 @@ class VersionSerializer(PassiveSerializer): | ||||
|  | ||||
|     def get_version_latest(self, _) -> str: | ||||
|         """Get latest version from cache""" | ||||
|         if get_current_tenant().schema_name == get_public_schema_name(): | ||||
|             return __version__ | ||||
|         version_in_cache = cache.get(VERSION_CACHE_KEY) | ||||
|         if not version_in_cache:  # pragma: no cover | ||||
|             update_latest_version.delay() | ||||
|  | ||||
| @ -14,19 +14,3 @@ class AuthentikAdminConfig(ManagedAppConfig): | ||||
|     label = "authentik_admin" | ||||
|     verbose_name = "authentik Admin" | ||||
|     default = True | ||||
|  | ||||
|     @ManagedAppConfig.reconcile_global | ||||
|     def clear_update_notifications(self): | ||||
|         """Clear update notifications on startup if the notification was for the version | ||||
|         we're running now.""" | ||||
|         from packaging.version import parse | ||||
|  | ||||
|         from authentik.admin.tasks import LOCAL_VERSION | ||||
|         from authentik.events.models import EventAction, Notification | ||||
|  | ||||
|         for notification in Notification.objects.filter(event__action=EventAction.UPDATE_AVAILABLE): | ||||
|             if "new_version" not in notification.event.context: | ||||
|                 continue | ||||
|             notification_version = notification.event.context["new_version"] | ||||
|             if LOCAL_VERSION >= parse(notification_version): | ||||
|                 notification.delete() | ||||
|  | ||||
| @ -1,7 +1,6 @@ | ||||
| """authentik admin settings""" | ||||
|  | ||||
| from celery.schedules import crontab | ||||
| from django_tenants.utils import get_public_schema_name | ||||
|  | ||||
| from authentik.lib.utils.time import fqdn_rand | ||||
|  | ||||
| @ -9,7 +8,6 @@ CELERY_BEAT_SCHEDULE = { | ||||
|     "admin_latest_version": { | ||||
|         "task": "authentik.admin.tasks.update_latest_version", | ||||
|         "schedule": crontab(minute=fqdn_rand("admin_latest_version"), hour="*"), | ||||
|         "tenant_schemas": [get_public_schema_name()], | ||||
|         "options": {"queue": "authentik_scheduled"}, | ||||
|     } | ||||
| } | ||||
|  | ||||
| @ -1,6 +1,7 @@ | ||||
| """authentik admin tasks""" | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django.db import DatabaseError, InternalError, ProgrammingError | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
| from packaging.version import parse | ||||
| from requests import RequestException | ||||
| @ -8,7 +9,7 @@ from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik import __version__, get_build_hash | ||||
| from authentik.admin.apps import PROM_INFO | ||||
| from authentik.events.models import Event, EventAction | ||||
| from authentik.events.models import Event, EventAction, Notification | ||||
| from authentik.events.system_tasks import SystemTask, TaskStatus, prefill_task | ||||
| from authentik.lib.config import CONFIG | ||||
| from authentik.lib.utils.http import get_http_session | ||||
| @ -32,6 +33,20 @@ def _set_prom_info(): | ||||
|     ) | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task( | ||||
|     throws=(DatabaseError, ProgrammingError, InternalError), | ||||
| ) | ||||
| def clear_update_notifications(): | ||||
|     """Clear update notifications on startup if the notification was for the version | ||||
|     we're running now.""" | ||||
|     for notification in Notification.objects.filter(event__action=EventAction.UPDATE_AVAILABLE): | ||||
|         if "new_version" not in notification.event.context: | ||||
|             continue | ||||
|         notification_version = notification.event.context["new_version"] | ||||
|         if LOCAL_VERSION >= parse(notification_version): | ||||
|             notification.delete() | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task(bind=True, base=SystemTask) | ||||
| @prefill_task | ||||
| def update_latest_version(self: SystemTask): | ||||
|  | ||||
| @ -36,6 +36,11 @@ class TestAdminAPI(TestCase): | ||||
|         body = loads(response.content) | ||||
|         self.assertEqual(len(body), 0) | ||||
|  | ||||
|     def test_metrics(self): | ||||
|         """Test metrics API""" | ||||
|         response = self.client.get(reverse("authentik_api:admin_metrics")) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_apps(self): | ||||
|         """Test apps API""" | ||||
|         response = self.client.get(reverse("authentik_api:apps-list")) | ||||
|  | ||||
| @ -1,12 +1,12 @@ | ||||
| """test admin tasks""" | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.core.cache import cache | ||||
| from django.test import TestCase | ||||
| from requests_mock import Mocker | ||||
|  | ||||
| from authentik.admin.tasks import ( | ||||
|     VERSION_CACHE_KEY, | ||||
|     clear_update_notifications, | ||||
|     update_latest_version, | ||||
| ) | ||||
| from authentik.events.models import Event, EventAction | ||||
| @ -72,13 +72,12 @@ class TestAdminTasks(TestCase): | ||||
|  | ||||
|     def test_clear_update_notifications(self): | ||||
|         """Test clear of previous notification""" | ||||
|         admin_config = apps.get_app_config("authentik_admin") | ||||
|         Event.objects.create( | ||||
|             action=EventAction.UPDATE_AVAILABLE, context={"new_version": "99999999.9999999.9999999"} | ||||
|         ) | ||||
|         Event.objects.create(action=EventAction.UPDATE_AVAILABLE, context={"new_version": "1.1.1"}) | ||||
|         Event.objects.create(action=EventAction.UPDATE_AVAILABLE, context={}) | ||||
|         admin_config.clear_update_notifications() | ||||
|         clear_update_notifications() | ||||
|         self.assertFalse( | ||||
|             Event.objects.filter( | ||||
|                 action=EventAction.UPDATE_AVAILABLE, context__new_version="1.1" | ||||
|  | ||||
| @ -3,6 +3,7 @@ | ||||
| from django.urls import path | ||||
|  | ||||
| from authentik.admin.api.meta import AppsViewSet, ModelViewSet | ||||
| from authentik.admin.api.metrics import AdministrationMetricsViewSet | ||||
| from authentik.admin.api.system import SystemView | ||||
| from authentik.admin.api.version import VersionView | ||||
| from authentik.admin.api.version_history import VersionHistoryViewSet | ||||
| @ -11,6 +12,11 @@ from authentik.admin.api.workers import WorkerView | ||||
| api_urlpatterns = [ | ||||
|     ("admin/apps", AppsViewSet, "apps"), | ||||
|     ("admin/models", ModelViewSet, "models"), | ||||
|     path( | ||||
|         "admin/metrics/", | ||||
|         AdministrationMetricsViewSet.as_view(), | ||||
|         name="admin_metrics", | ||||
|     ), | ||||
|     path("admin/version/", VersionView.as_view(), name="admin_version"), | ||||
|     ("admin/version/history", VersionHistoryViewSet, "version_history"), | ||||
|     path("admin/workers/", WorkerView.as_view(), name="admin_workers"), | ||||
|  | ||||
| @ -1,13 +1,12 @@ | ||||
| """authentik API AppConfig""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikAPIConfig(ManagedAppConfig): | ||||
| class AuthentikAPIConfig(AppConfig): | ||||
|     """authentik API Config""" | ||||
|  | ||||
|     name = "authentik.api" | ||||
|     label = "authentik_api" | ||||
|     mountpoint = "api/" | ||||
|     verbose_name = "authentik API" | ||||
|     default = True | ||||
|  | ||||
| @ -1,12 +1,9 @@ | ||||
| """API Authentication""" | ||||
|  | ||||
| from hmac import compare_digest | ||||
| from pathlib import Path | ||||
| from tempfile import gettempdir | ||||
| from typing import Any | ||||
|  | ||||
| from django.conf import settings | ||||
| from django.contrib.auth.models import AnonymousUser | ||||
| from drf_spectacular.extensions import OpenApiAuthenticationExtension | ||||
| from rest_framework.authentication import BaseAuthentication, get_authorization_header | ||||
| from rest_framework.exceptions import AuthenticationFailed | ||||
| @ -14,17 +11,11 @@ from rest_framework.request import Request | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.core.middleware import CTX_AUTH_VIA | ||||
| from authentik.core.models import Token, TokenIntents, User, UserTypes | ||||
| from authentik.core.models import Token, TokenIntents, User | ||||
| from authentik.outposts.models import Outpost | ||||
| from authentik.providers.oauth2.constants import SCOPE_AUTHENTIK_API | ||||
|  | ||||
| LOGGER = get_logger() | ||||
| _tmp = Path(gettempdir()) | ||||
| try: | ||||
|     with open(_tmp / "authentik-core-ipc.key") as _f: | ||||
|         ipc_key = _f.read() | ||||
| except OSError: | ||||
|     ipc_key = None | ||||
|  | ||||
|  | ||||
| def validate_auth(header: bytes) -> str | None: | ||||
| @ -82,11 +73,6 @@ def auth_user_lookup(raw_header: bytes) -> User | None: | ||||
|     if user: | ||||
|         CTX_AUTH_VIA.set("secret_key") | ||||
|         return user | ||||
|     # then try to auth via secret key (for embedded outpost/etc) | ||||
|     user = token_ipc(auth_credentials) | ||||
|     if user: | ||||
|         CTX_AUTH_VIA.set("ipc") | ||||
|         return user | ||||
|     raise AuthenticationFailed("Token invalid/expired") | ||||
|  | ||||
|  | ||||
| @ -104,43 +90,6 @@ def token_secret_key(value: str) -> User | None: | ||||
|     return outpost.user | ||||
|  | ||||
|  | ||||
| class IPCUser(AnonymousUser): | ||||
|     """'Virtual' user for IPC communication between authentik core and the authentik router""" | ||||
|  | ||||
|     username = "authentik:system" | ||||
|     is_active = True | ||||
|     is_superuser = True | ||||
|  | ||||
|     @property | ||||
|     def type(self): | ||||
|         return UserTypes.INTERNAL_SERVICE_ACCOUNT | ||||
|  | ||||
|     def has_perm(self, perm, obj=None): | ||||
|         return True | ||||
|  | ||||
|     def has_perms(self, perm_list, obj=None): | ||||
|         return True | ||||
|  | ||||
|     def has_module_perms(self, module): | ||||
|         return True | ||||
|  | ||||
|     @property | ||||
|     def is_anonymous(self): | ||||
|         return False | ||||
|  | ||||
|     @property | ||||
|     def is_authenticated(self): | ||||
|         return True | ||||
|  | ||||
|  | ||||
| def token_ipc(value: str) -> User | None: | ||||
|     """Check if the token is the secret key | ||||
|     and return the service account for the managed outpost""" | ||||
|     if not ipc_key or not compare_digest(value, ipc_key): | ||||
|         return None | ||||
|     return IPCUser() | ||||
|  | ||||
|  | ||||
| class TokenAuthentication(BaseAuthentication): | ||||
|     """Token-based authentication using HTTP Bearer authentication""" | ||||
|  | ||||
|  | ||||
| @ -54,7 +54,7 @@ def create_component(generator: SchemaGenerator, name, schema, type_=ResolvedCom | ||||
|     return component | ||||
|  | ||||
|  | ||||
| def postprocess_schema_responses(result, generator: SchemaGenerator, **kwargs): | ||||
| def postprocess_schema_responses(result, generator: SchemaGenerator, **kwargs):  # noqa: W0613 | ||||
|     """Workaround to set a default response for endpoints. | ||||
|     Workaround suggested at | ||||
|     <https://github.com/tfranzel/drf-spectacular/issues/119#issuecomment-656970357> | ||||
|  | ||||
| @ -72,33 +72,20 @@ class Command(BaseCommand): | ||||
|                     "additionalProperties": True, | ||||
|                 }, | ||||
|                 "entries": { | ||||
|                     "anyOf": [ | ||||
|                         { | ||||
|                             "type": "array", | ||||
|                             "items": {"$ref": "#/$defs/blueprint_entry"}, | ||||
|                         }, | ||||
|                         { | ||||
|                             "type": "object", | ||||
|                             "additionalProperties": { | ||||
|                                 "type": "array", | ||||
|                                 "items": {"$ref": "#/$defs/blueprint_entry"}, | ||||
|                             }, | ||||
|                         }, | ||||
|                     ], | ||||
|                     "type": "array", | ||||
|                     "items": { | ||||
|                         "oneOf": [], | ||||
|                     }, | ||||
|                 }, | ||||
|             }, | ||||
|             "$defs": {"blueprint_entry": {"oneOf": []}}, | ||||
|             "$defs": {}, | ||||
|         } | ||||
|  | ||||
|     def add_arguments(self, parser): | ||||
|         parser.add_argument("--file", type=str) | ||||
|  | ||||
|     @no_translations | ||||
|     def handle(self, *args, file: str, **options): | ||||
|     def handle(self, *args, **options): | ||||
|         """Generate JSON Schema for blueprints""" | ||||
|         self.build() | ||||
|         with open(file, "w") as _schema: | ||||
|             _schema.write(dumps(self.schema, indent=4, default=Command.json_default)) | ||||
|         self.stdout.write(dumps(self.schema, indent=4, default=Command.json_default)) | ||||
|  | ||||
|     @staticmethod | ||||
|     def json_default(value: Any) -> Any: | ||||
| @ -125,7 +112,7 @@ class Command(BaseCommand): | ||||
|                 } | ||||
|             ) | ||||
|             model_path = f"{model._meta.app_label}.{model._meta.model_name}" | ||||
|             self.schema["$defs"]["blueprint_entry"]["oneOf"].append( | ||||
|             self.schema["properties"]["entries"]["items"]["oneOf"].append( | ||||
|                 self.template_entry(model_path, model, serializer) | ||||
|             ) | ||||
|  | ||||
| @ -147,7 +134,7 @@ class Command(BaseCommand): | ||||
|                 "id": {"type": "string"}, | ||||
|                 "state": { | ||||
|                     "type": "string", | ||||
|                     "enum": sorted([s.value for s in BlueprintEntryDesiredState]), | ||||
|                     "enum": [s.value for s in BlueprintEntryDesiredState], | ||||
|                     "default": "present", | ||||
|                 }, | ||||
|                 "conditions": {"type": "array", "items": {"type": "boolean"}}, | ||||
| @ -218,7 +205,7 @@ class Command(BaseCommand): | ||||
|                 "type": "object", | ||||
|                 "required": ["permission"], | ||||
|                 "properties": { | ||||
|                     "permission": {"type": "string", "enum": sorted(perms)}, | ||||
|                     "permission": {"type": "string", "enum": perms}, | ||||
|                     "user": {"type": "integer"}, | ||||
|                     "role": {"type": "string"}, | ||||
|                 }, | ||||
|  | ||||
| @ -1,11 +1,10 @@ | ||||
| version: 1 | ||||
| entries: | ||||
|   foo: | ||||
|       - identifiers: | ||||
|             name: "%(id)s" | ||||
|             slug: "%(id)s" | ||||
|         model: authentik_flows.flow | ||||
|         state: present | ||||
|         attrs: | ||||
|             designation: stage_configuration | ||||
|             title: foo | ||||
|     - identifiers: | ||||
|           name: "%(id)s" | ||||
|           slug: "%(id)s" | ||||
|       model: authentik_flows.flow | ||||
|       state: present | ||||
|       attrs: | ||||
|           designation: stage_configuration | ||||
|           title: foo | ||||
|  | ||||
| @ -37,7 +37,6 @@ entries: | ||||
|     - attrs: | ||||
|           attributes: | ||||
|               env_null: !Env [bar-baz, null] | ||||
|               json_parse: !ParseJSON '{"foo": "bar"}' | ||||
|               policy_pk1: | ||||
|                   !Format [ | ||||
|                       "%s-%s", | ||||
|  | ||||
| @ -1,14 +0,0 @@ | ||||
| from django.test import TestCase | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from authentik.enterprise.apps import EnterpriseConfig | ||||
| from authentik.lib.utils.reflection import get_apps | ||||
|  | ||||
|  | ||||
| class TestManagedAppConfig(TestCase): | ||||
|     def test_apps_use_managed_app_config(self): | ||||
|         for app in get_apps(): | ||||
|             if app.name.startswith("authentik.enterprise"): | ||||
|                 self.assertIn(EnterpriseConfig, app.__class__.__bases__) | ||||
|             else: | ||||
|                 self.assertIn(ManagedAppConfig, app.__class__.__bases__) | ||||
| @ -35,6 +35,6 @@ def blueprint_tester(file_name: Path) -> Callable: | ||||
|  | ||||
|  | ||||
| for blueprint_file in Path("blueprints/").glob("**/*.yaml"): | ||||
|     if "local" in str(blueprint_file) or "testing" in str(blueprint_file): | ||||
|     if "local" in str(blueprint_file): | ||||
|         continue | ||||
|     setattr(TestPackaged, f"test_blueprint_{blueprint_file}", blueprint_tester(blueprint_file)) | ||||
|  | ||||
| @ -5,6 +5,7 @@ from collections.abc import Callable | ||||
| from django.apps import apps | ||||
| from django.test import TestCase | ||||
|  | ||||
| from authentik.blueprints.v1.importer import is_model_allowed | ||||
| from authentik.lib.models import SerializerModel | ||||
| from authentik.providers.oauth2.models import RefreshToken | ||||
|  | ||||
| @ -21,13 +22,10 @@ def serializer_tester_factory(test_model: type[SerializerModel]) -> Callable: | ||||
|             return | ||||
|         model_class = test_model() | ||||
|         self.assertTrue(isinstance(model_class, SerializerModel)) | ||||
|         # Models that have subclasses don't have to have a serializer | ||||
|         if len(test_model.__subclasses__()) > 0: | ||||
|             return | ||||
|         self.assertIsNotNone(model_class.serializer) | ||||
|         if model_class.serializer.Meta().model == RefreshToken: | ||||
|             return | ||||
|         self.assertTrue(issubclass(test_model, model_class.serializer.Meta().model)) | ||||
|         self.assertEqual(model_class.serializer.Meta().model, test_model) | ||||
|  | ||||
|     return tester | ||||
|  | ||||
| @ -36,6 +34,6 @@ for app in apps.get_app_configs(): | ||||
|     if not app.label.startswith("authentik"): | ||||
|         continue | ||||
|     for model in app.get_models(): | ||||
|         if not issubclass(model, SerializerModel): | ||||
|         if not is_model_allowed(model): | ||||
|             continue | ||||
|         setattr(TestModels, f"test_{app.label}_{model.__name__}", serializer_tester_factory(model)) | ||||
|  | ||||
| @ -215,7 +215,6 @@ class TestBlueprintsV1(TransactionTestCase): | ||||
|                     }, | ||||
|                     "nested_context": "context-nested-value", | ||||
|                     "env_null": None, | ||||
|                     "json_parse": {"foo": "bar"}, | ||||
|                     "at_index_sequence": "foo", | ||||
|                     "at_index_sequence_default": "non existent", | ||||
|                     "at_index_mapping": 2, | ||||
|  | ||||
| @ -6,7 +6,6 @@ from copy import copy | ||||
| from dataclasses import asdict, dataclass, field, is_dataclass | ||||
| from enum import Enum | ||||
| from functools import reduce | ||||
| from json import JSONDecodeError, loads | ||||
| from operator import ixor | ||||
| from os import getenv | ||||
| from typing import Any, Literal, Union | ||||
| @ -165,7 +164,9 @@ class BlueprintEntry: | ||||
|         """Get the blueprint model, with yaml tags resolved if present""" | ||||
|         return str(self.tag_resolver(self.model, blueprint)) | ||||
|  | ||||
|     def get_permissions(self, blueprint: "Blueprint") -> Generator[BlueprintEntryPermission]: | ||||
|     def get_permissions( | ||||
|         self, blueprint: "Blueprint" | ||||
|     ) -> Generator[BlueprintEntryPermission, None, None]: | ||||
|         """Get permissions of this entry, with all yaml tags resolved""" | ||||
|         for perm in self.permissions: | ||||
|             yield BlueprintEntryPermission( | ||||
| @ -192,18 +193,11 @@ class Blueprint: | ||||
|     """Dataclass used for a full export""" | ||||
|  | ||||
|     version: int = field(default=1) | ||||
|     entries: list[BlueprintEntry] | dict[str, list[BlueprintEntry]] = field(default_factory=list) | ||||
|     entries: list[BlueprintEntry] = field(default_factory=list) | ||||
|     context: dict = field(default_factory=dict) | ||||
|  | ||||
|     metadata: BlueprintMetadata | None = field(default=None) | ||||
|  | ||||
|     def iter_entries(self) -> Iterable[BlueprintEntry]: | ||||
|         if isinstance(self.entries, dict): | ||||
|             for _section, entries in self.entries.items(): | ||||
|                 yield from entries | ||||
|         else: | ||||
|             yield from self.entries | ||||
|  | ||||
|  | ||||
| class YAMLTag: | ||||
|     """Base class for all YAML Tags""" | ||||
| @ -234,7 +228,7 @@ class KeyOf(YAMLTag): | ||||
|         self.id_from = node.value | ||||
|  | ||||
|     def resolve(self, entry: BlueprintEntry, blueprint: Blueprint) -> Any: | ||||
|         for _entry in blueprint.iter_entries(): | ||||
|         for _entry in blueprint.entries: | ||||
|             if _entry.id == self.id_from and _entry._state.instance: | ||||
|                 # Special handling for PolicyBindingModels, as they'll have a different PK | ||||
|                 # which is used when creating policy bindings | ||||
| @ -292,22 +286,6 @@ class Context(YAMLTag): | ||||
|         return value | ||||
|  | ||||
|  | ||||
| class ParseJSON(YAMLTag): | ||||
|     """Parse JSON from context/env/etc value""" | ||||
|  | ||||
|     raw: str | ||||
|  | ||||
|     def __init__(self, loader: "BlueprintLoader", node: ScalarNode) -> None: | ||||
|         super().__init__() | ||||
|         self.raw = node.value | ||||
|  | ||||
|     def resolve(self, entry: BlueprintEntry, blueprint: Blueprint) -> Any: | ||||
|         try: | ||||
|             return loads(self.raw) | ||||
|         except JSONDecodeError as exc: | ||||
|             raise EntryInvalidError.from_entry(exc, entry) from exc | ||||
|  | ||||
|  | ||||
| class Format(YAMLTag): | ||||
|     """Format a string""" | ||||
|  | ||||
| @ -683,7 +661,6 @@ class BlueprintLoader(SafeLoader): | ||||
|         self.add_constructor("!Value", Value) | ||||
|         self.add_constructor("!Index", Index) | ||||
|         self.add_constructor("!AtIndex", AtIndex) | ||||
|         self.add_constructor("!ParseJSON", ParseJSON) | ||||
|  | ||||
|  | ||||
| class EntryInvalidError(SentryIgnoredException): | ||||
|  | ||||
| @ -384,7 +384,7 @@ class Importer: | ||||
|     def _apply_models(self, raise_errors=False) -> bool: | ||||
|         """Apply (create/update) models yaml""" | ||||
|         self.__pk_map = {} | ||||
|         for entry in self._import.iter_entries(): | ||||
|         for entry in self._import.entries: | ||||
|             model_app_label, model_name = entry.get_model(self._import).split(".") | ||||
|             try: | ||||
|                 model: type[SerializerModel] = registry.get_model(model_app_label, model_name) | ||||
|  | ||||
| @ -47,7 +47,7 @@ class MetaModelRegistry: | ||||
|         models = apps.get_models() | ||||
|         for _, value in self.models.items(): | ||||
|             models.append(value) | ||||
|         return sorted(models, key=str) | ||||
|         return models | ||||
|  | ||||
|     def get_model(self, app_label: str, model_id: str) -> type[Model]: | ||||
|         """Get model checks if any virtual models are registered, and falls back | ||||
|  | ||||
| @ -59,7 +59,6 @@ class BrandSerializer(ModelSerializer): | ||||
|             "flow_device_code", | ||||
|             "default_application", | ||||
|             "web_certificate", | ||||
|             "client_certificates", | ||||
|             "attributes", | ||||
|         ] | ||||
|         extra_kwargs = { | ||||
| @ -121,7 +120,6 @@ class BrandViewSet(UsedByMixin, ModelViewSet): | ||||
|         "domain", | ||||
|         "branding_title", | ||||
|         "web_certificate__name", | ||||
|         "client_certificates__name", | ||||
|     ] | ||||
|     filterset_fields = [ | ||||
|         "brand_uuid", | ||||
| @ -138,7 +136,6 @@ class BrandViewSet(UsedByMixin, ModelViewSet): | ||||
|         "flow_user_settings", | ||||
|         "flow_device_code", | ||||
|         "web_certificate", | ||||
|         "client_certificates", | ||||
|     ] | ||||
|     ordering = ["domain"] | ||||
|  | ||||
|  | ||||
| @ -1,9 +1,9 @@ | ||||
| """authentik brands app""" | ||||
|  | ||||
| from authentik.blueprints.apps import ManagedAppConfig | ||||
| from django.apps import AppConfig | ||||
|  | ||||
|  | ||||
| class AuthentikBrandsConfig(ManagedAppConfig): | ||||
| class AuthentikBrandsConfig(AppConfig): | ||||
|     """authentik Brand app""" | ||||
|  | ||||
|     name = "authentik.brands" | ||||
| @ -12,4 +12,3 @@ class AuthentikBrandsConfig(ManagedAppConfig): | ||||
|     mountpoints = { | ||||
|         "authentik.brands.urls_root": "", | ||||
|     } | ||||
|     default = True | ||||
|  | ||||
| @ -16,7 +16,7 @@ def migrate_custom_css(apps: Apps, schema_editor: BaseDatabaseSchemaEditor): | ||||
|     if not path.exists(): | ||||
|         return | ||||
|     css = path.read_text() | ||||
|     Brand.objects.using(db_alias).all().update(branding_custom_css=css) | ||||
|     Brand.objects.using(db_alias).update(branding_custom_css=css) | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
| @ -1,37 +0,0 @@ | ||||
| # Generated by Django 5.1.9 on 2025-05-19 15:09 | ||||
|  | ||||
| import django.db.models.deletion | ||||
| from django.db import migrations, models | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
|     dependencies = [ | ||||
|         ("authentik_brands", "0009_brand_branding_default_flow_background"), | ||||
|         ("authentik_crypto", "0004_alter_certificatekeypair_name"), | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.AddField( | ||||
|             model_name="brand", | ||||
|             name="client_certificates", | ||||
|             field=models.ManyToManyField( | ||||
|                 blank=True, | ||||
|                 default=None, | ||||
|                 help_text="Certificates used for client authentication.", | ||||
|                 to="authentik_crypto.certificatekeypair", | ||||
|             ), | ||||
|         ), | ||||
|         migrations.AlterField( | ||||
|             model_name="brand", | ||||
|             name="web_certificate", | ||||
|             field=models.ForeignKey( | ||||
|                 default=None, | ||||
|                 help_text="Web Certificate used by the authentik Core webserver.", | ||||
|                 null=True, | ||||
|                 on_delete=django.db.models.deletion.SET_DEFAULT, | ||||
|                 related_name="+", | ||||
|                 to="authentik_crypto.certificatekeypair", | ||||
|             ), | ||||
|         ), | ||||
|     ] | ||||
| @ -73,13 +73,6 @@ class Brand(SerializerModel): | ||||
|         default=None, | ||||
|         on_delete=models.SET_DEFAULT, | ||||
|         help_text=_("Web Certificate used by the authentik Core webserver."), | ||||
|         related_name="+", | ||||
|     ) | ||||
|     client_certificates = models.ManyToManyField( | ||||
|         CertificateKeyPair, | ||||
|         default=None, | ||||
|         blank=True, | ||||
|         help_text=_("Certificates used for client authentication."), | ||||
|     ) | ||||
|     attributes = models.JSONField(default=dict, blank=True) | ||||
|  | ||||
|  | ||||
| @ -148,14 +148,3 @@ class TestBrands(APITestCase): | ||||
|                 "default_locale": "", | ||||
|             }, | ||||
|         ) | ||||
|  | ||||
|     def test_custom_css(self): | ||||
|         """Test custom_css""" | ||||
|         brand = create_test_brand() | ||||
|         brand.branding_custom_css = """* { | ||||
|             font-family: "Foo bar"; | ||||
|         }""" | ||||
|         brand.save() | ||||
|         res = self.client.get(reverse("authentik_core:if-user")) | ||||
|         self.assertEqual(res.status_code, 200) | ||||
|         self.assertIn(brand.branding_custom_css, res.content.decode()) | ||||
|  | ||||
| @ -5,12 +5,10 @@ from typing import Any | ||||
| from django.db.models import F, Q | ||||
| from django.db.models import Value as V | ||||
| from django.http.request import HttpRequest | ||||
| from django.utils.html import _json_script_escapes | ||||
| from django.utils.safestring import mark_safe | ||||
| from sentry_sdk import get_current_span | ||||
|  | ||||
| from authentik import get_full_version | ||||
| from authentik.brands.models import Brand | ||||
| from authentik.lib.sentry import get_http_meta | ||||
| from authentik.tenants.models import Tenant | ||||
|  | ||||
| _q_default = Q(default=True) | ||||
| @ -34,14 +32,13 @@ def context_processor(request: HttpRequest) -> dict[str, Any]: | ||||
|     """Context Processor that injects brand object into every template""" | ||||
|     brand = getattr(request, "brand", DEFAULT_BRAND) | ||||
|     tenant = getattr(request, "tenant", Tenant()) | ||||
|     # similarly to `json_script` we escape everything HTML-related, however django | ||||
|     # only directly exposes this as a function that also wraps it in a <script> tag | ||||
|     # which we dont want for CSS | ||||
|     brand_css = mark_safe(str(brand.branding_custom_css).translate(_json_script_escapes))  # nosec | ||||
|     trace = "" | ||||
|     span = get_current_span() | ||||
|     if span: | ||||
|         trace = span.to_traceparent() | ||||
|     return { | ||||
|         "brand": brand, | ||||
|         "brand_css": brand_css, | ||||
|         "footer_links": tenant.footer_links, | ||||
|         "html_meta": {**get_http_meta()}, | ||||
|         "sentry_trace": trace, | ||||
|         "version": get_full_version(), | ||||
|     } | ||||
|  | ||||
| @ -2,9 +2,11 @@ | ||||
|  | ||||
| from collections.abc import Iterator | ||||
| from copy import copy | ||||
| from datetime import timedelta | ||||
|  | ||||
| from django.core.cache import cache | ||||
| from django.db.models import QuerySet | ||||
| from django.db.models.functions import ExtractHour | ||||
| from django.shortcuts import get_object_or_404 | ||||
| from drf_spectacular.types import OpenApiTypes | ||||
| from drf_spectacular.utils import OpenApiParameter, OpenApiResponse, extend_schema | ||||
| @ -18,6 +20,7 @@ from rest_framework.response import Response | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.admin.api.metrics import CoordinateSerializer | ||||
| from authentik.api.pagination import Pagination | ||||
| from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT | ||||
| from authentik.core.api.providers import ProviderSerializer | ||||
| @ -25,6 +28,7 @@ from authentik.core.api.used_by import UsedByMixin | ||||
| from authentik.core.api.utils import ModelSerializer | ||||
| from authentik.core.models import Application, User | ||||
| from authentik.events.logs import LogEventSerializer, capture_logs | ||||
| from authentik.events.models import EventAction | ||||
| from authentik.lib.utils.file import ( | ||||
|     FilePathSerializer, | ||||
|     FileUploadSerializer, | ||||
| @ -317,3 +321,18 @@ class ApplicationViewSet(UsedByMixin, ModelViewSet): | ||||
|         """Set application icon (as URL)""" | ||||
|         app: Application = self.get_object() | ||||
|         return set_file_url(request, app, "meta_icon") | ||||
|  | ||||
|     @permission_required("authentik_core.view_application", ["authentik_events.view_event"]) | ||||
|     @extend_schema(responses={200: CoordinateSerializer(many=True)}) | ||||
|     @action(detail=True, pagination_class=None, filter_backends=[]) | ||||
|     def metrics(self, request: Request, slug: str): | ||||
|         """Metrics for application logins""" | ||||
|         app = self.get_object() | ||||
|         return Response( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION, | ||||
|                 context__authorized_application__pk=app.pk.hex, | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
| @ -1,6 +1,8 @@ | ||||
| """Authenticator Devices API Views""" | ||||
|  | ||||
| from drf_spectacular.utils import extend_schema | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
| from drf_spectacular.types import OpenApiTypes | ||||
| from drf_spectacular.utils import OpenApiParameter, extend_schema | ||||
| from guardian.shortcuts import get_objects_for_user | ||||
| from rest_framework.fields import ( | ||||
|     BooleanField, | ||||
| @ -13,7 +15,6 @@ from rest_framework.request import Request | ||||
| from rest_framework.response import Response | ||||
| from rest_framework.viewsets import ViewSet | ||||
|  | ||||
| from authentik.core.api.users import ParamUserSerializer | ||||
| from authentik.core.api.utils import MetaNameSerializer | ||||
| from authentik.enterprise.stages.authenticator_endpoint_gdtc.models import EndpointDevice | ||||
| from authentik.stages.authenticator import device_classes, devices_for_user | ||||
| @ -22,7 +23,7 @@ from authentik.stages.authenticator_webauthn.models import WebAuthnDevice | ||||
|  | ||||
|  | ||||
| class DeviceSerializer(MetaNameSerializer): | ||||
|     """Serializer for authenticator devices""" | ||||
|     """Serializer for Duo authenticator devices""" | ||||
|  | ||||
|     pk = CharField() | ||||
|     name = CharField() | ||||
| @ -32,27 +33,22 @@ class DeviceSerializer(MetaNameSerializer): | ||||
|     last_updated = DateTimeField(read_only=True) | ||||
|     last_used = DateTimeField(read_only=True, allow_null=True) | ||||
|     extra_description = SerializerMethodField() | ||||
|     external_id = SerializerMethodField() | ||||
|  | ||||
|     def get_type(self, instance: Device) -> str: | ||||
|         """Get type of device""" | ||||
|         return instance._meta.label | ||||
|  | ||||
|     def get_extra_description(self, instance: Device) -> str | None: | ||||
|     def get_extra_description(self, instance: Device) -> str: | ||||
|         """Get extra description""" | ||||
|         if isinstance(instance, WebAuthnDevice): | ||||
|             return instance.device_type.description if instance.device_type else None | ||||
|             return ( | ||||
|                 instance.device_type.description | ||||
|                 if instance.device_type | ||||
|                 else _("Extra description not available") | ||||
|             ) | ||||
|         if isinstance(instance, EndpointDevice): | ||||
|             return instance.data.get("deviceSignals", {}).get("deviceModel") | ||||
|         return None | ||||
|  | ||||
|     def get_external_id(self, instance: Device) -> str | None: | ||||
|         """Get external Device ID""" | ||||
|         if isinstance(instance, WebAuthnDevice): | ||||
|             return instance.device_type.aaguid if instance.device_type else None | ||||
|         if isinstance(instance, EndpointDevice): | ||||
|             return instance.data.get("deviceSignals", {}).get("deviceModel") | ||||
|         return None | ||||
|         return "" | ||||
|  | ||||
|  | ||||
| class DeviceViewSet(ViewSet): | ||||
| @ -61,6 +57,7 @@ class DeviceViewSet(ViewSet): | ||||
|     serializer_class = DeviceSerializer | ||||
|     permission_classes = [IsAuthenticated] | ||||
|  | ||||
|     @extend_schema(responses={200: DeviceSerializer(many=True)}) | ||||
|     def list(self, request: Request) -> Response: | ||||
|         """Get all devices for current user""" | ||||
|         devices = devices_for_user(request.user) | ||||
| @ -82,11 +79,18 @@ class AdminDeviceViewSet(ViewSet): | ||||
|             yield from device_set | ||||
|  | ||||
|     @extend_schema( | ||||
|         parameters=[ParamUserSerializer], | ||||
|         parameters=[ | ||||
|             OpenApiParameter( | ||||
|                 name="user", | ||||
|                 location=OpenApiParameter.QUERY, | ||||
|                 type=OpenApiTypes.INT, | ||||
|             ) | ||||
|         ], | ||||
|         responses={200: DeviceSerializer(many=True)}, | ||||
|     ) | ||||
|     def list(self, request: Request) -> Response: | ||||
|         """Get all devices for current user""" | ||||
|         args = ParamUserSerializer(data=request.query_params) | ||||
|         args.is_valid(raise_exception=True) | ||||
|         return Response(DeviceSerializer(self.get_devices(**args.validated_data), many=True).data) | ||||
|         kwargs = {} | ||||
|         if "user" in request.query_params: | ||||
|             kwargs = {"user": request.query_params["user"]} | ||||
|         return Response(DeviceSerializer(self.get_devices(**kwargs), many=True).data) | ||||
|  | ||||
| @ -99,17 +99,18 @@ class GroupSerializer(ModelSerializer): | ||||
|             if superuser | ||||
|             else "authentik_core.disable_group_superuser" | ||||
|         ) | ||||
|         if self.instance or superuser: | ||||
|             has_perm = user.has_perm(perm) or user.has_perm(perm, self.instance) | ||||
|             if not has_perm: | ||||
|                 raise ValidationError( | ||||
|                     _( | ||||
|                         ( | ||||
|                             "User does not have permission to set " | ||||
|                             "superuser status to {superuser_status}." | ||||
|                         ).format_map({"superuser_status": superuser}) | ||||
|                     ) | ||||
|         has_perm = user.has_perm(perm) | ||||
|         if self.instance and not has_perm: | ||||
|             has_perm = user.has_perm(perm, self.instance) | ||||
|         if not has_perm: | ||||
|             raise ValidationError( | ||||
|                 _( | ||||
|                     ( | ||||
|                         "User does not have permission to set " | ||||
|                         "superuser status to {superuser_status}." | ||||
|                     ).format_map({"superuser_status": superuser}) | ||||
|                 ) | ||||
|             ) | ||||
|         return superuser | ||||
|  | ||||
|     class Meta: | ||||
|  | ||||
| @ -6,6 +6,7 @@ from typing import Any | ||||
|  | ||||
| from django.contrib.auth import update_session_auth_hash | ||||
| from django.contrib.auth.models import Permission | ||||
| from django.db.models.functions import ExtractHour | ||||
| from django.db.transaction import atomic | ||||
| from django.db.utils import IntegrityError | ||||
| from django.urls import reverse_lazy | ||||
| @ -51,6 +52,7 @@ from rest_framework.validators import UniqueValidator | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.admin.api.metrics import CoordinateSerializer | ||||
| from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT | ||||
| from authentik.brands.models import Brand | ||||
| from authentik.core.api.used_by import UsedByMixin | ||||
| @ -82,7 +84,6 @@ from authentik.flows.views.executor import QS_KEY_TOKEN | ||||
| from authentik.lib.avatars import get_avatar | ||||
| from authentik.rbac.decorators import permission_required | ||||
| from authentik.rbac.models import get_permission_choices | ||||
| from authentik.stages.email.flow import pickle_flow_token_for_email | ||||
| from authentik.stages.email.models import EmailStage | ||||
| from authentik.stages.email.tasks import send_mails | ||||
| from authentik.stages.email.utils import TemplateEmailMessage | ||||
| @ -90,12 +91,6 @@ from authentik.stages.email.utils import TemplateEmailMessage | ||||
| LOGGER = get_logger() | ||||
|  | ||||
|  | ||||
| class ParamUserSerializer(PassiveSerializer): | ||||
|     """Partial serializer for query parameters to select a user""" | ||||
|  | ||||
|     user = PrimaryKeyRelatedField(queryset=User.objects.all().exclude_anonymous(), required=False) | ||||
|  | ||||
|  | ||||
| class UserGroupSerializer(ModelSerializer): | ||||
|     """Simplified Group Serializer for user's groups""" | ||||
|  | ||||
| @ -321,6 +316,53 @@ class SessionUserSerializer(PassiveSerializer): | ||||
|     original = UserSelfSerializer(required=False) | ||||
|  | ||||
|  | ||||
| class UserMetricsSerializer(PassiveSerializer): | ||||
|     """User Metrics""" | ||||
|  | ||||
|     logins = SerializerMethodField() | ||||
|     logins_failed = SerializerMethodField() | ||||
|     authorizations = SerializerMethodField() | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins(self, _): | ||||
|         """Get successful logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN, user__pk=user.pk | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_logins_failed(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.LOGIN_FAILED, context__username=user.username | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|     @extend_schema_field(CoordinateSerializer(many=True)) | ||||
|     def get_authorizations(self, _): | ||||
|         """Get failed logins per 8 hours for the last 7 days""" | ||||
|         user = self.context["user"] | ||||
|         request = self.context["request"] | ||||
|         return ( | ||||
|             get_objects_for_user(request.user, "authentik_events.view_event").filter( | ||||
|                 action=EventAction.AUTHORIZE_APPLICATION, user__pk=user.pk | ||||
|             ) | ||||
|             # 3 data points per day, so 8 hour spans | ||||
|             .get_events_per(timedelta(days=7), ExtractHour, 7 * 3) | ||||
|         ) | ||||
|  | ||||
|  | ||||
| class UsersFilter(FilterSet): | ||||
|     """Filter for users""" | ||||
|  | ||||
| @ -392,23 +434,8 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|     queryset = User.objects.none() | ||||
|     ordering = ["username"] | ||||
|     serializer_class = UserSerializer | ||||
|     filterset_class = UsersFilter | ||||
|     search_fields = ["username", "name", "is_active", "email", "uuid", "attributes"] | ||||
|  | ||||
|     def get_ql_fields(self): | ||||
|         from djangoql.schema import BoolField, StrField | ||||
|  | ||||
|         from authentik.enterprise.search.fields import ChoiceSearchField, JSONSearchField | ||||
|  | ||||
|         return [ | ||||
|             StrField(User, "username"), | ||||
|             StrField(User, "name"), | ||||
|             StrField(User, "email"), | ||||
|             StrField(User, "path"), | ||||
|             BoolField(User, "is_active", nullable=True), | ||||
|             ChoiceSearchField(User, "type"), | ||||
|             JSONSearchField(User, "attributes", suggest_nested=False), | ||||
|         ] | ||||
|     filterset_class = UsersFilter | ||||
|  | ||||
|     def get_queryset(self): | ||||
|         base_qs = User.objects.all().exclude_anonymous() | ||||
| @ -424,7 +451,7 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|     def list(self, request, *args, **kwargs): | ||||
|         return super().list(request, *args, **kwargs) | ||||
|  | ||||
|     def _create_recovery_link(self, for_email=False) -> tuple[str, Token]: | ||||
|     def _create_recovery_link(self) -> tuple[str, Token]: | ||||
|         """Create a recovery link (when the current brand has a recovery flow set), | ||||
|         that can either be shown to an admin or sent to the user directly""" | ||||
|         brand: Brand = self.request._request.brand | ||||
| @ -446,16 +473,12 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|             raise ValidationError( | ||||
|                 {"non_field_errors": "Recovery flow not applicable to user"} | ||||
|             ) from None | ||||
|         _plan = FlowToken.pickle(plan) | ||||
|         if for_email: | ||||
|             _plan = pickle_flow_token_for_email(plan) | ||||
|         token, __ = FlowToken.objects.update_or_create( | ||||
|             identifier=f"{user.uid}-password-reset", | ||||
|             defaults={ | ||||
|                 "user": user, | ||||
|                 "flow": flow, | ||||
|                 "_plan": _plan, | ||||
|                 "revoke_on_execution": not for_email, | ||||
|                 "_plan": FlowToken.pickle(plan), | ||||
|             }, | ||||
|         ) | ||||
|         querystring = urlencode({QS_KEY_TOKEN: token.key}) | ||||
| @ -579,6 +602,17 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|             update_session_auth_hash(self.request, user) | ||||
|         return Response(status=204) | ||||
|  | ||||
|     @permission_required("authentik_core.view_user", ["authentik_events.view_event"]) | ||||
|     @extend_schema(responses={200: UserMetricsSerializer(many=False)}) | ||||
|     @action(detail=True, pagination_class=None, filter_backends=[]) | ||||
|     def metrics(self, request: Request, pk: int) -> Response: | ||||
|         """User metrics per 1h""" | ||||
|         user: User = self.get_object() | ||||
|         serializer = UserMetricsSerializer(instance={}) | ||||
|         serializer.context["user"] = user | ||||
|         serializer.context["request"] = request | ||||
|         return Response(serializer.data) | ||||
|  | ||||
|     @permission_required("authentik_core.reset_user_password") | ||||
|     @extend_schema( | ||||
|         responses={ | ||||
| @ -614,7 +648,7 @@ class UserViewSet(UsedByMixin, ModelViewSet): | ||||
|         if for_user.email == "": | ||||
|             LOGGER.debug("User doesn't have an email address") | ||||
|             raise ValidationError({"non_field_errors": "User does not have an email address set."}) | ||||
|         link, token = self._create_recovery_link(for_email=True) | ||||
|         link, token = self._create_recovery_link() | ||||
|         # Lookup the email stage to assure the current user can access it | ||||
|         stages = get_objects_for_user( | ||||
|             request.user, "authentik_stages_email.view_emailstage" | ||||
|  | ||||
| @ -2,7 +2,6 @@ | ||||
|  | ||||
| from typing import Any | ||||
|  | ||||
| from django.db import models | ||||
| from django.db.models import Model | ||||
| from drf_spectacular.extensions import OpenApiSerializerFieldExtension | ||||
| from drf_spectacular.plumbing import build_basic_type | ||||
| @ -31,27 +30,7 @@ def is_dict(value: Any): | ||||
|     raise ValidationError("Value must be a dictionary, and not have any duplicate keys.") | ||||
|  | ||||
|  | ||||
| class JSONDictField(JSONField): | ||||
|     """JSON Field which only allows dictionaries""" | ||||
|  | ||||
|     default_validators = [is_dict] | ||||
|  | ||||
|  | ||||
| class JSONExtension(OpenApiSerializerFieldExtension): | ||||
|     """Generate API Schema for JSON fields as""" | ||||
|  | ||||
|     target_class = "authentik.core.api.utils.JSONDictField" | ||||
|  | ||||
|     def map_serializer_field(self, auto_schema, direction): | ||||
|         return build_basic_type(OpenApiTypes.OBJECT) | ||||
|  | ||||
|  | ||||
| class ModelSerializer(BaseModelSerializer): | ||||
|  | ||||
|     # By default, JSON fields we have are used to store dictionaries | ||||
|     serializer_field_mapping = BaseModelSerializer.serializer_field_mapping.copy() | ||||
|     serializer_field_mapping[models.JSONField] = JSONDictField | ||||
|  | ||||
|     def create(self, validated_data): | ||||
|         instance = super().create(validated_data) | ||||
|  | ||||
| @ -92,6 +71,21 @@ class ModelSerializer(BaseModelSerializer): | ||||
|         return instance | ||||
|  | ||||
|  | ||||
| class JSONDictField(JSONField): | ||||
|     """JSON Field which only allows dictionaries""" | ||||
|  | ||||
|     default_validators = [is_dict] | ||||
|  | ||||
|  | ||||
| class JSONExtension(OpenApiSerializerFieldExtension): | ||||
|     """Generate API Schema for JSON fields as""" | ||||
|  | ||||
|     target_class = "authentik.core.api.utils.JSONDictField" | ||||
|  | ||||
|     def map_serializer_field(self, auto_schema, direction): | ||||
|         return build_basic_type(OpenApiTypes.OBJECT) | ||||
|  | ||||
|  | ||||
| class PassiveSerializer(Serializer): | ||||
|     """Base serializer class which doesn't implement create/update methods""" | ||||
|  | ||||
|  | ||||
| @ -13,6 +13,7 @@ class Command(TenantCommand): | ||||
|         parser.add_argument("usernames", nargs="*", type=str) | ||||
|  | ||||
|     def handle_per_tenant(self, **options): | ||||
|         print(options) | ||||
|         new_type = UserTypes(options["type"]) | ||||
|         qs = ( | ||||
|             User.objects.exclude_anonymous() | ||||
|  | ||||
| @ -2,7 +2,6 @@ | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.contrib.auth.management import create_permissions | ||||
| from django.core.management import call_command | ||||
| from django.core.management.base import BaseCommand, no_translations | ||||
| from guardian.management import create_anonymous_user | ||||
|  | ||||
| @ -17,10 +16,6 @@ class Command(BaseCommand): | ||||
|         """Check permissions for all apps""" | ||||
|         for tenant in Tenant.objects.filter(ready=True): | ||||
|             with tenant: | ||||
|                 # See https://code.djangoproject.com/ticket/28417 | ||||
|                 # Remove potential lingering old permissions | ||||
|                 call_command("remove_stale_contenttypes", "--no-input") | ||||
|  | ||||
|                 for app in apps.get_app_configs(): | ||||
|                     self.stdout.write(f"Checking app {app.name} ({app.label})\n") | ||||
|                     create_permissions(app, verbosity=0) | ||||
|  | ||||
| @ -31,10 +31,7 @@ class PickleSerializer: | ||||
|  | ||||
|     def loads(self, data): | ||||
|         """Unpickle data to be loaded from redis""" | ||||
|         try: | ||||
|             return pickle.loads(data)  # nosec | ||||
|         except Exception: | ||||
|             return {} | ||||
|         return pickle.loads(data)  # nosec | ||||
|  | ||||
|  | ||||
| def _migrate_session( | ||||
| @ -79,7 +76,6 @@ def _migrate_session( | ||||
|         AuthenticatedSession.objects.using(db_alias).create( | ||||
|             session=session, | ||||
|             user=old_auth_session.user, | ||||
|             uuid=old_auth_session.uuid, | ||||
|         ) | ||||
|  | ||||
|  | ||||
|  | ||||
| @ -1,103 +0,0 @@ | ||||
| # Generated by Django 5.1.9 on 2025-05-14 11:15 | ||||
|  | ||||
| from django.apps.registry import Apps, apps as global_apps | ||||
| from django.db import migrations | ||||
| from django.contrib.contenttypes.management import create_contenttypes | ||||
| from django.contrib.auth.management import create_permissions | ||||
| from django.db.backends.base.schema import BaseDatabaseSchemaEditor | ||||
|  | ||||
|  | ||||
| def migrate_authenticated_session_permissions(apps: Apps, schema_editor: BaseDatabaseSchemaEditor): | ||||
|     """Migrate permissions from OldAuthenticatedSession to AuthenticatedSession""" | ||||
|     db_alias = schema_editor.connection.alias | ||||
|  | ||||
|     # `apps` here is just an instance of `django.db.migrations.state.AppConfigStub`, we need the | ||||
|     # real config for creating permissions and content types | ||||
|     authentik_core_config = global_apps.get_app_config("authentik_core") | ||||
|     # These are only ran by django after all migrations, but we need them right now. | ||||
|     # `global_apps` is needed, | ||||
|     create_permissions(authentik_core_config, using=db_alias, verbosity=1) | ||||
|     create_contenttypes(authentik_core_config, using=db_alias, verbosity=1) | ||||
|  | ||||
|     # But from now on, this is just a regular migration, so use `apps` | ||||
|     Permission = apps.get_model("auth", "Permission") | ||||
|     ContentType = apps.get_model("contenttypes", "ContentType") | ||||
|  | ||||
|     try: | ||||
|         old_ct = ContentType.objects.using(db_alias).get( | ||||
|             app_label="authentik_core", model="oldauthenticatedsession" | ||||
|         ) | ||||
|         new_ct = ContentType.objects.using(db_alias).get( | ||||
|             app_label="authentik_core", model="authenticatedsession" | ||||
|         ) | ||||
|     except ContentType.DoesNotExist: | ||||
|         # This should exist at this point, but if not, let's cut our losses | ||||
|         return | ||||
|  | ||||
|     # Get all permissions for the old content type | ||||
|     old_perms = Permission.objects.using(db_alias).filter(content_type=old_ct) | ||||
|  | ||||
|     # Create equivalent permissions for the new content type | ||||
|     for old_perm in old_perms: | ||||
|         new_perm = ( | ||||
|             Permission.objects.using(db_alias) | ||||
|             .filter( | ||||
|                 content_type=new_ct, | ||||
|                 codename=old_perm.codename, | ||||
|             ) | ||||
|             .first() | ||||
|         ) | ||||
|         if not new_perm: | ||||
|             # This should exist at this point, but if not, let's cut our losses | ||||
|             continue | ||||
|  | ||||
|         # Global user permissions | ||||
|         User = apps.get_model("authentik_core", "User") | ||||
|         User.user_permissions.through.objects.using(db_alias).filter( | ||||
|             permission=old_perm | ||||
|         ).all().update(permission=new_perm) | ||||
|  | ||||
|         # Global role permissions | ||||
|         DjangoGroup = apps.get_model("auth", "Group") | ||||
|         DjangoGroup.permissions.through.objects.using(db_alias).filter( | ||||
|             permission=old_perm | ||||
|         ).all().update(permission=new_perm) | ||||
|  | ||||
|         # Object user permissions | ||||
|         UserObjectPermission = apps.get_model("guardian", "UserObjectPermission") | ||||
|         UserObjectPermission.objects.using(db_alias).filter(permission=old_perm).all().update( | ||||
|             permission=new_perm, content_type=new_ct | ||||
|         ) | ||||
|  | ||||
|         # Object role permissions | ||||
|         GroupObjectPermission = apps.get_model("guardian", "GroupObjectPermission") | ||||
|         GroupObjectPermission.objects.using(db_alias).filter(permission=old_perm).all().update( | ||||
|             permission=new_perm, content_type=new_ct | ||||
|         ) | ||||
|  | ||||
|  | ||||
| def remove_old_authenticated_session_content_type( | ||||
|     apps: Apps, schema_editor: BaseDatabaseSchemaEditor | ||||
| ): | ||||
|     db_alias = schema_editor.connection.alias | ||||
|     ContentType = apps.get_model("contenttypes", "ContentType") | ||||
|  | ||||
|     ContentType.objects.using(db_alias).filter(model="oldauthenticatedsession").delete() | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
|     dependencies = [ | ||||
|         ("authentik_core", "0047_delete_oldauthenticatedsession"), | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.RunPython( | ||||
|             code=migrate_authenticated_session_permissions, | ||||
|             reverse_code=migrations.RunPython.noop, | ||||
|         ), | ||||
|         migrations.RunPython( | ||||
|             code=remove_old_authenticated_session_content_type, | ||||
|             reverse_code=migrations.RunPython.noop, | ||||
|         ), | ||||
|     ] | ||||
| @ -18,7 +18,7 @@ from django.http import HttpRequest | ||||
| from django.utils.functional import SimpleLazyObject, cached_property | ||||
| from django.utils.timezone import now | ||||
| from django.utils.translation import gettext_lazy as _ | ||||
| from django_cte import CTE, with_cte | ||||
| from django_cte import CTEQuerySet, With | ||||
| from guardian.conf import settings | ||||
| from guardian.mixins import GuardianUserMixin | ||||
| from model_utils.managers import InheritanceManager | ||||
| @ -136,7 +136,7 @@ class AttributesMixin(models.Model): | ||||
|         return instance, False | ||||
|  | ||||
|  | ||||
| class GroupQuerySet(QuerySet): | ||||
| class GroupQuerySet(CTEQuerySet): | ||||
|     def with_children_recursive(self): | ||||
|         """Recursively get all groups that have the current queryset as parents | ||||
|         or are indirectly related.""" | ||||
| @ -165,9 +165,9 @@ class GroupQuerySet(QuerySet): | ||||
|             ) | ||||
|  | ||||
|         # Build the recursive query, see above | ||||
|         cte = CTE.recursive(make_cte) | ||||
|         cte = With.recursive(make_cte) | ||||
|         # Return the result, as a usable queryset for Group. | ||||
|         return with_cte(cte, select=cte.join(Group, group_uuid=cte.col.group_uuid)) | ||||
|         return cte.join(Group, group_uuid=cte.col.group_uuid).with_cte(cte) | ||||
|  | ||||
|  | ||||
| class Group(SerializerModel, AttributesMixin): | ||||
| @ -1082,12 +1082,6 @@ class AuthenticatedSession(SerializerModel): | ||||
|  | ||||
|     user = models.ForeignKey(User, on_delete=models.CASCADE) | ||||
|  | ||||
|     @property | ||||
|     def serializer(self) -> type[Serializer]: | ||||
|         from authentik.core.api.authenticated_sessions import AuthenticatedSessionSerializer | ||||
|  | ||||
|         return AuthenticatedSessionSerializer | ||||
|  | ||||
|     class Meta: | ||||
|         verbose_name = _("Authenticated Session") | ||||
|         verbose_name_plural = _("Authenticated Sessions") | ||||
|  | ||||
| @ -16,14 +16,12 @@ | ||||
|         {% block head_before %} | ||||
|         {% endblock %} | ||||
|         <link rel="stylesheet" type="text/css" href="{% static 'dist/authentik.css' %}"> | ||||
|         <style>{{ brand_css }}</style> | ||||
|         <style>{{ brand.branding_custom_css }}</style> | ||||
|         <script src="{% versioned_script 'dist/poly-%v.js' %}" type="module"></script> | ||||
|         <script src="{% versioned_script 'dist/standalone/loading/index-%v.js' %}" type="module"></script> | ||||
|         {% block head %} | ||||
|         {% endblock %} | ||||
|         {% for key, value in html_meta.items %} | ||||
|         <meta name="{{key}}" content="{{ value }}" /> | ||||
|         {% endfor %} | ||||
|         <meta name="sentry-trace" content="{{ sentry_trace }}" /> | ||||
|     </head> | ||||
|     <body> | ||||
|         {% block body %} | ||||
|  | ||||
| @ -10,7 +10,7 @@ | ||||
| {% endblock %} | ||||
|  | ||||
| {% block body %} | ||||
| <ak-message-container alignment="bottom"></ak-message-container> | ||||
| <ak-message-container></ak-message-container> | ||||
| <ak-interface-admin> | ||||
|     <ak-loading></ak-loading> | ||||
| </ak-interface-admin> | ||||
|  | ||||
| @ -114,7 +114,6 @@ class TestApplicationsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
| @ -168,7 +167,6 @@ class TestApplicationsAPI(APITestCase): | ||||
|         self.assertJSONEqual( | ||||
|             response.content.decode(), | ||||
|             { | ||||
|                 "autocomplete": {}, | ||||
|                 "pagination": { | ||||
|                     "next": 0, | ||||
|                     "previous": 0, | ||||
|  | ||||
| @ -124,16 +124,6 @@ class TestGroupsAPI(APITestCase): | ||||
|             {"is_superuser": ["User does not have permission to set superuser status to True."]}, | ||||
|         ) | ||||
|  | ||||
|     def test_superuser_no_perm_no_superuser(self): | ||||
|         """Test creating a group without permission and without superuser flag""" | ||||
|         assign_perm("authentik_core.add_group", self.login_user) | ||||
|         self.client.force_login(self.login_user) | ||||
|         res = self.client.post( | ||||
|             reverse("authentik_api:group-list"), | ||||
|             data={"name": generate_id(), "is_superuser": False}, | ||||
|         ) | ||||
|         self.assertEqual(res.status_code, 201) | ||||
|  | ||||
|     def test_superuser_update_no_perm(self): | ||||
|         """Test updating a superuser group without permission""" | ||||
|         group = Group.objects.create(name=generate_id(), is_superuser=True) | ||||
|  | ||||
| @ -13,10 +13,7 @@ from authentik.core.models import ( | ||||
|     TokenIntents, | ||||
|     User, | ||||
| ) | ||||
| from authentik.core.tasks import ( | ||||
|     clean_expired_models, | ||||
|     clean_temporary_users, | ||||
| ) | ||||
| from authentik.core.tasks import clean_expired_models, clean_temporary_users | ||||
| from authentik.core.tests.utils import create_test_admin_user | ||||
| from authentik.lib.generators import generate_id | ||||
|  | ||||
|  | ||||
| @ -81,6 +81,22 @@ class TestUsersAPI(APITestCase): | ||||
|         response = self.client.get(reverse("authentik_api:user-list"), {"include_groups": "true"}) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_metrics(self): | ||||
|         """Test user's metrics""" | ||||
|         self.client.force_login(self.admin) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:user-metrics", kwargs={"pk": self.user.pk}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|  | ||||
|     def test_metrics_denied(self): | ||||
|         """Test user's metrics (non-superuser)""" | ||||
|         self.client.force_login(self.user) | ||||
|         response = self.client.get( | ||||
|             reverse("authentik_api:user-metrics", kwargs={"pk": self.user.pk}) | ||||
|         ) | ||||
|         self.assertEqual(response.status_code, 403) | ||||
|  | ||||
|     def test_recovery_no_flow(self): | ||||
|         """Test user recovery link (no recovery flow set)""" | ||||
|         self.client.force_login(self.admin) | ||||
|  | ||||
| @ -30,7 +30,6 @@ from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.core.api.used_by import UsedByMixin | ||||
| from authentik.core.api.utils import ModelSerializer, PassiveSerializer | ||||
| from authentik.core.models import UserTypes | ||||
| from authentik.crypto.apps import MANAGED_KEY | ||||
| from authentik.crypto.builder import CertificateBuilder, PrivateKeyAlg | ||||
| from authentik.crypto.models import CertificateKeyPair | ||||
| @ -273,12 +272,11 @@ class CertificateKeyPairViewSet(UsedByMixin, ModelViewSet): | ||||
|     def view_certificate(self, request: Request, pk: str) -> Response: | ||||
|         """Return certificate-key pairs certificate and log access""" | ||||
|         certificate: CertificateKeyPair = self.get_object() | ||||
|         if request.user.type != UserTypes.INTERNAL_SERVICE_ACCOUNT: | ||||
|             Event.new(  # noqa # nosec | ||||
|                 EventAction.SECRET_VIEW, | ||||
|                 secret=certificate, | ||||
|                 type="certificate", | ||||
|             ).from_http(request) | ||||
|         Event.new(  # noqa # nosec | ||||
|             EventAction.SECRET_VIEW, | ||||
|             secret=certificate, | ||||
|             type="certificate", | ||||
|         ).from_http(request) | ||||
|         if "download" in request.query_params: | ||||
|             # Mime type from https://pki-tutorial.readthedocs.io/en/latest/mime.html | ||||
|             response = HttpResponse( | ||||
| @ -304,12 +302,11 @@ class CertificateKeyPairViewSet(UsedByMixin, ModelViewSet): | ||||
|     def view_private_key(self, request: Request, pk: str) -> Response: | ||||
|         """Return certificate-key pairs private key and log access""" | ||||
|         certificate: CertificateKeyPair = self.get_object() | ||||
|         if request.user.type != UserTypes.INTERNAL_SERVICE_ACCOUNT: | ||||
|             Event.new(  # noqa # nosec | ||||
|                 EventAction.SECRET_VIEW, | ||||
|                 secret=certificate, | ||||
|                 type="private_key", | ||||
|             ).from_http(request) | ||||
|         Event.new(  # noqa # nosec | ||||
|             EventAction.SECRET_VIEW, | ||||
|             secret=certificate, | ||||
|             type="private_key", | ||||
|         ).from_http(request) | ||||
|         if "download" in request.query_params: | ||||
|             # Mime type from https://pki-tutorial.readthedocs.io/en/latest/mime.html | ||||
|             response = HttpResponse(certificate.key_data, content_type="application/x-pem-file") | ||||
|  | ||||
| @ -132,14 +132,13 @@ class LicenseKey: | ||||
|         """Get a summarized version of all (not expired) licenses""" | ||||
|         total = LicenseKey(get_license_aud(), 0, "Summarized license", 0, 0) | ||||
|         for lic in License.objects.all(): | ||||
|             if lic.is_valid: | ||||
|                 total.internal_users += lic.internal_users | ||||
|                 total.external_users += lic.external_users | ||||
|                 total.license_flags.extend(lic.status.license_flags) | ||||
|             total.internal_users += lic.internal_users | ||||
|             total.external_users += lic.external_users | ||||
|             exp_ts = int(mktime(lic.expiry.timetuple())) | ||||
|             if total.exp == 0: | ||||
|                 total.exp = exp_ts | ||||
|             total.exp = max(total.exp, exp_ts) | ||||
|             total.license_flags.extend(lic.status.license_flags) | ||||
|         return total | ||||
|  | ||||
|     @staticmethod | ||||
|  | ||||
| @ -39,10 +39,6 @@ class License(SerializerModel): | ||||
|     internal_users = models.BigIntegerField() | ||||
|     external_users = models.BigIntegerField() | ||||
|  | ||||
|     @property | ||||
|     def is_valid(self) -> bool: | ||||
|         return self.expiry >= now() | ||||
|  | ||||
|     @property | ||||
|     def serializer(self) -> type[BaseSerializer]: | ||||
|         from authentik.enterprise.api import LicenseSerializer | ||||
|  | ||||
| @ -1,27 +0,0 @@ | ||||
| from rest_framework.viewsets import ModelViewSet | ||||
|  | ||||
| from authentik.core.api.used_by import UsedByMixin | ||||
| from authentik.enterprise.api import EnterpriseRequiredMixin | ||||
| from authentik.enterprise.policies.unique_password.models import UniquePasswordPolicy | ||||
| from authentik.policies.api.policies import PolicySerializer | ||||
|  | ||||
|  | ||||
| class UniquePasswordPolicySerializer(EnterpriseRequiredMixin, PolicySerializer): | ||||
|     """Password Uniqueness Policy Serializer""" | ||||
|  | ||||
|     class Meta: | ||||
|         model = UniquePasswordPolicy | ||||
|         fields = PolicySerializer.Meta.fields + [ | ||||
|             "password_field", | ||||
|             "num_historical_passwords", | ||||
|         ] | ||||
|  | ||||
|  | ||||
| class UniquePasswordPolicyViewSet(UsedByMixin, ModelViewSet): | ||||
|     """Password Uniqueness Policy Viewset""" | ||||
|  | ||||
|     queryset = UniquePasswordPolicy.objects.all() | ||||
|     serializer_class = UniquePasswordPolicySerializer | ||||
|     filterset_fields = "__all__" | ||||
|     ordering = ["name"] | ||||
|     search_fields = ["name"] | ||||
| @ -1,10 +0,0 @@ | ||||
| """authentik Unique Password policy app config""" | ||||
|  | ||||
| from authentik.enterprise.apps import EnterpriseConfig | ||||
|  | ||||
|  | ||||
| class AuthentikEnterprisePoliciesUniquePasswordConfig(EnterpriseConfig): | ||||
|     name = "authentik.enterprise.policies.unique_password" | ||||
|     label = "authentik_policies_unique_password" | ||||
|     verbose_name = "authentik Enterprise.Policies.Unique Password" | ||||
|     default = True | ||||
| @ -1,81 +0,0 @@ | ||||
| # Generated by Django 5.0.13 on 2025-03-26 23:02 | ||||
|  | ||||
| import django.db.models.deletion | ||||
| from django.conf import settings | ||||
| from django.db import migrations, models | ||||
|  | ||||
|  | ||||
| class Migration(migrations.Migration): | ||||
|  | ||||
|     initial = True | ||||
|  | ||||
|     dependencies = [ | ||||
|         ("authentik_policies", "0011_policybinding_failure_result_and_more"), | ||||
|         migrations.swappable_dependency(settings.AUTH_USER_MODEL), | ||||
|     ] | ||||
|  | ||||
|     operations = [ | ||||
|         migrations.CreateModel( | ||||
|             name="UniquePasswordPolicy", | ||||
|             fields=[ | ||||
|                 ( | ||||
|                     "policy_ptr", | ||||
|                     models.OneToOneField( | ||||
|                         auto_created=True, | ||||
|                         on_delete=django.db.models.deletion.CASCADE, | ||||
|                         parent_link=True, | ||||
|                         primary_key=True, | ||||
|                         serialize=False, | ||||
|                         to="authentik_policies.policy", | ||||
|                     ), | ||||
|                 ), | ||||
|                 ( | ||||
|                     "password_field", | ||||
|                     models.TextField( | ||||
|                         default="password", | ||||
|                         help_text="Field key to check, field keys defined in Prompt stages are available.", | ||||
|                     ), | ||||
|                 ), | ||||
|                 ( | ||||
|                     "num_historical_passwords", | ||||
|                     models.PositiveIntegerField( | ||||
|                         default=1, help_text="Number of passwords to check against." | ||||
|                     ), | ||||
|                 ), | ||||
|             ], | ||||
|             options={ | ||||
|                 "verbose_name": "Password Uniqueness Policy", | ||||
|                 "verbose_name_plural": "Password Uniqueness Policies", | ||||
|                 "indexes": [ | ||||
|                     models.Index(fields=["policy_ptr_id"], name="authentik_p_policy__f559aa_idx") | ||||
|                 ], | ||||
|             }, | ||||
|             bases=("authentik_policies.policy",), | ||||
|         ), | ||||
|         migrations.CreateModel( | ||||
|             name="UserPasswordHistory", | ||||
|             fields=[ | ||||
|                 ( | ||||
|                     "id", | ||||
|                     models.AutoField( | ||||
|                         auto_created=True, primary_key=True, serialize=False, verbose_name="ID" | ||||
|                     ), | ||||
|                 ), | ||||
|                 ("old_password", models.CharField(max_length=128)), | ||||
|                 ("created_at", models.DateTimeField(auto_now_add=True)), | ||||
|                 ("hibp_prefix_sha1", models.CharField(max_length=5)), | ||||
|                 ("hibp_pw_hash", models.TextField()), | ||||
|                 ( | ||||
|                     "user", | ||||
|                     models.ForeignKey( | ||||
|                         on_delete=django.db.models.deletion.CASCADE, | ||||
|                         related_name="old_passwords", | ||||
|                         to=settings.AUTH_USER_MODEL, | ||||
|                     ), | ||||
|                 ), | ||||
|             ], | ||||
|             options={ | ||||
|                 "verbose_name": "User Password History", | ||||
|             }, | ||||
|         ), | ||||
|     ] | ||||
| @ -1,151 +0,0 @@ | ||||
| from hashlib import sha1 | ||||
|  | ||||
| from django.contrib.auth.hashers import identify_hasher, make_password | ||||
| from django.db import models | ||||
| from django.utils.translation import gettext as _ | ||||
| from rest_framework.serializers import BaseSerializer | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.core.models import User | ||||
| from authentik.policies.models import Policy | ||||
| from authentik.policies.types import PolicyRequest, PolicyResult | ||||
| from authentik.stages.prompt.stage import PLAN_CONTEXT_PROMPT | ||||
|  | ||||
| LOGGER = get_logger() | ||||
|  | ||||
|  | ||||
| class UniquePasswordPolicy(Policy): | ||||
|     """This policy prevents users from reusing old passwords.""" | ||||
|  | ||||
|     password_field = models.TextField( | ||||
|         default="password", | ||||
|         help_text=_("Field key to check, field keys defined in Prompt stages are available."), | ||||
|     ) | ||||
|  | ||||
|     # Limit on the number of previous passwords the policy evaluates | ||||
|     # Also controls number of old passwords the system stores. | ||||
|     num_historical_passwords = models.PositiveIntegerField( | ||||
|         default=1, | ||||
|         help_text=_("Number of passwords to check against."), | ||||
|     ) | ||||
|  | ||||
|     @property | ||||
|     def serializer(self) -> type[BaseSerializer]: | ||||
|         from authentik.enterprise.policies.unique_password.api import UniquePasswordPolicySerializer | ||||
|  | ||||
|         return UniquePasswordPolicySerializer | ||||
|  | ||||
|     @property | ||||
|     def component(self) -> str: | ||||
|         return "ak-policy-password-uniqueness-form" | ||||
|  | ||||
|     def passes(self, request: PolicyRequest) -> PolicyResult: | ||||
|         from authentik.enterprise.policies.unique_password.models import UserPasswordHistory | ||||
|  | ||||
|         password = request.context.get(PLAN_CONTEXT_PROMPT, {}).get( | ||||
|             self.password_field, request.context.get(self.password_field) | ||||
|         ) | ||||
|         if not password: | ||||
|             LOGGER.warning( | ||||
|                 "Password field not found in request when checking UniquePasswordPolicy", | ||||
|                 field=self.password_field, | ||||
|                 fields=request.context.keys(), | ||||
|             ) | ||||
|             return PolicyResult(False, _("Password not set in context")) | ||||
|         password = str(password) | ||||
|  | ||||
|         if not self.num_historical_passwords: | ||||
|             # Policy not configured to check against any passwords | ||||
|             return PolicyResult(True) | ||||
|  | ||||
|         num_to_check = self.num_historical_passwords | ||||
|         password_history = UserPasswordHistory.objects.filter(user=request.user).order_by( | ||||
|             "-created_at" | ||||
|         )[:num_to_check] | ||||
|  | ||||
|         if not password_history: | ||||
|             return PolicyResult(True) | ||||
|  | ||||
|         for record in password_history: | ||||
|             if not record.old_password: | ||||
|                 continue | ||||
|  | ||||
|             if self._passwords_match(new_password=password, old_password=record.old_password): | ||||
|                 # Return on first match. Authentik does not consider timing attacks | ||||
|                 # on old passwords to be an attack surface. | ||||
|                 return PolicyResult( | ||||
|                     False, | ||||
|                     _("This password has been used previously. Please choose a different one."), | ||||
|                 ) | ||||
|  | ||||
|         return PolicyResult(True) | ||||
|  | ||||
|     def _passwords_match(self, *, new_password: str, old_password: str) -> bool: | ||||
|         try: | ||||
|             hasher = identify_hasher(old_password) | ||||
|         except ValueError: | ||||
|             LOGGER.warning( | ||||
|                 "Skipping password; could not load hash algorithm", | ||||
|             ) | ||||
|             return False | ||||
|  | ||||
|         return hasher.verify(new_password, old_password) | ||||
|  | ||||
|     @classmethod | ||||
|     def is_in_use(cls): | ||||
|         """Check if any UniquePasswordPolicy is in use, either through policy bindings | ||||
|         or direct attachment to a PromptStage. | ||||
|  | ||||
|         Returns: | ||||
|             bool: True if any policy is in use, False otherwise | ||||
|         """ | ||||
|         from authentik.policies.models import PolicyBinding | ||||
|  | ||||
|         # Check if any policy is in use through bindings | ||||
|         if PolicyBinding.in_use.for_policy(cls).exists(): | ||||
|             return True | ||||
|  | ||||
|         # Check if any policy is attached to a PromptStage | ||||
|         if cls.objects.filter(promptstage__isnull=False).exists(): | ||||
|             return True | ||||
|  | ||||
|         return False | ||||
|  | ||||
|     class Meta(Policy.PolicyMeta): | ||||
|         verbose_name = _("Password Uniqueness Policy") | ||||
|         verbose_name_plural = _("Password Uniqueness Policies") | ||||
|  | ||||
|  | ||||
| class UserPasswordHistory(models.Model): | ||||
|     user = models.ForeignKey(User, on_delete=models.CASCADE, related_name="old_passwords") | ||||
|     # Mimic's column type of AbstractBaseUser.password | ||||
|     old_password = models.CharField(max_length=128) | ||||
|     created_at = models.DateTimeField(auto_now_add=True) | ||||
|  | ||||
|     hibp_prefix_sha1 = models.CharField(max_length=5) | ||||
|     hibp_pw_hash = models.TextField() | ||||
|  | ||||
|     class Meta: | ||||
|         verbose_name = _("User Password History") | ||||
|  | ||||
|     def __str__(self) -> str: | ||||
|         timestamp = f"{self.created_at:%Y/%m/%d %X}" if self.created_at else "N/A" | ||||
|         return f"Previous Password (user: {self.user_id}, recorded: {timestamp})" | ||||
|  | ||||
|     @classmethod | ||||
|     def create_for_user(cls, user: User, password: str): | ||||
|         # To check users' passwords against Have I been Pwned, we need the first 5 chars | ||||
|         # of the password hashed with SHA1 without a salt... | ||||
|         pw_hash_sha1 = sha1(password.encode("utf-8")).hexdigest()  # nosec | ||||
|         # ...however that'll give us a list of hashes from HIBP, and to compare that we still | ||||
|         # need a full unsalted SHA1 of the password. We don't want to save that directly in | ||||
|         # the database, so we hash that SHA1 again with a modern hashing alg, | ||||
|         # and then when we check users' passwords against HIBP we can use `check_password` | ||||
|         # which will take care of this. | ||||
|         hibp_hash_hash = make_password(pw_hash_sha1) | ||||
|         return cls.objects.create( | ||||
|             user=user, | ||||
|             old_password=password, | ||||
|             hibp_prefix_sha1=pw_hash_sha1[:5], | ||||
|             hibp_pw_hash=hibp_hash_hash, | ||||
|         ) | ||||
| @ -1,20 +0,0 @@ | ||||
| """Unique Password Policy settings""" | ||||
|  | ||||
| from celery.schedules import crontab | ||||
|  | ||||
| from authentik.lib.utils.time import fqdn_rand | ||||
|  | ||||
| CELERY_BEAT_SCHEDULE = { | ||||
|     "policies_unique_password_trim_history": { | ||||
|         "task": "authentik.enterprise.policies.unique_password.tasks.trim_password_histories", | ||||
|         "schedule": crontab(minute=fqdn_rand("policies_unique_password_trim"), hour="*/12"), | ||||
|         "options": {"queue": "authentik_scheduled"}, | ||||
|     }, | ||||
|     "policies_unique_password_check_purge": { | ||||
|         "task": ( | ||||
|             "authentik.enterprise.policies.unique_password.tasks.check_and_purge_password_history" | ||||
|         ), | ||||
|         "schedule": crontab(minute=fqdn_rand("policies_unique_password_purge"), hour="*/24"), | ||||
|         "options": {"queue": "authentik_scheduled"}, | ||||
|     }, | ||||
| } | ||||
| @ -1,23 +0,0 @@ | ||||
| """authentik policy signals""" | ||||
|  | ||||
| from django.dispatch import receiver | ||||
|  | ||||
| from authentik.core.models import User | ||||
| from authentik.core.signals import password_changed | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
|  | ||||
|  | ||||
| @receiver(password_changed) | ||||
| def copy_password_to_password_history(sender, user: User, *args, **kwargs): | ||||
|     """Preserve the user's old password if UniquePasswordPolicy is enabled anywhere""" | ||||
|     # Check if any UniquePasswordPolicy is in use | ||||
|     unique_pwd_policy_in_use = UniquePasswordPolicy.is_in_use() | ||||
|  | ||||
|     if unique_pwd_policy_in_use: | ||||
|         """NOTE: Because we run this in a signal after saving the user, | ||||
|         we are not atomically guaranteed to save password history. | ||||
|         """ | ||||
|         UserPasswordHistory.create_for_user(user, user.password) | ||||
| @ -1,66 +0,0 @@ | ||||
| from django.db.models.aggregates import Count | ||||
| from structlog import get_logger | ||||
|  | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
| from authentik.events.system_tasks import SystemTask, TaskStatus, prefill_task | ||||
| from authentik.root.celery import CELERY_APP | ||||
|  | ||||
| LOGGER = get_logger() | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task(bind=True, base=SystemTask) | ||||
| @prefill_task | ||||
| def check_and_purge_password_history(self: SystemTask): | ||||
|     """Check if any UniquePasswordPolicy exists, and if not, purge the password history table. | ||||
|     This is run on a schedule instead of being triggered by policy binding deletion. | ||||
|     """ | ||||
|     if not UniquePasswordPolicy.objects.exists(): | ||||
|         UserPasswordHistory.objects.all().delete() | ||||
|         LOGGER.debug("Purged UserPasswordHistory table as no policies are in use") | ||||
|         self.set_status(TaskStatus.SUCCESSFUL, "Successfully purged UserPasswordHistory") | ||||
|         return | ||||
|  | ||||
|     self.set_status( | ||||
|         TaskStatus.SUCCESSFUL, "Not purging password histories, a unique password policy exists" | ||||
|     ) | ||||
|  | ||||
|  | ||||
| @CELERY_APP.task(bind=True, base=SystemTask) | ||||
| def trim_password_histories(self: SystemTask): | ||||
|     """Removes rows from UserPasswordHistory older than | ||||
|     the `n` most recent entries. | ||||
|  | ||||
|     The `n` is defined by the largest configured value for all bound | ||||
|     UniquePasswordPolicy policies. | ||||
|     """ | ||||
|  | ||||
|     # No policy, we'll let the cleanup above do its thing | ||||
|     if not UniquePasswordPolicy.objects.exists(): | ||||
|         return | ||||
|  | ||||
|     num_rows_to_preserve = 0 | ||||
|     for policy in UniquePasswordPolicy.objects.all(): | ||||
|         num_rows_to_preserve = max(num_rows_to_preserve, policy.num_historical_passwords) | ||||
|  | ||||
|     all_pks_to_keep = [] | ||||
|  | ||||
|     # Get all users who have password history entries | ||||
|     users_with_history = ( | ||||
|         UserPasswordHistory.objects.values("user") | ||||
|         .annotate(count=Count("user")) | ||||
|         .filter(count__gt=0) | ||||
|         .values_list("user", flat=True) | ||||
|     ) | ||||
|     for user_pk in users_with_history: | ||||
|         entries = UserPasswordHistory.objects.filter(user__pk=user_pk) | ||||
|         pks_to_keep = entries.order_by("-created_at")[:num_rows_to_preserve].values_list( | ||||
|             "pk", flat=True | ||||
|         ) | ||||
|         all_pks_to_keep.extend(pks_to_keep) | ||||
|  | ||||
|     num_deleted, _ = UserPasswordHistory.objects.exclude(pk__in=all_pks_to_keep).delete() | ||||
|     LOGGER.debug("Deleted stale password history records", count=num_deleted) | ||||
|     self.set_status(TaskStatus.SUCCESSFUL, f"Delete {num_deleted} stale password history records") | ||||
| @ -1,108 +0,0 @@ | ||||
| """Unique Password Policy flow tests""" | ||||
|  | ||||
| from django.contrib.auth.hashers import make_password | ||||
| from django.urls.base import reverse | ||||
|  | ||||
| from authentik.core.tests.utils import create_test_flow, create_test_user | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
| from authentik.flows.models import FlowDesignation, FlowStageBinding | ||||
| from authentik.flows.tests import FlowTestCase | ||||
| from authentik.lib.generators import generate_id | ||||
| from authentik.stages.prompt.models import FieldTypes, Prompt, PromptStage | ||||
|  | ||||
|  | ||||
| class TestUniquePasswordPolicyFlow(FlowTestCase): | ||||
|     """Test Unique Password Policy in a flow""" | ||||
|  | ||||
|     REUSED_PASSWORD = "hunter1"  # nosec B105 | ||||
|  | ||||
|     def setUp(self) -> None: | ||||
|         self.user = create_test_user() | ||||
|         self.flow = create_test_flow(FlowDesignation.AUTHENTICATION) | ||||
|  | ||||
|         password_prompt = Prompt.objects.create( | ||||
|             name=generate_id(), | ||||
|             field_key="password", | ||||
|             label="PASSWORD_LABEL", | ||||
|             type=FieldTypes.PASSWORD, | ||||
|             required=True, | ||||
|             placeholder="PASSWORD_PLACEHOLDER", | ||||
|         ) | ||||
|  | ||||
|         self.policy = UniquePasswordPolicy.objects.create( | ||||
|             name="password_must_unique", | ||||
|             password_field=password_prompt.field_key, | ||||
|             num_historical_passwords=1, | ||||
|         ) | ||||
|         stage = PromptStage.objects.create(name="prompt-stage") | ||||
|         stage.validation_policies.set([self.policy]) | ||||
|         stage.fields.set( | ||||
|             [ | ||||
|                 password_prompt, | ||||
|             ] | ||||
|         ) | ||||
|         FlowStageBinding.objects.create(target=self.flow, stage=stage, order=2) | ||||
|  | ||||
|         # Seed the user's password history | ||||
|         UserPasswordHistory.create_for_user(self.user, make_password(self.REUSED_PASSWORD)) | ||||
|  | ||||
|     def test_prompt_data(self): | ||||
|         """Test policy attached to a prompt stage""" | ||||
|         # Test the policy directly | ||||
|         from authentik.policies.types import PolicyRequest | ||||
|         from authentik.stages.prompt.stage import PLAN_CONTEXT_PROMPT | ||||
|  | ||||
|         # Create a policy request with the reused password | ||||
|         request = PolicyRequest(user=self.user) | ||||
|         request.context[PLAN_CONTEXT_PROMPT] = {"password": self.REUSED_PASSWORD} | ||||
|  | ||||
|         # Test the policy directly | ||||
|         result = self.policy.passes(request) | ||||
|  | ||||
|         # Verify that the policy fails (returns False) with the expected error message | ||||
|         self.assertFalse(result.passing, "Policy should fail for reused password") | ||||
|         self.assertEqual( | ||||
|             result.messages[0], | ||||
|             "This password has been used previously. Please choose a different one.", | ||||
|             "Incorrect error message", | ||||
|         ) | ||||
|  | ||||
|         # API-based testing approach: | ||||
|  | ||||
|         self.client.force_login(self.user) | ||||
|  | ||||
|         # Send a POST request to the flow executor with the reused password | ||||
|         response = self.client.post( | ||||
|             reverse("authentik_api:flow-executor", kwargs={"flow_slug": self.flow.slug}), | ||||
|             {"password": self.REUSED_PASSWORD}, | ||||
|         ) | ||||
|         self.assertStageResponse( | ||||
|             response, | ||||
|             self.flow, | ||||
|             component="ak-stage-prompt", | ||||
|             fields=[ | ||||
|                 { | ||||
|                     "choices": None, | ||||
|                     "field_key": "password", | ||||
|                     "label": "PASSWORD_LABEL", | ||||
|                     "order": 0, | ||||
|                     "placeholder": "PASSWORD_PLACEHOLDER", | ||||
|                     "initial_value": "", | ||||
|                     "required": True, | ||||
|                     "type": "password", | ||||
|                     "sub_text": "", | ||||
|                 } | ||||
|             ], | ||||
|             response_errors={ | ||||
|                 "non_field_errors": [ | ||||
|                     { | ||||
|                         "code": "invalid", | ||||
|                         "string": "This password has been used previously. " | ||||
|                         "Please choose a different one.", | ||||
|                     } | ||||
|                 ] | ||||
|             }, | ||||
|         ) | ||||
| @ -1,77 +0,0 @@ | ||||
| """Unique Password Policy tests""" | ||||
|  | ||||
| from django.contrib.auth.hashers import make_password | ||||
| from django.test import TestCase | ||||
| from guardian.shortcuts import get_anonymous_user | ||||
|  | ||||
| from authentik.core.models import User | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
| from authentik.policies.types import PolicyRequest, PolicyResult | ||||
| from authentik.stages.prompt.stage import PLAN_CONTEXT_PROMPT | ||||
|  | ||||
|  | ||||
| class TestUniquePasswordPolicy(TestCase): | ||||
|     """Test Password Uniqueness Policy""" | ||||
|  | ||||
|     def setUp(self) -> None: | ||||
|         self.policy = UniquePasswordPolicy.objects.create( | ||||
|             name="test_unique_password", num_historical_passwords=1 | ||||
|         ) | ||||
|         self.user = User.objects.create(username="test-user") | ||||
|  | ||||
|     def test_invalid(self): | ||||
|         """Test without password present in request""" | ||||
|         request = PolicyRequest(get_anonymous_user()) | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertFalse(result.passing) | ||||
|         self.assertEqual(result.messages[0], "Password not set in context") | ||||
|  | ||||
|     def test_passes_no_previous_passwords(self): | ||||
|         request = PolicyRequest(get_anonymous_user()) | ||||
|         request.context = {PLAN_CONTEXT_PROMPT: {"password": "hunter2"}} | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertTrue(result.passing) | ||||
|  | ||||
|     def test_passes_passwords_are_different(self): | ||||
|         # Seed database with an old password | ||||
|         UserPasswordHistory.create_for_user(self.user, make_password("hunter1")) | ||||
|  | ||||
|         request = PolicyRequest(self.user) | ||||
|         request.context = {PLAN_CONTEXT_PROMPT: {"password": "hunter2"}} | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertTrue(result.passing) | ||||
|  | ||||
|     def test_passes_multiple_old_passwords(self): | ||||
|         # Seed with multiple old passwords | ||||
|         UserPasswordHistory.objects.bulk_create( | ||||
|             [ | ||||
|                 UserPasswordHistory(user=self.user, old_password=make_password("hunter1")), | ||||
|                 UserPasswordHistory(user=self.user, old_password=make_password("hunter2")), | ||||
|             ] | ||||
|         ) | ||||
|         request = PolicyRequest(self.user) | ||||
|         request.context = {PLAN_CONTEXT_PROMPT: {"password": "hunter3"}} | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertTrue(result.passing) | ||||
|  | ||||
|     def test_fails_password_matches_old_password(self): | ||||
|         # Seed database with an old password | ||||
|  | ||||
|         UserPasswordHistory.create_for_user(self.user, make_password("hunter1")) | ||||
|  | ||||
|         request = PolicyRequest(self.user) | ||||
|         request.context = {PLAN_CONTEXT_PROMPT: {"password": "hunter1"}} | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertFalse(result.passing) | ||||
|  | ||||
|     def test_fails_if_identical_password_with_different_hash_algos(self): | ||||
|         UserPasswordHistory.create_for_user( | ||||
|             self.user, make_password("hunter2", "somesalt", "scrypt") | ||||
|         ) | ||||
|         request = PolicyRequest(self.user) | ||||
|         request.context = {PLAN_CONTEXT_PROMPT: {"password": "hunter2"}} | ||||
|         result: PolicyResult = self.policy.passes(request) | ||||
|         self.assertFalse(result.passing) | ||||
| @ -1,90 +0,0 @@ | ||||
| from django.urls import reverse | ||||
|  | ||||
| from authentik.core.models import Group, Source, User | ||||
| from authentik.core.tests.utils import create_test_flow, create_test_user | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
| from authentik.flows.markers import StageMarker | ||||
| from authentik.flows.models import FlowStageBinding | ||||
| from authentik.flows.planner import PLAN_CONTEXT_PENDING_USER, FlowPlan | ||||
| from authentik.flows.tests import FlowTestCase | ||||
| from authentik.flows.views.executor import SESSION_KEY_PLAN | ||||
| from authentik.lib.generators import generate_key | ||||
| from authentik.policies.models import PolicyBinding, PolicyBindingModel | ||||
| from authentik.stages.prompt.stage import PLAN_CONTEXT_PROMPT | ||||
| from authentik.stages.user_write.models import UserWriteStage | ||||
|  | ||||
|  | ||||
| class TestUserWriteStage(FlowTestCase): | ||||
|     """Write tests""" | ||||
|  | ||||
|     def setUp(self): | ||||
|         super().setUp() | ||||
|         self.flow = create_test_flow() | ||||
|         self.group = Group.objects.create(name="test-group") | ||||
|         self.other_group = Group.objects.create(name="other-group") | ||||
|         self.stage: UserWriteStage = UserWriteStage.objects.create( | ||||
|             name="write", create_users_as_inactive=True, create_users_group=self.group | ||||
|         ) | ||||
|         self.binding = FlowStageBinding.objects.create(target=self.flow, stage=self.stage, order=2) | ||||
|         self.source = Source.objects.create(name="fake_source") | ||||
|  | ||||
|     def test_save_password_history_if_policy_binding_enforced(self): | ||||
|         """Test user's new password is recorded when ANY enabled UniquePasswordPolicy exists""" | ||||
|         unique_password_policy = UniquePasswordPolicy.objects.create(num_historical_passwords=5) | ||||
|         pbm = PolicyBindingModel.objects.create() | ||||
|         PolicyBinding.objects.create( | ||||
|             target=pbm, policy=unique_password_policy, order=0, enabled=True | ||||
|         ) | ||||
|  | ||||
|         test_user = create_test_user() | ||||
|         # Store original password for verification | ||||
|         original_password = test_user.password | ||||
|  | ||||
|         # We're changing our own password | ||||
|         self.client.force_login(test_user) | ||||
|  | ||||
|         new_password = generate_key() | ||||
|         plan = FlowPlan(flow_pk=self.flow.pk.hex, bindings=[self.binding], markers=[StageMarker()]) | ||||
|         plan.context[PLAN_CONTEXT_PENDING_USER] = test_user | ||||
|         plan.context[PLAN_CONTEXT_PROMPT] = { | ||||
|             "username": test_user.username, | ||||
|             "password": new_password, | ||||
|         } | ||||
|         session = self.client.session | ||||
|         session[SESSION_KEY_PLAN] = plan | ||||
|         session.save() | ||||
|         # Password history should be recorded | ||||
|         user_password_history_qs = UserPasswordHistory.objects.filter(user=test_user) | ||||
|         self.assertTrue(user_password_history_qs.exists(), "Password history should be recorded") | ||||
|         self.assertEqual(len(user_password_history_qs), 1, "expected 1 recorded password") | ||||
|  | ||||
|         # Create a password history entry manually to simulate the signal behavior | ||||
|         # This is what would happen if the signal worked correctly | ||||
|         UserPasswordHistory.objects.create(user=test_user, old_password=original_password) | ||||
|         user_password_history_qs = UserPasswordHistory.objects.filter(user=test_user) | ||||
|         self.assertTrue(user_password_history_qs.exists(), "Password history should be recorded") | ||||
|         self.assertEqual(len(user_password_history_qs), 2, "expected 2 recorded password") | ||||
|  | ||||
|         # Execute the flow by sending a POST request to the flow executor endpoint | ||||
|         response = self.client.post( | ||||
|             reverse("authentik_api:flow-executor", kwargs={"flow_slug": self.flow.slug}) | ||||
|         ) | ||||
|  | ||||
|         # Verify that the request was successful | ||||
|         self.assertEqual(response.status_code, 200) | ||||
|         user_qs = User.objects.filter(username=plan.context[PLAN_CONTEXT_PROMPT]["username"]) | ||||
|         self.assertTrue(user_qs.exists()) | ||||
|  | ||||
|         # Verify the password history entry exists | ||||
|         user_password_history_qs = UserPasswordHistory.objects.filter(user=test_user) | ||||
|         self.assertTrue(user_password_history_qs.exists(), "Password history should be recorded") | ||||
|  | ||||
|         self.assertEqual(len(user_password_history_qs), 3, "expected 3 recorded password") | ||||
|         # Verify that one of the entries contains the original password | ||||
|         self.assertTrue( | ||||
|             any(entry.old_password == original_password for entry in user_password_history_qs), | ||||
|             "original password should be in password history table", | ||||
|         ) | ||||
| @ -1,178 +0,0 @@ | ||||
| from datetime import datetime, timedelta | ||||
|  | ||||
| from django.test import TestCase | ||||
|  | ||||
| from authentik.core.tests.utils import create_test_user | ||||
| from authentik.enterprise.policies.unique_password.models import ( | ||||
|     UniquePasswordPolicy, | ||||
|     UserPasswordHistory, | ||||
| ) | ||||
| from authentik.enterprise.policies.unique_password.tasks import ( | ||||
|     check_and_purge_password_history, | ||||
|     trim_password_histories, | ||||
| ) | ||||
| from authentik.policies.models import PolicyBinding, PolicyBindingModel | ||||
|  | ||||
|  | ||||
| class TestUniquePasswordPolicyModel(TestCase): | ||||
|     """Test the UniquePasswordPolicy model methods""" | ||||
|  | ||||
|     def test_is_in_use_with_binding(self): | ||||
|         """Test is_in_use returns True when a policy binding exists""" | ||||
|         # Create a UniquePasswordPolicy and a PolicyBinding for it | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=5) | ||||
|         pbm = PolicyBindingModel.objects.create() | ||||
|         PolicyBinding.objects.create(target=pbm, policy=policy, order=0, enabled=True) | ||||
|  | ||||
|         # Verify is_in_use returns True | ||||
|         self.assertTrue(UniquePasswordPolicy.is_in_use()) | ||||
|  | ||||
|     def test_is_in_use_with_promptstage(self): | ||||
|         """Test is_in_use returns True when attached to a PromptStage""" | ||||
|         from authentik.stages.prompt.models import PromptStage | ||||
|  | ||||
|         # Create a UniquePasswordPolicy and attach it to a PromptStage | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=5) | ||||
|         prompt_stage = PromptStage.objects.create( | ||||
|             name="Test Prompt Stage", | ||||
|         ) | ||||
|         # Use the set() method for many-to-many relationships | ||||
|         prompt_stage.validation_policies.set([policy]) | ||||
|  | ||||
|         # Verify is_in_use returns True | ||||
|         self.assertTrue(UniquePasswordPolicy.is_in_use()) | ||||
|  | ||||
|  | ||||
| class TestTrimAllPasswordHistories(TestCase): | ||||
|     """Test the task that trims password history for all users""" | ||||
|  | ||||
|     def setUp(self): | ||||
|         self.user1 = create_test_user("test-user1") | ||||
|         self.user2 = create_test_user("test-user2") | ||||
|         self.pbm = PolicyBindingModel.objects.create() | ||||
|         # Create a policy with a limit of 1 password | ||||
|         self.policy = UniquePasswordPolicy.objects.create(num_historical_passwords=1) | ||||
|         PolicyBinding.objects.create( | ||||
|             target=self.pbm, | ||||
|             policy=self.policy, | ||||
|             enabled=True, | ||||
|             order=0, | ||||
|         ) | ||||
|  | ||||
|  | ||||
| class TestCheckAndPurgePasswordHistory(TestCase): | ||||
|     """Test the scheduled task that checks if any policy is in use and purges if not""" | ||||
|  | ||||
|     def setUp(self): | ||||
|         self.user = create_test_user("test-user") | ||||
|         self.pbm = PolicyBindingModel.objects.create() | ||||
|  | ||||
|     def test_purge_when_no_policy_in_use(self): | ||||
|         """Test that the task purges the table when no policy is in use""" | ||||
|         # Create some password history entries | ||||
|         UserPasswordHistory.create_for_user(self.user, "hunter2") | ||||
|  | ||||
|         # Verify we have entries | ||||
|         self.assertTrue(UserPasswordHistory.objects.exists()) | ||||
|  | ||||
|         # Run the task - should purge since no policy is in use | ||||
|         check_and_purge_password_history() | ||||
|  | ||||
|         # Verify the table is empty | ||||
|         self.assertFalse(UserPasswordHistory.objects.exists()) | ||||
|  | ||||
|     def test_no_purge_when_policy_in_use(self): | ||||
|         """Test that the task doesn't purge when a policy is in use""" | ||||
|         # Create a policy and binding | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=5) | ||||
|         PolicyBinding.objects.create( | ||||
|             target=self.pbm, | ||||
|             policy=policy, | ||||
|             enabled=True, | ||||
|             order=0, | ||||
|         ) | ||||
|  | ||||
|         # Create some password history entries | ||||
|         UserPasswordHistory.create_for_user(self.user, "hunter2") | ||||
|  | ||||
|         # Verify we have entries | ||||
|         self.assertTrue(UserPasswordHistory.objects.exists()) | ||||
|  | ||||
|         # Run the task - should NOT purge since a policy is in use | ||||
|         check_and_purge_password_history() | ||||
|  | ||||
|         # Verify the entries still exist | ||||
|         self.assertTrue(UserPasswordHistory.objects.exists()) | ||||
|  | ||||
|  | ||||
| class TestTrimPasswordHistory(TestCase): | ||||
|     """Test password history cleanup task""" | ||||
|  | ||||
|     def setUp(self): | ||||
|         self.user = create_test_user("test-user") | ||||
|         self.pbm = PolicyBindingModel.objects.create() | ||||
|  | ||||
|     def test_trim_password_history_ok(self): | ||||
|         """Test passwords over the define limit are deleted""" | ||||
|         _now = datetime.now() | ||||
|         UserPasswordHistory.objects.bulk_create( | ||||
|             [ | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter1",  # nosec | ||||
|                     created_at=_now - timedelta(days=3), | ||||
|                 ), | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter2",  # nosec | ||||
|                     created_at=_now - timedelta(days=2), | ||||
|                 ), | ||||
|                 UserPasswordHistory( | ||||
|                     user=self.user, | ||||
|                     old_password="hunter3",  # nosec | ||||
|                     created_at=_now, | ||||
|                 ), | ||||
|             ] | ||||
|         ) | ||||
|  | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=1) | ||||
|         PolicyBinding.objects.create( | ||||
|             target=self.pbm, | ||||
|             policy=policy, | ||||
|             enabled=True, | ||||
|             order=0, | ||||
|         ) | ||||
|         trim_password_histories.delay() | ||||
|         user_pwd_history_qs = UserPasswordHistory.objects.filter(user=self.user) | ||||
|         self.assertEqual(len(user_pwd_history_qs), 1) | ||||
|  | ||||
|     def test_trim_password_history_policy_diabled_no_op(self): | ||||
|         """Test no passwords removed if policy binding is disabled""" | ||||
|  | ||||
|         # Insert a record to ensure it's not deleted after executing task | ||||
|         UserPasswordHistory.create_for_user(self.user, "hunter2") | ||||
|  | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=1) | ||||
|         PolicyBinding.objects.create( | ||||
|             target=self.pbm, | ||||
|             policy=policy, | ||||
|             enabled=False, | ||||
|             order=0, | ||||
|         ) | ||||
|         trim_password_histories.delay() | ||||
|         self.assertTrue(UserPasswordHistory.objects.filter(user=self.user).exists()) | ||||
|  | ||||
|     def test_trim_password_history_fewer_records_than_maximum_is_no_op(self): | ||||
|         """Test no passwords deleted if fewer passwords exist than limit""" | ||||
|  | ||||
|         UserPasswordHistory.create_for_user(self.user, "hunter2") | ||||
|  | ||||
|         policy = UniquePasswordPolicy.objects.create(num_historical_passwords=2) | ||||
|         PolicyBinding.objects.create( | ||||
|             target=self.pbm, | ||||
|             policy=policy, | ||||
|             enabled=True, | ||||
|             order=0, | ||||
|         ) | ||||
|         trim_password_histories.delay() | ||||
|         self.assertTrue(UserPasswordHistory.objects.filter(user=self.user).exists()) | ||||
| @ -1,7 +0,0 @@ | ||||
| """API URLs""" | ||||
|  | ||||
| from authentik.enterprise.policies.unique_password.api import UniquePasswordPolicyViewSet | ||||
|  | ||||
| api_urlpatterns = [ | ||||
|     ("policies/unique_password", UniquePasswordPolicyViewSet), | ||||
| ] | ||||
| @ -25,7 +25,7 @@ class GoogleWorkspaceGroupClient( | ||||
|     """Google client for groups""" | ||||
|  | ||||
|     connection_type = GoogleWorkspaceProviderGroup | ||||
|     connection_attr = "googleworkspaceprovidergroup_set" | ||||
|     connection_type_query = "group" | ||||
|     can_discover = True | ||||
|  | ||||
|     def __init__(self, provider: GoogleWorkspaceProvider) -> None: | ||||
|  | ||||
| @ -20,7 +20,7 @@ class GoogleWorkspaceUserClient(GoogleWorkspaceSyncClient[User, GoogleWorkspaceP | ||||
|     """Sync authentik users into google workspace""" | ||||
|  | ||||
|     connection_type = GoogleWorkspaceProviderUser | ||||
|     connection_attr = "googleworkspaceprovideruser_set" | ||||
|     connection_type_query = "user" | ||||
|     can_discover = True | ||||
|  | ||||
|     def __init__(self, provider: GoogleWorkspaceProvider) -> None: | ||||
|  | ||||
| @ -132,11 +132,7 @@ class GoogleWorkspaceProvider(OutgoingSyncProvider, BackchannelProvider): | ||||
|         if type == User: | ||||
|             # Get queryset of all users with consistent ordering | ||||
|             # according to the provider's settings | ||||
|             base = ( | ||||
|                 User.objects.prefetch_related("googleworkspaceprovideruser_set") | ||||
|                 .all() | ||||
|                 .exclude_anonymous() | ||||
|             ) | ||||
|             base = User.objects.all().exclude_anonymous() | ||||
|             if self.exclude_users_service_account: | ||||
|                 base = base.exclude(type=UserTypes.SERVICE_ACCOUNT).exclude( | ||||
|                     type=UserTypes.INTERNAL_SERVICE_ACCOUNT | ||||
| @ -146,11 +142,7 @@ class GoogleWorkspaceProvider(OutgoingSyncProvider, BackchannelProvider): | ||||
|             return base.order_by("pk") | ||||
|         if type == Group: | ||||
|             # Get queryset of all groups with consistent ordering | ||||
|             return ( | ||||
|                 Group.objects.prefetch_related("googleworkspaceprovidergroup_set") | ||||
|                 .all() | ||||
|                 .order_by("pk") | ||||
|             ) | ||||
|             return Group.objects.all().order_by("pk") | ||||
|         raise ValueError(f"Invalid type {type}") | ||||
|  | ||||
|     def google_credentials(self): | ||||
|  | ||||
| @ -29,7 +29,7 @@ class MicrosoftEntraGroupClient( | ||||
|     """Microsoft client for groups""" | ||||
|  | ||||
|     connection_type = MicrosoftEntraProviderGroup | ||||
|     connection_attr = "microsoftentraprovidergroup_set" | ||||
|     connection_type_query = "group" | ||||
|     can_discover = True | ||||
|  | ||||
|     def __init__(self, provider: MicrosoftEntraProvider) -> None: | ||||
|  | ||||
| @ -24,7 +24,7 @@ class MicrosoftEntraUserClient(MicrosoftEntraSyncClient[User, MicrosoftEntraProv | ||||
|     """Sync authentik users into microsoft entra""" | ||||
|  | ||||
|     connection_type = MicrosoftEntraProviderUser | ||||
|     connection_attr = "microsoftentraprovideruser_set" | ||||
|     connection_type_query = "user" | ||||
|     can_discover = True | ||||
|  | ||||
|     def __init__(self, provider: MicrosoftEntraProvider) -> None: | ||||
|  | ||||
| @ -121,11 +121,7 @@ class MicrosoftEntraProvider(OutgoingSyncProvider, BackchannelProvider): | ||||
|         if type == User: | ||||
|             # Get queryset of all users with consistent ordering | ||||
|             # according to the provider's settings | ||||
|             base = ( | ||||
|                 User.objects.prefetch_related("microsoftentraprovideruser_set") | ||||
|                 .all() | ||||
|                 .exclude_anonymous() | ||||
|             ) | ||||
|             base = User.objects.all().exclude_anonymous() | ||||
|             if self.exclude_users_service_account: | ||||
|                 base = base.exclude(type=UserTypes.SERVICE_ACCOUNT).exclude( | ||||
|                     type=UserTypes.INTERNAL_SERVICE_ACCOUNT | ||||
| @ -135,11 +131,7 @@ class MicrosoftEntraProvider(OutgoingSyncProvider, BackchannelProvider): | ||||
|             return base.order_by("pk") | ||||
|         if type == Group: | ||||
|             # Get queryset of all groups with consistent ordering | ||||
|             return ( | ||||
|                 Group.objects.prefetch_related("microsoftentraprovidergroup_set") | ||||
|                 .all() | ||||
|                 .order_by("pk") | ||||
|             ) | ||||
|             return Group.objects.all().order_by("pk") | ||||
|         raise ValueError(f"Invalid type {type}") | ||||
|  | ||||
|     def microsoft_credentials(self): | ||||
|  | ||||
| @ -1,8 +1,10 @@ | ||||
| from hashlib import sha256 | ||||
|  | ||||
| from django.contrib.auth.signals import user_logged_out | ||||
| from django.db.models import Model | ||||
| from django.db.models.signals import post_delete, post_save, pre_delete | ||||
| from django.dispatch import receiver | ||||
| from django.http.request import HttpRequest | ||||
| from guardian.shortcuts import assign_perm | ||||
|  | ||||
| from authentik.core.models import ( | ||||
| @ -60,6 +62,31 @@ def ssf_providers_post_save(sender: type[Model], instance: SSFProvider, created: | ||||
|             instance.save() | ||||
|  | ||||
|  | ||||
| @receiver(user_logged_out) | ||||
| def ssf_user_logged_out_session_revoked(sender, request: HttpRequest, user: User, **_): | ||||
|     """Session revoked trigger (user logged out)""" | ||||
|     if not request.session or not request.session.session_key or not user: | ||||
|         return | ||||
|     send_ssf_event( | ||||
|         EventTypes.CAEP_SESSION_REVOKED, | ||||
|         { | ||||
|             "initiating_entity": "user", | ||||
|         }, | ||||
|         sub_id={ | ||||
|             "format": "complex", | ||||
|             "session": { | ||||
|                 "format": "opaque", | ||||
|                 "id": sha256(request.session.session_key.encode("ascii")).hexdigest(), | ||||
|             }, | ||||
|             "user": { | ||||
|                 "format": "email", | ||||
|                 "email": user.email, | ||||
|             }, | ||||
|         }, | ||||
|         request=request, | ||||
|     ) | ||||
|  | ||||
|  | ||||
| @receiver(pre_delete, sender=AuthenticatedSession) | ||||
| def ssf_user_session_delete_session_revoked(sender, instance: AuthenticatedSession, **_): | ||||
|     """Session revoked trigger (users' session has been deleted) | ||||
|  | ||||
| @ -1,12 +0,0 @@ | ||||
| """Enterprise app config""" | ||||
|  | ||||
| from authentik.enterprise.apps import EnterpriseConfig | ||||
|  | ||||
|  | ||||
| class AuthentikEnterpriseSearchConfig(EnterpriseConfig): | ||||
|     """Enterprise app config""" | ||||
|  | ||||
|     name = "authentik.enterprise.search" | ||||
|     label = "authentik_search" | ||||
|     verbose_name = "authentik Enterprise.Search" | ||||
|     default = True | ||||
| @ -1,128 +0,0 @@ | ||||
| """DjangoQL search""" | ||||
|  | ||||
| from collections import OrderedDict, defaultdict | ||||
| from collections.abc import Generator | ||||
|  | ||||
| from django.db import connection | ||||
| from django.db.models import Model, Q | ||||
| from djangoql.compat import text_type | ||||
| from djangoql.schema import StrField | ||||
|  | ||||
|  | ||||
| class JSONSearchField(StrField): | ||||
|     """JSON field for DjangoQL""" | ||||
|  | ||||
|     model: Model | ||||
|  | ||||
|     def __init__(self, model=None, name=None, nullable=None, suggest_nested=True): | ||||
|         # Set this in the constructor to not clobber the type variable | ||||
|         self.type = "relation" | ||||
|         self.suggest_nested = suggest_nested | ||||
|         super().__init__(model, name, nullable) | ||||
|  | ||||
|     def get_lookup(self, path, operator, value): | ||||
|         search = "__".join(path) | ||||
|         op, invert = self.get_operator(operator) | ||||
|         q = Q(**{f"{search}{op}": self.get_lookup_value(value)}) | ||||
|         return ~q if invert else q | ||||
|  | ||||
|     def json_field_keys(self) -> Generator[tuple[str]]: | ||||
|         with connection.cursor() as cursor: | ||||
|             cursor.execute( | ||||
|                 f""" | ||||
|                 WITH RECURSIVE "{self.name}_keys" AS ( | ||||
|                     SELECT | ||||
|                         ARRAY[jsonb_object_keys("{self.name}")] AS key_path_array, | ||||
|                         "{self.name}" -> jsonb_object_keys("{self.name}") AS value | ||||
|                     FROM {self.model._meta.db_table} | ||||
|                     WHERE "{self.name}" IS NOT NULL | ||||
|                         AND jsonb_typeof("{self.name}") = 'object' | ||||
|  | ||||
|                     UNION ALL | ||||
|  | ||||
|                     SELECT | ||||
|                         ck.key_path_array || jsonb_object_keys(ck.value), | ||||
|                         ck.value -> jsonb_object_keys(ck.value) AS value | ||||
|                     FROM "{self.name}_keys" ck | ||||
|                     WHERE jsonb_typeof(ck.value) = 'object' | ||||
|                 ), | ||||
|  | ||||
|                 unique_paths AS ( | ||||
|                     SELECT DISTINCT key_path_array | ||||
|                     FROM "{self.name}_keys" | ||||
|                 ) | ||||
|  | ||||
|                 SELECT key_path_array FROM unique_paths; | ||||
|             """  # nosec | ||||
|             ) | ||||
|             return (x[0] for x in cursor.fetchall()) | ||||
|  | ||||
|     def get_nested_options(self) -> OrderedDict: | ||||
|         """Get keys of all nested objects to show autocomplete""" | ||||
|         if not self.suggest_nested: | ||||
|             return OrderedDict() | ||||
|         base_model_name = f"{self.model._meta.app_label}.{self.model._meta.model_name}_{self.name}" | ||||
|  | ||||
|         def recursive_function(parts: list[str], parent_parts: list[str] | None = None): | ||||
|             if not parent_parts: | ||||
|                 parent_parts = [] | ||||
|             path = parts.pop(0) | ||||
|             parent_parts.append(path) | ||||
|             relation_key = "_".join(parent_parts) | ||||
|             if len(parts) > 1: | ||||
|                 out_dict = { | ||||
|                     relation_key: { | ||||
|                         parts[0]: { | ||||
|                             "type": "relation", | ||||
|                             "relation": f"{relation_key}_{parts[0]}", | ||||
|                         } | ||||
|                     } | ||||
|                 } | ||||
|                 child_paths = recursive_function(parts.copy(), parent_parts.copy()) | ||||
|                 child_paths.update(out_dict) | ||||
|                 return child_paths | ||||
|             else: | ||||
|                 return {relation_key: {parts[0]: {}}} | ||||
|  | ||||
|         relation_structure = defaultdict(dict) | ||||
|  | ||||
|         for relations in self.json_field_keys(): | ||||
|             result = recursive_function([base_model_name] + relations) | ||||
|             for relation_key, value in result.items(): | ||||
|                 for sub_relation_key, sub_value in value.items(): | ||||
|                     if not relation_structure[relation_key].get(sub_relation_key, None): | ||||
|                         relation_structure[relation_key][sub_relation_key] = sub_value | ||||
|                     else: | ||||
|                         relation_structure[relation_key][sub_relation_key].update(sub_value) | ||||
|  | ||||
|         final_dict = defaultdict(dict) | ||||
|  | ||||
|         for key, value in relation_structure.items(): | ||||
|             for sub_key, sub_value in value.items(): | ||||
|                 if not sub_value: | ||||
|                     final_dict[key][sub_key] = { | ||||
|                         "type": "str", | ||||
|                         "nullable": True, | ||||
|                     } | ||||
|                 else: | ||||
|                     final_dict[key][sub_key] = sub_value | ||||
|         return OrderedDict(final_dict) | ||||
|  | ||||
|     def relation(self) -> str: | ||||
|         return f"{self.model._meta.app_label}.{self.model._meta.model_name}_{self.name}" | ||||
|  | ||||
|  | ||||
| class ChoiceSearchField(StrField): | ||||
|     def __init__(self, model=None, name=None, nullable=None): | ||||
|         super().__init__(model, name, nullable, suggest_options=True) | ||||
|  | ||||
|     def get_options(self, search): | ||||
|         result = [] | ||||
|         choices = self._field_choices() | ||||
|         if choices: | ||||
|             search = search.lower() | ||||
|             for c in choices: | ||||
|                 choice = text_type(c[0]) | ||||
|                 if search in choice.lower(): | ||||
|                     result.append(choice) | ||||
|         return result | ||||
| @ -1,53 +0,0 @@ | ||||
| from rest_framework.response import Response | ||||
|  | ||||
| from authentik.api.pagination import Pagination | ||||
| from authentik.enterprise.search.ql import AUTOCOMPLETE_COMPONENT_NAME, QLSearch | ||||
|  | ||||
|  | ||||
| class AutocompletePagination(Pagination): | ||||
|  | ||||
|     def paginate_queryset(self, queryset, request, view=None): | ||||
|         self.view = view | ||||
|         return super().paginate_queryset(queryset, request, view) | ||||
|  | ||||
|     def get_autocomplete(self): | ||||
|         schema = QLSearch().get_schema(self.request, self.view) | ||||
|         introspections = {} | ||||
|         if hasattr(self.view, "get_ql_fields"): | ||||
|             from authentik.enterprise.search.schema import AKQLSchemaSerializer | ||||
|  | ||||
|             introspections = AKQLSchemaSerializer().serialize( | ||||
|                 schema(self.page.paginator.object_list.model) | ||||
|             ) | ||||
|         return introspections | ||||
|  | ||||
|     def get_paginated_response(self, data): | ||||
|         previous_page_number = 0 | ||||
|         if self.page.has_previous(): | ||||
|             previous_page_number = self.page.previous_page_number() | ||||
|         next_page_number = 0 | ||||
|         if self.page.has_next(): | ||||
|             next_page_number = self.page.next_page_number() | ||||
|         return Response( | ||||
|             { | ||||
|                 "pagination": { | ||||
|                     "next": next_page_number, | ||||
|                     "previous": previous_page_number, | ||||
|                     "count": self.page.paginator.count, | ||||
|                     "current": self.page.number, | ||||
|                     "total_pages": self.page.paginator.num_pages, | ||||
|                     "start_index": self.page.start_index(), | ||||
|                     "end_index": self.page.end_index(), | ||||
|                 }, | ||||
|                 "results": data, | ||||
|                 "autocomplete": self.get_autocomplete(), | ||||
|             } | ||||
|         ) | ||||
|  | ||||
|     def get_paginated_response_schema(self, schema): | ||||
|         final_schema = super().get_paginated_response_schema(schema) | ||||
|         final_schema["properties"]["autocomplete"] = { | ||||
|             "$ref": f"#/components/schemas/{AUTOCOMPLETE_COMPONENT_NAME}" | ||||
|         } | ||||
|         final_schema["required"].append("autocomplete") | ||||
|         return final_schema | ||||
| @ -1,80 +0,0 @@ | ||||
| """DjangoQL search""" | ||||
|  | ||||
| from django.apps import apps | ||||
| from django.db.models import QuerySet | ||||
| from djangoql.ast import Name | ||||
| from djangoql.exceptions import DjangoQLError | ||||
| from djangoql.queryset import apply_search | ||||
| from djangoql.schema import DjangoQLSchema | ||||
| from rest_framework.filters import BaseFilterBackend, SearchFilter | ||||
| from rest_framework.request import Request | ||||
| from structlog.stdlib import get_logger | ||||
|  | ||||
| from authentik.enterprise.search.fields import JSONSearchField | ||||
|  | ||||
| LOGGER = get_logger() | ||||
| AUTOCOMPLETE_COMPONENT_NAME = "Autocomplete" | ||||
| AUTOCOMPLETE_SCHEMA = { | ||||
|     "type": "object", | ||||
|     "additionalProperties": {}, | ||||
| } | ||||
|  | ||||
|  | ||||
| class BaseSchema(DjangoQLSchema): | ||||
|     """Base Schema which deals with JSON Fields""" | ||||
|  | ||||
|     def resolve_name(self, name: Name): | ||||
|         model = self.model_label(self.current_model) | ||||
|         root_field = name.parts[0] | ||||
|         field = self.models[model].get(root_field) | ||||
|         # If the query goes into a JSON field, return the root | ||||
|         # field as the JSON field will do the rest | ||||
|         if isinstance(field, JSONSearchField): | ||||
|             # This is a workaround; build_filter will remove the right-most | ||||
|             # entry in the path as that is intended to be the same as the field | ||||
|             # however for JSON that is not the case | ||||
|             if name.parts[-1] != root_field: | ||||
|                 name.parts.append(root_field) | ||||
|             return field | ||||
|         return super().resolve_name(name) | ||||
|  | ||||
|  | ||||
| class QLSearch(BaseFilterBackend): | ||||
|     """rest_framework search filter which uses DjangoQL""" | ||||
|  | ||||
|     def __init__(self): | ||||
|         super().__init__() | ||||
|         self._fallback = SearchFilter() | ||||
|  | ||||
|     @property | ||||
|     def enabled(self): | ||||
|         return apps.get_app_config("authentik_enterprise").enabled() | ||||
|  | ||||
|     def get_search_terms(self, request: Request) -> str: | ||||
|         """Search terms are set by a ?search=... query parameter, | ||||
|         and may be comma and/or whitespace delimited.""" | ||||
|         params = request.query_params.get("search", "") | ||||
|         params = params.replace("\x00", "")  # strip null characters | ||||
|         return params | ||||
|  | ||||
|     def get_schema(self, request: Request, view) -> BaseSchema: | ||||
|         ql_fields = [] | ||||
|         if hasattr(view, "get_ql_fields"): | ||||
|             ql_fields = view.get_ql_fields() | ||||
|  | ||||
|         class InlineSchema(BaseSchema): | ||||
|             def get_fields(self, model): | ||||
|                 return ql_fields or [] | ||||
|  | ||||
|         return InlineSchema | ||||
|  | ||||
|     def filter_queryset(self, request: Request, queryset: QuerySet, view) -> QuerySet: | ||||
|         search_query = self.get_search_terms(request) | ||||
|         schema = self.get_schema(request, view) | ||||
|         if len(search_query) == 0 or not self.enabled: | ||||
|             return self._fallback.filter_queryset(request, queryset, view) | ||||
|         try: | ||||
|             return apply_search(queryset, search_query, schema=schema) | ||||
|         except DjangoQLError as exc: | ||||
|             LOGGER.debug("Failed to parse search expression", exc=exc) | ||||
|             return self._fallback.filter_queryset(request, queryset, view) | ||||
| @ -1,29 +0,0 @@ | ||||
| from djangoql.serializers import DjangoQLSchemaSerializer | ||||
| from drf_spectacular.generators import SchemaGenerator | ||||
|  | ||||
| from authentik.api.schema import create_component | ||||
| from authentik.enterprise.search.fields import JSONSearchField | ||||
| from authentik.enterprise.search.ql import AUTOCOMPLETE_COMPONENT_NAME, AUTOCOMPLETE_SCHEMA | ||||
|  | ||||
|  | ||||
| class AKQLSchemaSerializer(DjangoQLSchemaSerializer): | ||||
|     def serialize(self, schema): | ||||
|         serialization = super().serialize(schema) | ||||
|         for _, fields in schema.models.items(): | ||||
|             for _, field in fields.items(): | ||||
|                 if not isinstance(field, JSONSearchField): | ||||
|                     continue | ||||
|                 serialization["models"].update(field.get_nested_options()) | ||||
|         return serialization | ||||
|  | ||||
|     def serialize_field(self, field): | ||||
|         result = super().serialize_field(field) | ||||
|         if isinstance(field, JSONSearchField): | ||||
|             result["relation"] = field.relation() | ||||
|         return result | ||||
|  | ||||
|  | ||||
| def postprocess_schema_search_autocomplete(result, generator: SchemaGenerator, **kwargs): | ||||
|     create_component(generator, AUTOCOMPLETE_COMPONENT_NAME, AUTOCOMPLETE_SCHEMA) | ||||
|  | ||||
|     return result | ||||
Some files were not shown because too many files have changed in this diff Show More
		Reference in New Issue
	
	Block a user
	