Run Codex from a GitHub Actions workflow while keeping tight control over the privileges available to Codex. This action handles installing the Codex CLI and configuring it with a secure proxy to the Responses API.
Users must provide an API key for their chosen provider (for example, OPENAI_API_KEY or AZURE_OPENAI_API_KEY if using Azure for OpenAI models) as a GitHub Actions secret to use this action.
While Codex cloud offers a powerful code review tool that you can use today, here is an example of how you can build your own code review workflow with openai/codex-action if you want to have more control over the experience.
In the following example, we define a workflow that is triggered whenever a user creates a pull request that:
- Creates a shallow clone of the repo.
- Ensures the
baseandheadrefs for the PR are available locally. - Runs Codex with a
promptthat includes the details specific to the PR. - Takes the output from Codex and posts it as a comment on the PR.
See security.md for tips on using openai/codex-action securely.
name: Perform a code review when a pull request is created.
on:
pull_request:
types: [opened]
jobs:
codex:
runs-on: ubuntu-latest
permissions:
contents: read
outputs:
final_message: ${{ steps.run_codex.outputs.final-message }}
steps:
- uses: actions/checkout@v5
with:
# Explicitly check out the PR's merge commit.
ref: refs/pull/${{ github.event.pull_request.number }}/merge
- name: Pre-fetch base and head refs for the PR
run: |
git fetch --no-tags origin \
${{ github.event.pull_request.base.ref }} \
+refs/pull/${{ github.event.pull_request.number }}/head
# If you want Codex to build and run code, install any dependencies that
# need to be downloaded before the "Run Codex" step because Codex's
# default sandbox disables network access.
- name: Run Codex
id: run_codex
uses: openai/codex-action@v1
with:
openai-api-key: ${{ secrets.OPENAI_API_KEY }}
prompt: |
This is PR #${{ github.event.pull_request.number }} for ${{ github.repository }}.
Review ONLY the changes introduced by the PR, so consider:
git log --oneline ${{ github.event.pull_request.base.sha }}...${{ github.event.pull_request.head.sha }}
Suggest any improvements, potential bugs, or issues.
Be concise and specific in your feedback.
Pull request title and body:
----
${{ github.event.pull_request.title }}
${{ github.event.pull_request.body }}
post_feedback:
runs-on: ubuntu-latest
needs: codex
if: needs.codex.outputs.final_message != ''
permissions:
issues: write
pull-requests: write
steps:
- name: Report Codex feedback
uses: actions/github-script@v7
env:
CODEX_FINAL_MESSAGE: ${{ needs.codex.outputs.final_message }}
with:
github-token: ${{ github.token }}
script: |
await github.rest.issues.createComment({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.payload.pull_request.number,
body: process.env.CODEX_FINAL_MESSAGE,
});| Name | Description | Default |
|---|---|---|
openai-api-key |
Secret used to start the Responses API proxy when you are using OpenAI (default). Store it in secrets. |
"" |
responses-api-endpoint |
Optional Responses API endpoint override, e.g. https://example.openai.azure.com/openai/v1/responses. Leave empty to use the proxy's default. |
"" |
prompt |
Inline prompt text. Provide this or prompt-file. |
"" |
prompt-file |
Path (relative to the repository root) of a file that contains the prompt. Provide this or prompt. |
"" |
output-file |
File where the final Codex message is written. Leave empty to skip writing a file. | "" |
working-directory |
Directory passed to codex exec --cd. Defaults to the repository root. |
"" |
sandbox |
Sandbox mode for Codex. One of workspace-write (default), read-only or danger-full-access. |
"" |
codex-version |
Version of @openai/codex to install. |
"" |
codex-args |
Extra arguments forwarded to codex exec. Accepts JSON arrays (["--flag", "value"]) or shell-style strings. |
"" |
output-schema |
Inline schema contents written to a temp file and passed to codex exec --output-schema. Mutually exclusive with output-schema-file. |
"" |
output-schema-file |
Schema file forwarded to codex exec --output-schema. Leave empty to skip passing the option. |
"" |
model |
Model the agent should use. Leave empty to let Codex pick its default. | "" |
effort |
Reasoning effort the agent should use. Leave empty to let Codex pick its default. | "" |
codex-home |
Directory to use as the Codex CLI home (config/cache). Uses the CLI default when empty. | "" |
safety-strategy |
Controls how the action restricts Codex privileges. See Safety strategy. | drop-sudo |
codex-user |
Username to run Codex as when safety-strategy is unprivileged-user. |
"" |
allow-users |
List of GitHub usernames who can trigger the action in addition to those who have write access to the repo. | "" |
allow-bots |
Allow runs triggered by GitHub Apps/bot accounts to bypass the write-access check. | false |
The safety-strategy input determines how much access Codex receives on the runner. Choosing the right option is critical, especially when sensitive secrets (like your OpenAI API key) are present.
See Protecting your OPENAI_API_KEY on the Security page for important details on this topic.
drop-sudo(default) — On Linux and macOS runners, the action revokes the default user’ssudomembership before invoking Codex. Codex then runs as that user without superuser privileges. This change lasts for the rest of the job, so subsequent steps cannot rely onsudo. This is usually the safest choice on GitHub-hosted runners.unprivileged-user— Runs Codex as the user provided viacodex-user. Use this if you manage your own runner with a pre-created unprivileged account. Ensure the user can read the repository checkout and any files Codex needs. Seeunprivileged-user.ymlfor an example of how to configure such an account onubuntu-latest.read-only— Executes Codex in a read-only sandbox. Codex can view files but cannot mutate the filesystem or access the network directly. The OpenAI API key still flows through the proxy, so Codex could read it if it can reach process memory.unsafe— No privilege reduction. Codex runs as the defaultrunneruser (which typically hassudo). Only use this when you fully trust the prompt. On Windows runners this is the only supported choice and the action will fail if another option is provided.
- Windows: GitHub-hosted Windows runners lack a supported sandbox. Set
safety-strategy: unsafe. The action validates this and exits early otherwise. - Linux/macOS: All options for
safety-strategyare supported. Again, if you pickdrop-sudo, remember that later steps in yourjobthat rely onsudowill fail. If you do need to run code that requiressudoafteropenai/codex-actionhas run, one option is to pipe the output ofopenai/codex-actionto a freshjobon a new host and to continue your workflow from there.
| Name | Description |
|---|---|
final-message |
Final message returned by codex exec. |
As we saw in the example above, we took the final-message output of the run_codex step and made it an output of the codex job in the workflow:
jobs:
codex:
# ...
outputs:
final_message: ${{ steps.run_codex.outputs.final-message }}- Run this action after
actions/checkout@v5so Codex has access to your repository contents. - To use a non-default Responses endpoint (for example Azure OpenAI), set
responses-api-endpointto the provider's URL while keepingopenai-api-keypopulated; the proxy will still sendAuthorization: Bearer <key>upstream. - If you want Codex to have access to a narrow set of privileged functionality, consider running a local MCP server that can perform these actions and configure Codex to use it.
- If you need more control over the CLI invocation, pass flags through
codex-argsor create aconfig.tomlincodex-home. - Once
openai/codex-actionis run once withopenai-api-key, you can also callcodexfrom subsequent scripts in your job. (You can omitpromptandprompt-filefrom the action in this case.)
To configure the Action to use OpenAI models hosted on Azure, pay close attention to the following:
- The
responses-api-endpointmust be set to the full URL (including any required query parameters) that Codex willPOSTto for a Responses API request. For Azure, this might look likehttps://YOUR_PROJECT_NAME.openai.azure.com/openai/v1/responses. Note that unlike when customizing a model provider in Codex, you must include thev1/responsessuffix to the URL yourself, if appropriate. - The
openai-api-keyinput must be a valid key that can be used with theAuthorization: Bearer <KEY>header when making aPOSTrequest to your Responses API endpoint. (This is also true for the value of theenv_keywhen setting a custom provider using the Codex CLI.)
Ultimately, your configured Action might look something like the following:
- name: Start Codex proxy
uses: openai/codex-action@v1
with:
openai-api-key: ${{ secrets.AZURE_OPENAI_API_KEY }}
responses-api-endpoint: "https://bolinfest-7804-resource.cognitiveservices.azure.com/openai/v1/responses"
prompt: "Debug all the things."See the CHANGELOG for details.
This project is licensed under the Apache License 2.0.