Setup Ollama

Install and configure Ollama to work with Job matcher

Download and Install

Download Ollama from the official page: https://ollama.com/download

The installation process varies slightly depending on your operating system.

Choosing model

Llama3 is a good model to use as baseline. If you have more capable machine we recommend models from OpenAI like gpt-oss-20b.

To use llama3, run this on terminal:

  ollama pull llama3
  ollama run llama3

To use OpenAI model:

  ollama pull gpt-oss-20b
  ollama run gpt-oss-20b

Change the model to use in the Criteria setting on Match page.

Allow Job Matcher to access Ollama

By default local Ollama instance is not accessible from external websites for security reasons. We need to explicitly allow Job Matcher to talk to our local Ollama.

MacOS

Run the following command:

launchctl setenv OLLAMA_ORIGINS "https://job-matcher.senyas.io"

Restart the Ollama application to apply the changes.

Linux

For Linux users running OLLAMA as a systemd service, the systemctl command is used to set environment variables:

Use systemctl edit ollama.service to open the service file in an editor. In the [Service] section, add the Environment line with your CORS settings.

[Service]
Environment="OLLAMA_ORIGINS=https://job-matcher.senyas.io"

Save your changes, then reload systemd and restart OLLAMA with

systemctl daemon-reload
systemctl restart ollama

Windows

Ensure Ollama is not running by quitting the application from the taskbar. Open the Control Panel and navigate to “Edit system environment variables.” Choose to edit or create a new variable named OLLAMA_ORIGINS.

Variable name: OLLAMA_ORIGINS
Variable value: https://job-matcher.senyas.io

Press OK and close the control panel. Run Ollama from a new terminal window to ensure it picks up the updated environment variables.