Local AI vs ChatGPT Plus for job hunting
An honest comparison for résumé tailoring, cover letters, and application tracking. Where local wins (privacy of career data, no per-prompt cost, full ownership of your tracker), where ChatGPT wins (zero setup, web-aware research), and the hard ethical lines you don't cross either way.
Answer first
For a one-week search where you apply to a handful of roles, ChatGPT Plus is fine. For anything longer than that — a multi-month transition, a sensitive industry switch, a search you don't want anyone outside your home to know about — running a model locally for the résumé and cover-letter work is the better call. The reason is not capability; both can draft a competent first pass. The reason is the data: your career history, salary, employer names, and the dozens of cover letters you draft for jobs you never get all live somewhere when the work is done. Local AI keeps that on your laptop. ChatGPT Plus keeps it on someone else's server, possibly logged for safety review, possibly used for evaluation depending on the consumer-tier toggles you may or may not have set.
The full operator-grade workflow lives at /workflows/private-career-assistant. The five rules of honest AI-assisted application use are at /guides/how-to-use-ai-in-job-applications-ethically. This page is the head-to-head.
What you actually do during a search
A real job search is not one task. It is a stack of recurring small tasks that compound over weeks. Honestly cataloging them clarifies which assistant fits where:
- Tailoring an existing résumé to a specific role. You feed the JD and your master CV; the model surfaces the four bullets most relevant and tightens language. You verify every claim is yours.
- Drafting cover letters. Three-paragraph structure: hook, fit-with-two-examples, close. The model drafts; you rewrite each paragraph in your own voice.
- Summarizing long JDs and company pages. A 1,200-word JD has 200 useful words. The model extracts must-haves vs nice-to-haves vs cultural-fit signals.
- Tracking applications. Companies, dates, contact people, status, follow-up cadences. A small SQLite-backed tracker plus a chat interface beats a Notion board you abandon by week three.
- Practicing behavioral interviews. The model asks the ten most likely questions for the JD; you answer out loud; optionally AnythingLLM indexes your past answers so you build a personal library across the search.
- Researching companies. Recent press, funding, public Glassdoor patterns, who works there from your network. This is the one task where web access matters.
Where local AI wins for job hunting
Three categories where the answer is not even close.
Privacy of career data. Every cover letter you draft is a snapshot of who you are, what you've done, what you make, and where you're trying to go. Multiplied across a multi-month search, that data set is more sensitive than most people's tax shoebox. Cloud-based assistants log it for retention windows ranging from 30 days to indefinitely depending on the tier and the toggles. Local inference produces none of that off-system trace. The model file is on your disk; the prompts go through a runtime on your machine; nothing leaves the laptop unless you decide it does.
No per-prompt cost. A serious search can easily run hundreds of prompts a week — drafts, rewrites, JD summaries, interview practice. With ChatGPT Plus you stay under the cap most weeks. With a local LM Studio + Qwen 2.5 14B setup you stop counting at all. The cost shifts from per-message to electricity, which lands at a few dollars a month even on heavy GPU use.
Full ownership of the tracker. The end-state of a multi-month search should be a personal database: companies, contacts, JDs, your responses, what worked. Cloud assistants rebuild this every conversation; local stacks like AnythingLLM or a small SQLite store keep it forever. When you start your next search two years from now, the data is still on your disk and still searchable.
Where ChatGPT Plus wins
Operator-grade honesty: ChatGPT Plus has real wins, and pretending otherwise wastes your time.
Zero setup, ten-second start. If your search is one or two roles and you don't already own a local stack, the time cost of installing Ollama, picking a model, and learning a frontend is more than the value local would deliver for that search. Plus is a paid login.
Web-aware company research. Recent press, current funding, the team page that updated last week — local models cannot fetch this without you adding a search tool yourself. Plus does it natively. For the research-and-prep portion of the workflow, this is a real edge.
Frontier reasoning on the hardest behavioral questions. If you're prepping for senior leadership rounds at competitive companies, the gap between a frontier reasoning model and a local 14B-32B is most noticeable on the “tell me about a time you had to make a hard call between two values” questions. The local model gets you 80% of the way there; the frontier model nudges the last 20%.
Ethics — the lines you don't cross either way
These rules apply to both local and cloud assistants. The choice of where the model runs does not change the ethics of how you use it.
- No impersonation. Do not pipe an assistant into a hidden earpiece during a live interview. Do not run an “interview copilot” overlay that paraphrases your answers in real time. This is fraud, it is being detected at increasing rates, and the blast radius is “quietly blacklisted at every adjacent company,” not just the one you're interviewing with.
- No live-interview cheating. Live coding rounds, live system-design rounds, live behavioral rounds are yours alone. AI prep before the call is fine. AI during the call is not.
- No fabricated credentials. Do not let any model — local or cloud — add certifications you didn't earn, languages you don't speak, projects you didn't ship, or employers you didn't work for. You will be asked to demonstrate these in week one of the role and you will not be able to.
- No bypassing legitimate ATS filters with deceptive content. Hidden white-on-white keyword stuffing, false metadata, and résumé text designed to game an ATS into surfacing claims you can't back up are deception. ATS optimization is fine when it is honest reformatting; it is not fine when it is trickery.
- Disclose when asked. If an employer asks whether you used AI in preparing your application, the answer is the truth: “I drafted with AI and rewrote and verified every line.” That is a defensible answer. “No” when the answer is “yes” is the failure mode.
These rules are part of our editorial policy. They exist because the failure mode is not abstract — it is “offer rescinded, name on a list.” The full ethical framing is in /guides/how-to-use-ai-in-job-applications-ethically.
The hybrid setup most candidates end at
After a few weeks of running a real search, most candidates settle into a split that uses each tool where it fits. Local handles the bulk-and-private work: tailoring the résumé to every JD, drafting cover letters, organizing the tracker, rehearsing behavioral answers. Cloud handles the bursty research: company news, funding history, executive backgrounds. The cover-letter draft never leaves the laptop. The summary of last quarter's public earnings call comes back from the web.
The hardware floor for the local side is modest — 8 GB+ RAM is enough for a 7-8B model that handles tailoring and drafting, and 12 GB+ VRAM unlocks the 14B class that feels indistinguishable from a paid chat tier for this work. /will-it-run/custom will tell you what your machine can run in under a minute. The full stack assembly is in /workflows/private-career-assistant; the toolset for résumé-specific tasks is in /guides/local-ai-tools-for-resume-optimization; common breakages are in /guides/how-to-troubleshoot-local-ai-job-tools.
Next recommended step
Full stack: model + retrieval + tracker, with hardware tiers and failure modes.
The break-even math is straightforward: at $240 per year for ChatGPT Plus, a budget GPU that runs the same class of models pays for itself within twelve to eighteen months while keeping your search history, resume drafts, and salary negotiation data entirely off OpenAI's servers. The privacy dividend alone — never having your job-search patterns absorbed into a public training corpus — makes the hardware route compelling before you even factor in the subscription savings.
The hardware that replaces the subscription for good: best budget GPU for local AI.