Taking the Sting Out of the Job Hunt
- clone the repo
- navigate to the project root
- create your venv
python -m venv venv
- activate and enter your venv
source venv/bin/activate
- install dependencies
pip install -r requirements.txt
- create your .env file
SECRET_KEY=[YOUR SECRET KEY HERE] or ask someone on the team for the secret key
DATABASE_URL=postgresql:///job_crawler
TEST_DATABASE_URL=postgresql:///job_crawler
DATABASE_NAME=job_crawler
OPEN_AI_API_KEY=[ROBOT API KEY]
- Create your local job_crawler database:
If you already have a job_crawler DB and want to use the backup data:
-
drop table, create table, import backup_database.sql:
dropdb job_crawlercreatedb job_crawlerpsql -d job_crawler -f data/backup_database.sql
or start from scratch
-
create the tables and run the scraper (this will take a while):
createdb job_crawlerpsql -d job_crawler -f data/migrate.sqlpython full_scrape.py
-
Who's it for? - Bootcamp grads & junior devs What are we looking for? - Job descriptions that fit our needs (bootcamp grads, no degree) Where are we looking for it? - (for now) https://stillhiring.today/
-
Scrape for the URLs
- Get company "career" URLs
-
Scrape those job URLs (company websites) for jobs in our field
- top 5 job boards
- jobs.lever.co: 191
- boards.greenhouse.io: 113
- jobs.ashbyhq.com: 37
- jobs.jobvite.com: 8
- careers.smartrecruiters.com: 7
- top 5 job boards
-
Scrape and save the job descriptions
-
Run the job descriptions through GPT
Please make sure to run the following commands before you start working on the codebase to ensure you have the most up-to-date packages and code:
git pull
pip install -r requirements.txt