-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feature/workday_scraper #67
base: master
Are you sure you want to change the base?
feature/workday_scraper #67
Conversation
Saw your comment on the #23 ticket. Is that an error inside docker? Is your docker instance not able to install chrome? Are you able to run it using python + venv? |
Just looked over the PR, I like it. Please include a test for the scraper. It can be something very simple and straightforward that tests if the scraper is properly retrieving info from the site. Thank you |
@nicobrenner Hey! I am able to run it locally through venv, and I also managed to build with docker just now but running it is causing some issues atm. I pushed a commit with a successful docker build |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you @noameron for this PR. Excellent work. One last request: please update the README.md file to include an update at the top. It should be 1 or 2 sentences about what is new. Extra brownie points if you also make a short video using it 🙏🏽🤗:
@@ -10,6 +10,11 @@ WORKDIR /commandjobs | |||
# Install any needed packages specified in requirements.txt | |||
RUN pip3 install --no-cache-dir -r config/requirements.txt | |||
|
|||
# Install required packages, including Chromium and ChromeDriver |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Excellent :)
@@ -14,7 +16,7 @@ | |||
class WorkdayScraper: | |||
def __init__(self, db_path='job_listings.db', update_func=None, done_event=None, result_queue=None): | |||
self.db_path = db_path | |||
self.driver = webdriver.Chrome(options=self.get_selenium_configs()) | |||
self.driver = webdriver.Chrome(service=Service(ChromeDriverManager().install()), options=self.get_selenium_configs()) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Very helpful 👍🏼
"🕸 Scrape \"Work at a Startup jobs\"", | ||
"🕸 Scrape \"Workday\"", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
😄
Added support for generic Workday scraper, and updated requirements.
I am facing issues trying to build docker, due to chrome / chromeDriver.