Forage AI | Hiring | Data Engineer (ETL Pipelines / Web Crawling) | BigDataKB.com | 2022-09-16

Job Location: India

Join Forage AI as a Data Engineer (Python)! 

In this role, youโ€™ll be working with an amazingly passionate and talented team of engineers and data scientists who are working at the bleeding edge of data science and data automation.

Responsibilities:

Our data engineering team is integral to our business operation in growing through data extraction. You will play a key role in coding and designing unique data extraction methodologies. You will also have responsibilities in aiding the business growth through technological developments. You will have end-to-end ownership in working on projects independently. Hereโ€™s what youโ€™ll do:

  • Create and maintain an effective data pipeline architecture.
  • Build, improve, and run our generic robots to extract data from the web while handling critical information in a wide variety of structures and formats without error
  • Assemble large and complex data sets based on business requirements.
  • Set up processes (manual and automation) for optimizing data delivery at scale.
  • Independently ensure that delivery cycles are completed end-to-end through effective collaboration with various stakeholders, including our clients, engineers, and QA specialists.
  • Monitor and improve back-end performance.
  • Will be responsible for handling data and problem solving with databases and python

Requirements:

  • Must have 3-4 years of experience working with Python programming and various python libraries
  • Must be deeply familiar either with web crawling OR core backend development –
  • Web crawling experience must include exposure to python packages and frameworks like Requests, Scrapy, Pandas, Urllib, or BeautifulSoup (BS4). Experience with web-based automation tools (Selenium, Puppeteer, or Mechanize, etc.) would be an added advantage.  
  • Core backend development experience must include crafting ETL pipelines that handle large volumes of data, working in Linux environments, etc.
  • Must be familiar with API development, including web frameworks like FLASK, or Django, or FastAPI, etc.  
  • Have experience in working with at least one standard RDBMS, preferably PostgreSQL. Knowledge of NoSql, preferably MongoDB would be an added advantage.
  • Excellent troubleshooting and debugging skills.
  • Strong attention to detail and demonstrated ability in a professional, programming position.
  • Exposure to working with cloud platforms, preferably AWS would be an added advantage.

Other Infrastructure Requirements:

Since this is a completely work-from-home position, you will also require the following –

  • High-speed internet connectivity for video calls and efficient work.
  • Capable business-grade computer (e.g., modern processor, 8 GB+ of RAM, and no other obstacles to interrupted, efficient work).
  • Headphones with clear audio quality.
  • Stable power connection and backups in case of internet/power failure.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

๐Ÿ” Explore All Related ITSM Jobs Below! ๐Ÿš€ โœ… Select your preferred “Job Category” in the Job Category Filter ๐ŸŽฏ ๐Ÿ”Ž Hit “Search” to find matching jobs ๐Ÿ”ฅ โž• Click the “+” icon that appears just before the company name to see the Job Detail & Apply Link ๐Ÿ“๐Ÿ’ผ

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *