Forage AI | Hiring | Data Engineer (ETL Pipelines / Web Crawling) | BigDataKB.com | 2022-09-16

0

Job Location: India

Join Forage AI as a Data Engineer (Python)! 

In this role, you’ll be working with an amazingly passionate and talented team of engineers and data scientists who are working at the bleeding edge of data science and data automation.

Responsibilities:

Our data engineering team is integral to our business operation in growing through data extraction. You will play a key role in coding and designing unique data extraction methodologies. You will also have responsibilities in aiding the business growth through technological developments. You will have end-to-end ownership in working on projects independently. Here’s what you’ll do:

  • Create and maintain an effective data pipeline architecture.
  • Build, improve, and run our generic robots to extract data from the web while handling critical information in a wide variety of structures and formats without error
  • Assemble large and complex data sets based on business requirements.
  • Set up processes (manual and automation) for optimizing data delivery at scale.
  • Independently ensure that delivery cycles are completed end-to-end through effective collaboration with various stakeholders, including our clients, engineers, and QA specialists.
  • Monitor and improve back-end performance.
  • Will be responsible for handling data and problem solving with databases and python

Requirements:

  • Must have 3-4 years of experience working with Python programming and various python libraries
  • Must be deeply familiar either with web crawling OR core backend development –
  • Web crawling experience must include exposure to python packages and frameworks like Requests, Scrapy, Pandas, Urllib, or BeautifulSoup (BS4). Experience with web-based automation tools (Selenium, Puppeteer, or Mechanize, etc.) would be an added advantage.  
  • Core backend development experience must include crafting ETL pipelines that handle large volumes of data, working in Linux environments, etc.
  • Must be familiar with API development, including web frameworks like FLASK, or Django, or FastAPI, etc.  
  • Have experience in working with at least one standard RDBMS, preferably PostgreSQL. Knowledge of NoSql, preferably MongoDB would be an added advantage.
  • Excellent troubleshooting and debugging skills.
  • Strong attention to detail and demonstrated ability in a professional, programming position.
  • Exposure to working with cloud platforms, preferably AWS would be an added advantage.

Other Infrastructure Requirements:

Since this is a completely work-from-home position, you will also require the following –

  • High-speed internet connectivity for video calls and efficient work.
  • Capable business-grade computer (e.g., modern processor, 8 GB+ of RAM, and no other obstacles to interrupted, efficient work).
  • Headphones with clear audio quality.
  • Stable power connection and backups in case of internet/power failure.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

🔍 Explore All Related ITSM Jobs Below! 🚀 ✅ Select your preferred "Job Category" in the Job Category Filter 🎯 🔎 Hit "Search" to find matching jobs 🔥 ➕ Click the "+" icon that appears just before the company name to see the Job Detail & Apply Link 📝💼

LEAVE A REPLY

Please enter your comment!
Please enter your name here