Data Engineer - Backend/Database/Devops

Bountiful

Date listed

2 months ago

Employment Type

Contract

Remote

Yes

Company Overview: Bountiful is a data analytics company that is digitizing and organizing data in the agriculture specialty crop industry. We digitize records, benchmark against peers, forecast yield, and give market intelligence to farmers, buyers and sellers. We recently launched our Operating System (OS) product to digitize and synthesize farm level data into one place (dashboard) for farmers and track production, costs and income year over year. This data feeds into training our yield forecasts to provide more accurate estimates at the end of the season at the field level. We have seen tremendous growth in this new product with farmers growing 3x last year and on track for over 3x this year. OS has opportunities for applications beyond a dashboard for farmers such as creating a data moat for ML models, due diligence for lending, insurance reporting/verification and trading. 

Job description: Contract, Data Engineer - the primary role is to improve our Operating System (OS) product backend and database infrastructure so operations can be more automated and robust. Python is the primary scripting language that is used to digitize various pdf records, write to a PostgreSQL database and be served to the user in table/graphical form in Retool via React. Several upgrades are needed throughout each step in the process. Opportunities for follow on work upon completion of these tasks.

Technology Stack: Python/Node (Backend), React/JS (Frontend), AWS services (RDS, EC2, S3) transitioning to GCS (SQL, Compute, Storage), PostgreSQL (Database), Docker, Retool, Mixpanel, Team: 1 engineer

Responsibilities/Tasks

  • Refactor python ingest processes for digitizing three major domain specific pdf document formats (Fieldruns, Turnouts, USDAs). Outcome: improved accuracy, quality, and automation of the ingest process  using either OpenAI’s gpt generative models or Google’s gemini generative models. - 6 weeks
  • Complete end-to-end automation process of pushing to the database once a user (near real time) uploads a pdf document. This means - 6 weeks
    • kicking off the ingest scripts in the user import page and handling co-current jobs to several tables in the database
    • Set up a notification system for users when new data is available within OS via email
    • User flagging and editing of displayed data via postgres queries in Retool. Updating edited data to corresponding tables to the database. 
  • Improve best practices for database and development - 3 weeks
    • Updating database performance, security and logging tools. General upgrades. 
    • Set up proper workflow for dev, staging and production databases.
  • This role will work directly with the Head of Product to understand current capabilities, expected outcomes of tasks, and perform code & product reviews to ensure we maintain product quality.

Qualifications: 

  • 3-5 years of full stack experience in professional setting
  • Small team experience < 10 people
  • Curious and problem solver
  • Great communicator
  • Expert in Python
  • Experience with AI model integration and a strong interest in AI
  • Specific focus on database and backend engineering
  • Expert with PostgreSQL databases
  • Built out CI/CD pipelines
  • Proficient with AWS (S3, RDS, EC2)  or GCS (Storage, SQL, Compute) cloud products
  • Familiar with Retool or dashboard like tools that leverage API and database connections

Findwork Copyright © 2023

Newsletter


Let's simplify your job search. Receive your tailored set of opportunities today.

Subscribe to our Jobs