Hey there! I'm working to build out a Data Engineering team for nCino and we need your expertise! We are in the very early stages of capturing and cataloging data from various sources. We need people to help us build out a system that will make this data easily accessible to analyst and data scientists. We're going to be building pipelines to ingest data from the Fed, our clients, and other sources of economic data. We'll be using this data to explore ideas like ensuring that banks are lending fairly, how we can optimize deals so that they're better for the bank and their customers, and how we can reduce fraud among a few other ideas we have.
Our stack is based in AWS using their Lake Formation tools as well as other tools like Kinesis, Glue, and Athena. We'll be using some AWS CDK, Spark, Typescript, Python, etc to setup infrastructure for, ingest, clean, and catalog data. The data we're receiving will be in various formats from APIs that we'll need to hook into in Salesforce to CSV and TSV files that the Fed publish and everything in between. Then we'll be publishing this data internally in an easy to query format so that whether it's in a notebook or an application we can get data quickly and easily.
I get really excited talking about this effort! We want to change how banking and our company makes use of all the data out there. If any of this sounds interesting at all I'd love to hear from you! You'd be working with me and another junior level data engineer as well as a team from our AWS platform effort.