Senior Data Engineer
Santa Ana, California
Company SummaryCome join First American's Digital Title Group, newly formed to re-imagine and digitize the title search and examination process through Big Data, AI, document automation and modern, cloud-native application development. As a market leading title insurance company, powered by the nation's largest and most complete property information, ownership and recorded document database, First American is committed to advancing title automation and removing friction from the real estate closing process. Our modern title decisioning solutions create certainty and speed through data and analytics, delivered to real estate agents, lenders, title agents and homebuyers. Join a team that puts its People First! Since 1889, First American (NYSE: FAF) has held an unwavering belief in its people. They are passionate about what they do, and we are equally passionate about fostering an environment where all feel welcome, supported, and empowered to be innovative and reach their full potential. Our inclusive, people-first culture has earned our company numerous accolades, including being named to the Fortune 100 Best Companies to Work For® list for seven consecutive years. We have also earned awards as a best place to work for women, diversity and LGBTQ+ employees, and have been included on more than 50 regional best places to work lists. First American will always strive to be a great place to work, for all. For more information, please visit www.careers.firstam.com.
Senior Data Engineer (remote)
We're looking for Senior Data Engineers interested in transforming our industry by solving cutting-edge problems with modern technologies, want the benefit of working for an established real estate insurance leader and seek a culture awarded as a Fortune 100 Best Companies to Work.
What you’ll do:
Develop ETL pipelines involving transformation of nested data stored in JSON and Parquet files, using GLUE + Pyspark
Develop and maintain scalable data pipelines and build out new API integrations for data transfer.
Provide guidance on data architecture for tables, indices, caching strategies, etc. to help improve performance
Collaborate with analysts to perform data analysis and troubleshoot data issues.
3+ years of development experience with any of the following software languages: Python, R, Scala, Java, Kotlin, or C# (we use Python)
2+ years of experience in any of these technologies: Snowflake, BigQuery, Databricks, Spark in any flavor, HIVE, Hadoop, Cloudera or RedShift (we use PySpark & Snowflake)
Strong Database/SQL experience in any RDBMS (We use PostgreSQL)
Experience in schema design, data ingestion/processing experience on Snowflake (or equivalent MPP)
Experience developing in a containerized local environment like: Docker, Rancher, Kubernetes
Experience in orchestrating data processing jobs using Step Function/Glue workflow/Apache Airflow (MWAA)
Bachelor's degree in Computer Science (or related field) or equivalent combination of education and experience