Data Engineer
Clarasight
Clarasight provides global enterprises with a real-time Carbon Planning & Intelligence platform to power forward-looking emissions management, so they can close the gap between climate-related actions and intentions. Some of the world's most recognizable companies, representing over $4T in market value, rely on Clarasight's software to reduce risks, save costs, protect revenues, and manage emissions aligned with business objectives and sustainability goals.
Clarasight is seeking a talented Data Engineer with exceptional data engineering skills who is dedicated to transforming how enterprises manage and optimize travel data through our innovative Carbon Planning and Intelligence platform. Join us in bridging the gap between data-driven insights and actionable solutions for a more sustainable future.
About You
You are passionate about data, APIs, and building seamless data pipelines.
You thrive on ensuring data accuracy, integrity, and reliability.
You enjoy collaborating with customers and internal teams to solve complex data challenges.
You are detail-oriented and meticulous when it comes to data quality and continuous improvement.
You are excited to work in a fast-paced startup environment with a remote, international team of top talent.
Responsibilities
Customer Discovery and Requirements Elicitation:
Collaborate directly with customers through discovery sessions to define technical and business requirements, including identifying necessary data fields for extraction.
Engage directly with customers to align technical solutions with business requirements, ensuring successful data integration and reporting.
API Integrations:
Work on direct API integrations to extract data from client systems, ensuring seamless and accurate data flow into Clarasight's platform.
ETL Pipeline Development:
Design, implement, and maintain scalable ETL processes to extract, transform, and load data from business critical systems including Travel Management, Expense, and HR aContinuously optimize ETL workflows for efficiency, scalability, and reliability, incorporating feedback and evolving business needs.
Quality Control and Database Integrity:
Implement processes to clean and validate data, ensuring high data integrity and consistency across all pipelines.
Work closely with database engineers to develop consistency controls and automated tests, ensuring high-quality connections between data pipelines and the Clarasight client databases
Proactively monitor and resolve data-related issues for customers, ensuring consistent and reliable data availability.
Cross-Functional Collaboration:
Partner with Reporting experts, Product Management, Customer Success, and other internal teams to identify patterns in customer implementations and opportunities to productize custom data pipelines into reusable connectors.
Become an expert in travel and expense data, understanding key industry standards and trends to better serve Clarasight's customers.
-
Share knowledge with the team through documentation, code reviews, and collaborative sessions to foster continuous learning and improvement within the data engineering function.
A Few Things We’re Looking For
5+ years of data architecture, data modeling, data warehousing concepts, methodologies, and best practices
Experience working directly with customers
Experience scaling and re-architecting data platforms and infrastructure through orders of magnitude of growth and scaling data volume
Experience working on data infrastructure-focused engineering teams building a data lake or data warehouse from scratch
Experience with state-of-the-art ETL tools, techniques and processes
RDBMS; SQL; query optimization; performance troubleshooting
3+ years of python programming experience
Familiarity with DBT or SQLMesh
0 to 1 startup experience
Exceptional common sense and the ability to manage ambiguity
Bonus
Experience in any of these areas is a bonus:
Experience working with data in the Travel industry (OBTs and TMCs, expense systems)
Experience with open-source libraries, frameworks, and systems such as DuckDB, Spark, Trino, Iceberg, and Deltalake
Building or using Financial Planning & Analysis (FP&A) systems
Data visualization tools (e.g., Tableau in particular, Power BI)
TypeScript – It’s what our ETL is written in now
Data Science
-
Data Wrangling
No visa sponsorship is available for this position.