Senior Engineer, Data Products & AI Innovation
About Us
Prokeep is an Order Engine on a mission to empower distributors to win by turning every customer interaction into an order opportunity.
Trusted by 8,000+ distributor locations and 40,000+ users, Prokeep makes it easy for over 3 million contractors, technicians, plumbers, electricians, mechanics, and other skilled professionals to order parts, troubleshoot issues, and generally, get what they need and get on with their day.
Built for the $100+ billion distribution market, Prokeep eliminates missed opportunities caused by busy phone lines, scattered customer data, and passive sales habits. By centralizing communication and enabling proactive outreach, we help teams take more orders from any channel and get more orders by driving demand—fueling growth, stronger relationships, and greater impact across the supply chain.
Having recently secured a $25M Series A investment, we’re poised for rapid growth and innovation. If you're passionate about tackling complex challenges in a collaborative, fast-paced environment while contributing to an often-overlooked yet vital sector of the economy, we’d love to have you on board.
Let’s shape the future together—one order at a time!
Our Values
Build Together. – Collaboration, camaraderie, and customer-focus are core to how we collectively work and win.
Own It. – Persistence and personal growth drive the results - and the rewards! - that we can all enjoy.
Be Humble. Be Human. – Respect and authenticity allow us to build meaningful and lasting relationships.
Have Fun! – Positivity, perseverance, and a bit of laughter all allow us to make an impact while enjoying the rollercoaster!
The RoleWe're building a small, high-impact team to create new data-driven capabilities and AI-powered features that enhance our existing SaaS products or serve as standalone offerings. We’re looking for a Data Engineer who’s excited to be part of an entrepreneurial, product-minded team where creativity, curiosity, and experimentation are core to how we work.
This isn’t a role for someone who wants to be handed requirements and optimize pipelines in isolation. We're looking for someone who thrives in ambiguous environments, loves turning messy data into structured assets, and has a nose for product opportunities hiding in the data.
You’ll be joining a team led by a deeply technical founder, alongside a founding engineer focused on AI and machine learning. Your role will focus on designing and building robust, scalable data pipelines, shaping our storage layer, and collaborating on new data products and proofs of concept.
What You’ll DoDesign, build, and maintain ETL/ELT pipelines using modern data tools and frameworks
Own and optimize our data lake/data warehouse setup (e.g. Parquet, Iceberg, Delta Lake)
Collaborate closely with AI/ML engineers to build and serve training datasets, inference pipelines, and feedback loops
Partner with product and engineering to turn raw data into product-ready insights, APIs, or feature sets
Prototype and iterate on new data product ideas quickly, focusing on speed-to-learning over polish
Take initiative in identifying data-driven opportunities within our platform or market
Contribute to defining best practices, architectural choices, and internal data standards
Strong engineering background with 5-7 years of hands-on experience in data systems and distributed processing (e.g., Apache Spark, dbt, Airflow, etc.)
Deep understanding of file formats and table formats like Parquet, Apache Iceberg, or Delta Lake
Experience building and operating data infrastructure in AWS or similar
Proven ability to work independently and ship data products, features, or APIs—not just internal dashboards
Familiarity with data modeling for analytics and product use cases
Curiosity and creativity in exploring what’s possible with data + AI
A founder’s mindset — someone who wants to shape not just the code, but the product and direction of the team
Comfort navigating ambiguity, iterating quickly, and learning through experimentation
Experience with real-time/streaming data tools (e.g., Kafka, Spark Streaming, Flink)
Experience with MLOps practices like deploying and monitoring models, managing feature stores, or automating retraining workflows
Familiarity with building LLM-powered features, including pipelines, embeddings, vector databases, or semantic search
Exposure to tools like Trino, DuckDB, or OLAP engines used for ad-hoc querying over data lakes
You’ll be part of a small, senior team with the autonomy to move fast and explore big ideas
You’ll have direct access to a production-scale SaaS platform, customer usage data, and real market needs
You’ll help shape entirely new data products, not just pipelines behind the scenes
You’ll work with a team that values craft, experimentation, and impact over red tape
Why Prokeep?
At Prokeep, we offer a dynamic, passionate, and collaborative work environment where innovation thrives. Here’s what you can expect:
Competitive Compensation: ($125K-135K) Reflecting your expertise and impact.
Equity Package: Your success is our success, share in the growth you’ll help create.
Comprehensive Benefits: Health, dental, vision, life, short & long-term disability, 401(k), and employee assistance program (EAP).
Flexible PTO: Recharge and refocus with the flexibility to manage your time with no preset limits
Hybrid Work: Hybrid from one of our office in Portland
Continuous Growth: Yearly education stipend to support your professional development.
Apply to this Job