Establishing cannabis as a force for good.
PAX Labs is an award-winning consumer technology brand with a mission to establish cannabis as a force for good. Our products set a high bar for quality and, together with the PAX App, offer unique features such as safety locking, adjustable temperature and draw control. As millions of consumers try cannabis for the first time or return to it after a long hiatus, PAX believes that quality, transparency and predictability are the keys to safe, responsible use by adults.
Headquartered in San Francisco, PAX is backed by leading technology investors including Tiger Global and Tao Invest (Pritzker family fund). We are an exceptional team with hardware and software backgrounds in consumer technology, healthcare, and biotech, and are growing to deliver on our mission. Our internal culture values diversity, integrity, having an impact, and passion for the larger movement that PAX is a part of.
ROLE AND RESPONSIBILITIES:
Our team is looking for a senior engineer to help craft, build and maintain an efficient data pipeline architecture. You will work closely with internal partners on identifying data-related technical issues and supporting infrastructure needs. We are looking for an integral member of the Software team that will help define the Data Engineering function at PAX.
As the PAX Senior Software Engineer, you will:
Assemble large, sophisticated data sets that meet functional / non-functional business requirements;
Identify, design, and implement internal processes and process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc;
Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Apache Airflow, AWS Athena, AWS EMR, Spark, and other AWS big data tools;
Build analytics tools that utilize the data pipelines to provide meaningful insights into customer acquisition, operational efficiency and other key business performance metrics;
Work with partners including the Executive, Software, Product, Data and Design teams to assist with data-related issues and support their data infrastructure needs;
Keep our data separated and secured within logical boundaries via strict role based access controls;
Maintain data tools like Apache Superset, and Zeppelin for analytics and data scientist team members that assist them in building and optimizing our product for future innovations;
Work with data scientists and analytics specialists to strive for greater functionality in our data systems
7+ years experience working as a Data Engineer (or equivalent);
Comfort working in an agile and/or lean development environment;
Experience with modern data pipelines, data streaming, and real time analytics using systems such as Apache Airflow, AWS Kinesis, AWS EMR, Spark, AWS Lambda, AWS Athena (Presto), Zeppelin, Jupyter Notebook, or similar technologies;
Strong programming abilities with Python, Scala, Java or other similar languages;
A strong ability to understand and organize data from disparate sources into structures that are easy to digest and query;
Sophisticated working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases and their tradeoffs;
Experience building and optimizing big data ETL pipelines that are idempotent, incremental and partitioned for performance and cost efficiency;
Experience performing root cause analysis on internal and external data and processes to answer specific business questions and find opportunities for improvement;
A successful history of manipulating, processing and extracting value from large disconnected datasets;
Solid understanding of message queuing, stream processing, and highly scalable ‘big data’ data stores;
Strong project management and organizational skills
Write clear, concise and maintainable code;
Must be a strong, self-motivated individual that can work with minimal day to day supervision and able to objectively rank and prioritize development work;
Excellent English verbal and written communication skills
Experience with the following software/tools:
Big data systems: Hadoop, Spark, Sqoop;
Relational SQL and NoSQL databases, including MySQL and Presto;
Data pipeline and workflow management tools: Apache Airflow;
AWS cloud services: EC2, EMR, Kinesis, Lambda, Redshift, Athena, Glue
Stream-processing systems: Storm, Spark-Streaming, etc;
Object-oriented/object function scripting languages: Python, Java, Scala;
Experience with bash scripting, Terraform, machine learning tools and concepts is a plus
Graduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field preferred or equivalent experience
PAX LABS PERKS & BENEFITS:
Competitive compensation, equity & bi-annual performance reviews
Fully funded comprehensive medical, dental, and vision coverage
Generous PTO policy
Paid Parental Leave
Monthly wellness reimbursement
Cell Phone reimbursement
Employee Purchase Program for discounted PAX devices
Weekly catered lunch, endless snacks and beverages
Dog Friendly HQ in the Mission District of San Francisco
Employee Assistance Program including access to online legal support
We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.
Recruiter screen, hiring manager/tech screen, virtual panel interview, offer.
Made with ❤️ by Veni Kunche.