Job description
General Responsibilities:
- Development of high-quality code for the core data stack including data integration hub, data warehouse and data pipelines.
- Build data flows for data acquisition, aggregation, and modeling, using both batch and streaming paradigms
- Empower data scientists and data analysts to be as self-sufficient as possible by building core systems and developing reusable library code
- Support and optimize data tools and associated cloud environments for consumption by downstream systems, data analysts and data scientists
- Ensure code, configuration and other technology artifacts are delivered within agreed time schedules and any potential delays are escalated in advance
- Collaborate across developers as part of a SCRUM team ensuring collective team productivity
- Participate in peer reviews and QA processes to drive higher quality
- Ensure that 100% of code is well documented and maintained in source code repository.
INNOVATIVE INFRASTRUCTURE & PROBLEM SOLVING
- Strive to engineering excellence by simplifying, optimizing, and automating processes and workflows.
TEAMWORK & COMMUNICATION
- Proactively educate others on basic data management concepts such as data governance, master data management, data warehousing, big data, reporting, data quality, and database performance.
- Superior & demonstrated team building & development skills to harness powerful teams
- Ability to communicate effectively with different levels of leadership within the organization
- Provide timely updates so that progress against each individual incident can be updated as required
- Write and review high quality technical documentation
CONTROL & AUDIT
- Ensures their workstation and all processes and procedures, follow organization standards
CONTINUOUS IMPROVEMENT
- Encourages and maintains a ‘Best practice sharing culture’, always striving to find ways to improve service and change mind set.
- Extensive professional experience as a data engineer, software engineer, data analyst, data scientist or related role
- Strongly prefer hands on experience with DataBricks or Palantir
- Experience with relational and dimensional database modelling (Relational, Kimball, or Data Vault)
- Proven experience with all aspects of the Data Pipeline (Data Sourcing, Transformations, Data Quality, Etc…)
- Bachelors or Masters in Computer Science, Information Systems, or an engineering field or equivalent work experience
- Prefer experience with event driven architectures and data streaming pub/sub technologies such as IBM MQ, Kafka, or Amazon Kinesis.
- Strong capabilities in a scripting language such as Python, R, Scala, etc.
- Strong capabilities in SQL and experience with stored procedures
- Strong interpersonal and communication skills with Agile/Scrum experience.
- Strong problem solving and critical thinking skills with a proven record for identifying and diagnosing problems, and solving complex problems with simple, logical solutions.
- Outstanding verbal, written, presentation, facilitation, and interaction skills, including ability to effectively communicate technical and non-technical issues and concepts to multiple organization levels.
- Prefer Travel, transportation, or hospitality experience
- Prefer experience with designing application data models for mobile or web applications
- Excellent written and verbal communication skills.
- Flexibility in scheduling which may include nights, weekends, and holidays preferred
seankuhnke.com is the go-to platform for job seekers looking for the best job postings from around the web. With a focus on quality, the platform guarantees that all job postings are from reliable sources and are up-to-date. It also offers a variety of tools to help users find the perfect job for them, such as searching by location and filtering by industry. Furthermore, seankuhnke.com provides helpful resources like resume tips and career advice to give job seekers an edge in their search. With its commitment to quality and user-friendliness, seankuhnke.com is the ideal place to find your next job.