Sr. Data Engineer Job Description Template
Our company is looking for a Sr. Data Engineer to join our team.
Responsibilities:
- Achieve proficiency with Big Data SQL query engines such as Drill, Spark Sql and Hive;
- Assist the Sales Team (Sales Rep and Sales Engineer) in positioning and selling MapR Service products and Service offerings;
- Build and maintain one or more data lakes to support scalable ingesting, manipulation, and reporting of data;
- Manipulate data to produce and maintain new data elements using repeatable, automated processes;
- Demonstrates knowledge of industry trends, our infrastructure, technologies, tools and systems;
- Experience in measuring and communicating the value of data platforms and tools;
- Partner end-to-end with Product Managers and Data Scientists to understand customer requirements and design prototypes and bring ideas to production;
- Formulate and refine analytics algorithms leveraging data from search, social and e-commerce in time series, text and image-based formats;
- Expertise across data related disciplines and environments;
- Develop accurate time, cost and resource estimates for developing and maintaining systems;
- Keeps abreast of new and emerging business systems design and development techniques and makes appropriate recommendations for their use;
- 5+ years IT experience;
- Partner with the security and risk teams to build appropriate security and compliance into the data platforms;
- Demonstrated experience building and maintaining strong relationships with business and IT senior leaders and executives;
- Experience in multiple enterprise environments, industries and development methodologies.
Requirements:
- 2+ years in a customer facing, professional services software delivery role preferred but not required;
- Certifications on upcoming NoSQL/Hadoop oriented databases like MongoDB, Cassandra are preferred but not needed for non-relational databases;
- Strong experience with advanced analytics tools for Object-oriented/object function scripting using languages such as R, Python, Java, and C++;
- Metadata;
- Bachelor’s Degree in Computer Science, Computer Engineering or a closely related field;
- Strong experience with performance and scalability design and testing;
- DD254;
- Strong experience with data processing software in Azure, AWS and GCP;
- Handle large volumes of data and integrate our platform with a range of internal and external systems;
- Knowledge of Application development platforms/languages;
- Experience connecting business requirements to data mining objectives and measuring the business benefit;
- Responsible for building and maintaining processes for ingestion of data to data lake;
- Current Stack: Python, AWS Redshift, AWS S3, CloudWatch, EMR Spark, Airflow, Okta;
- Ability to execute the strategic vision with tactical efficiency;
- Experience with Teradata & Hadoop.