Principal Data Engineer Resume Samples

4.9 (83 votes) for Principal Data Engineer Resume Samples

The Guide To Resume Tailoring

Guide the recruiter to the conclusion that you are the best candidate for the principal data engineer job. It’s actually very simple. Tailor your resume by picking relevant responsibilities from the examples below and then add your accomplishments. This way, you can position yourself in the best way to get hired.

Craft your perfect resume by picking job responsibilities written by professional recruiters

Pick from the thousands of curated job responsibilities used by the leading companies

Tailor your resume & cover letter with wording that best fits for each job you apply

Resume Builder

Create a Resume in Minutes with Professional Resume Templates

Resume Builder
CHOOSE THE BEST TEMPLATE - Choose from 15 Leading Templates. No need to think about design details.
USE PRE-WRITTEN BULLET POINTS - Select from thousands of pre-written bullet points.
SAVE YOUR DOCUMENTS IN PDF FILES - Instantly download in PDF format or share a custom link.

Resume Builder

Create a Resume in Minutes with Professional Resume Templates

Create a Resume in Minutes
LS
L Schuster
Lina
Schuster
741 Saige Manors
New York
NY
+1 (555) 425 3523
741 Saige Manors
New York
NY
Phone
p +1 (555) 425 3523
Experience Experience
New York, NY
Principal Data Engineer
New York, NY
Gerhold-Farrell
New York, NY
Principal Data Engineer
  • Design data integration architecture in a diverse data environment and develop it from scratch working with a global team with a strong execution rigor
  • Develop, enhance, and automate processes for queuing and prioritizing data management and curation requests
  • As we expand into other platforms later this year, design, code and implement data movement across diverse technology platforms ie Oracle databases, enterprise big data cluster, data science servers and cloud platforms like Heroku
  • This is a unique opportunity to join an innovative group of engineers creating the next generation big data platform that will be accessed and used by hundreds of people across domains within Mercedes-Benz
  • Create ETLs to take data from various operational systems and create a unified dimensional or star schema data model for analytics and reporting
  • Closely work with Data Scientists and ML engineers to build the next generation data integration capabilities which in turn will support a variety of predictive applications
  • Design and implement big data analytics models and solutions, as well as integrate new data management technologies and software engineering tools into existing structures
Los Angeles, CA
Principal Data Engineer
Los Angeles, CA
Gutmann-Williamson
Los Angeles, CA
Principal Data Engineer
  • The individual will be responsible for design and development of high-performance data architectures which support data warehousing, real-time ETL, and batch big-data – processing
  • Actively contribute to Celgene’s R/ED Data Assets initiative to build Celgene-specific knowledge collections for cell lines, animal models and drug compounds based on internal and external clinical, genomic, proteomic, transcriptional and epigenetic assay data
  • The individual will work as part of cross-functional teams in Global Marketing, Research & Development, Information Technology, Quality, Operations, Regulatory and Business Development to deliver informatics products
  • Ensure accurate, complete and timely collection, delivery and tracking of analytical information from internal or contract laboratory service providers or collaborating laboratories for curation, ingestion and delivery to computational and translational scientists
  • Support Technical Program Manager, Research Scientist, and a growing virtual team aimed at analyzing usage data to derive new insights and fuel customer success
  • Support the development of performance dashboards that encompass key metrics to be reviewed with senior leadership and sales management
  • Take responsibility for supporting your changes all the way through to proving that they work well – and continue to work well – in production
present
Phoenix, AZ
Think Big Principal Data Engineer
Phoenix, AZ
Abernathy and Sons
present
Phoenix, AZ
Think Big Principal Data Engineer
present
  • Supporting and managing internal asset development projects
  • Mentoring junior members on the team in application architecture, design and development best practices
  • Other duties as assigned
  • With the number of connected Mercedes-Benz vehicles on the road increasing daily, the amount of data being generated and its significance is also increasing rapidly. The Vehicle Infrastructure & Data Analytics team is looking for passionate and versatile Data Engineers to support with the collection, storage, processing, and analysis of large amounts of vehicle data
  • You will be contributing to technology strategy for the tracking product line and take ownership of technical product architecture of analytics and machine learning functionality
  • As the Principal Data Engineer for HERE IoT Tracking products you have the opportunity to develop analytics tooling and pipelines to process location data from millions of devices
  • Building next-generation Big Data analytics framework developed on a group of core
Education Education
Bachelor’s Degree in Computer Science
Bachelor’s Degree in Computer Science
North Carolina State University
Bachelor’s Degree in Computer Science
Skills Skills
  • Experience in distributed and highly optimized data stores highly desirable
  • Excellent communication skills and the ability to work effectively with others in a team environment
  • Strong analytical and problem solving skills, good understanding of different data structures, algorithms and their usage in solving business problems
  • Strong knowledge on batch and streaming data sources with structured and unstructured data
  • Excellent communication skills, and the ability to work effectively with others
  • A strong voice for data integrity and reporting quality utilizing best practices and industry standards
  • Strong knowledge on Data Warehouses, RDBMS and MPP database skills, including query optimization, and performance tuning
  • Able to take individual ownership of very large complex projects from start to finish. Insatiable curiosity, high aptitude with a flair for business impact
  • Experience with managing database systems on the cloud is highly desirable
  • Experience with ETL frameworks like Airflow is highly desirable
Create a Resume in Minutes

15 Principal Data Engineer resume templates

1

Principal Data Engineer Resume Examples & Samples

  • Leads and supports efforts to provide timely, relevant, and clean data to various groups within Disney inside and outside of Disney Studios
  • Identify, propose, develop, validate, and deploy innovative data insight systems to business partners within Disney Studios
  • Provide business insights by integrating data from disparate systems and services in maintainable ways and developing validated solutions to provide those insights
  • Provide support to other engineers in the form of solution identification, mentoring, automation and data or effort validation
  • Identify opportunities for data sharing and develop federated systems to supply that data
  • 8+ years of software engineering 5+ years of large scale systems
  • A strong knowledge of the Java programming language
  • Working experience with Hadoop batch processing system and the map-reduce horizontally scalable paradigm Experience with the Linux operating system
  • Strong grounding in object oriented programming, aspect oriented programming, design patterns, concurrency, algorithms & data structures
  • Development experience using service oriented architecture, JAX-RS and JAXB Strong ability to research solutions, processes, industry trends and best practices
  • Experience with alternative data processing platforms including Storm, Spark, Shark, Apache Mesos, Hive, PIG, and Apache Crunch
  • Experience with an analytical approach to tuning models
  • Experience with machine learning and machine learning libraries like Apache Mahout
  • Experience with deployment environments and systems including Amazon Web Services & Chef
  • Experience with diverse storage systems, platforms and methods including hBase, MongoDB, Apache Avro, PostgreSQL, Greenplum, Teradata
  • Experience with data integration including Apache Camel, Spring Batch and Talend
  • Experience with pilot application frameworks including Spring Roo, GWT 2.0, Portlets and Spring MVC
  • Experience with disciplined development practices including using tools like Maven, TestNG, Findbugs, CheckStyle, Checker, Sonar, wikis, coding standards and Git
  • Experience working in an evolved technical organization including contributing to and following coding guidelines, best practices, documenting, and presenting at brown bags, professional groups and conferences
  • Experienced in participating in scrum or other agile development environments
2

Principal Data Engineer Resume Examples & Samples

  • Strong Java programming skills required. Other scripting skills, such as Python or Perl a plus
  • Strong knowledge on Data Warehouses, RDBMS and MPP database skills, including query optimization, and performance tuning
  • Expert knowledge on large-scale/distributed SQL, Hadoop, NoSQL, HBase, Columnar databases
  • Expert knowledge on Hadoop-related technologies, Hive, Impala, MapReduce, Spark, etc
  • Strong knowledge on batch and streaming data sources with structured and unstructured data
  • Experience delivering self-service analytics solutions to a wide variety of users to enable data discovery
  • Familiarity with analytical tools such as R, SAS, other statistical packages, data visualization, Business Intelligence, Reporting and other advanced analytics tools and how they access data from different data stores
  • Strong process and technical documentation skills
  • Strong communications skills with different levels in the organization as well as a business acumen to understand various types of business problems
  • Experience working with large datasets, Big Data and distributed database environments
  • Experience working with groups which have highly analytical skills
  • Desire to learn new technologies and try implementing them
  • Ability to drive a project from the beginning to the end
  • Java programming, Object Oriented design, large-scale software application design
  • Software development on Unix/Linux, Unix shell scripting
  • Hadoop experience, Java Map/Reduce
  • Must be self-motivated and capable of working with minimal supervision
  • JavaScript, D3, HTML, CSS, web application development
  • Web Services development, SOAP/REST
  • Spring, Spring Boot, Spring Data, JPA/Hibernate, JDBC
3

Principal Data Engineer Resume Examples & Samples

  • This is a technical role with the candidate expected to be a critical contributor to building compelling informatics products within Abbott’s Diagnostics Division
  • The individual will develop software on informatics programs from concept through launch. She/he will also be responsible for supporting on-market products post-launch
  • The individual will be responsible for design and development of high-performance data architectures which support data warehousing, real-time ETL, and batch big-data – processing
  • She/he will be responsible for leading the execution of product features as part of a cross functional team
  • The individual is responsible for executing the various software development lifecycle (SDLC) activities – user stories, use cases, software requirements, design, development and design verification (incl. unit, functional and system testing)
  • The individual is responsible for root cause analysis including troubleshooting of informatics work products during development and for on-market products
  • She/he will be responsible for creating and maintaining documentation for the artifacts created during the SDLC
  • The individual will be responsible for ensuring the continuous improvement and evolution of software work products
  • The individual should independently conceive plans and implement development artifacts (requirements, analysis, design, documentation and software work products) as-needed to meet program goals
  • The individual may travel occasionally per needs of the assigned project
  • The individual is responsible for the design, development and commercialization of informatics products in compliance with internal policies and procedures, both US and Foreign regulatory standards and requirements
  • She/he is accountable for leading the design and implementation of one or more features as assigned on individual programs
  • The individual will lead development efforts using Agile development frameworks such as Scrum or Kanban
  • The individual will interact with project managers, product owners and R&D leadership on project scope, timeline, risks and budget
4

Think Big Principal Data Engineer Resume Examples & Samples

  • 5+ years of experience leading technical teams
  • 7+ years of experience programming in Java
  • Unix experience
  • Prior experience with Hadoop, HBase, Hive, Pig, and Map/Reduce
  • Experience with other programming languages (Python, C, C++, etc.)
  • Proficiency in SQL, NoSQL, relational database design and methods
  • Experience with Hadoop, Hive, Pig
  • Experience with Avro, Thrift, Protobufs
  • JMS: ActiveMQ, RabbitMQ, JBoss, etc
  • Dynamic and/or funtional languages (Python, Ruby, Scala, Clojure, etc.)
  • Familiarity with different development methodologies (e.g. agile, waterfall, XP, scrum, etc.)
5

Think Big Principal Data Engineer Resume Examples & Samples

  • Prior consulting experience required
  • Prior experience managing client relationships
  • 5+ years of experience managing software teams
  • 7+ years of experience programing in Java or C/C++
  • Prior experience with Hadoop, HBase, Hive, Pig and Map/Reduce
  • Must be adept at sizing and estimating new projects, as well as have experience supporting new business development
  • Adept at learning and applying new technologies
  • Experience with JMS: ActiveMQ, RabbitMQ, JBoss, etc
  • Prior experience with data warehousing and business intelligence systems
  • Familiarity with different development methodologies (e.g., agile, waterfall, XP, scrum, etc.)
  • Firm understanding of Python memory model, classes, subclassing, designing classes for re-use, static string constants rather than in-line constants
  • Experience with Elasticsearch, SolrWeb, and Lucene
  • Experience with Star Schema, fact vs dimensions, updates/restatements and views
  • Able to create storm topology to filter or transform a stream of data. Ability to track state and isolation in Trident or similar
  • Implement a graph (line or pie etc.) baked by a live (changing) data set, something like “request per minute” or similar
  • Understand basic modeling techniques and tool sets. Implement simple Python or R analytic routines
6

Principal Data Engineer Resume Examples & Samples

  • Design, implement, and support data warehouse infrastructure using AWS cloud technologies
  • Create ETLs to take data from various operational systems and create a unified dimensional or star schema data model for analytics and reporting
  • Use business intelligence and visualization software (e.g., OBIEE, Tableau Server, etc.) to empower non-technical, internal customers to drive their own analytics and reporting
  • Monitor and maintain database security and database software
  • Bachelors degree in CS or related technical field and 6+ years of experience in data warehousing
  • Excellent knowledge of SQL and Unix/Linux
  • 8+ years of relevant experience with ETL, data modeling, and business intelligence architectures
  • Deep understanding of advanced data warehousing concepts and track record of applying these concepts on the job
  • Experience building self-service reporting solutions using business intelligence software (e.g., OBIEE, Tableau Server, etc.)
  • Experience with Amazon Redshift or other distributed computing technology
7

Principal Data Engineer Resume Examples & Samples

  • Architect and build applications using the Hadoop Stack
  • Design and build data pipelines solutions to process large amounts of data using the latest Big Data technologies
  • Provide a common interface into the data platform to leverage structured and unstructured data integrated from multiple sources
  • Work with the product managers to understand the business requirements and translate them into data requirements and data models (logical and physical)
  • Work closely with stakeholders (Data Scientists, QA, Business) and the platform team
  • Work closely with Systems Operations on the deployment architecture
  • Mentor junior members of the data team
  • Evangelize best practices in Big Data stack
  • 15+ years of experience in Data Architecture
  • Experience in engineering large-scale systems in a product environment
  • In-depth understanding of the inner workings of Hadoop
  • Experience designing and implementing data pipelines with a combination of Hadoop, Map Reduce, Hive, Impala, Spark, Kafka, Storm, SQL, Hive, Pig, Impala, Oozie, Sqoop, Zookeeper, Mahout and NoSQL data warehouses
  • BS or MS in Computer Science / Engineering or equivalent
8

Principal Data Engineer Resume Examples & Samples

  • 10 or more years of progressively complex related experience in data science engineering
  • In-depth knowledge of large scale search applications and building high volume data pipelines
  • Expert experience implementing complex, enterprise-wide data transformation and processing solutions
  • In-depth knowledge on Hadoop architecture, HDFS commands and experience designing & optimizing queries to build scalable, modular, and efficient data pipelines
  • Strong leadership and mentoring skills
9

Principal Data Engineer Resume Examples & Samples

  • Develop infrastructure to inform on key metrics, recommend changes, and predict future results
  • Build systems to answer business questions
  • Analyze and improve efficiency, scalability and stability of Vertica clusters
  • Be part of a team that owns Data Management, Data Modelling and ETL processes for Zynga company-wide
  • Help build out the framework as we transition ETL from Vertica to newer technologies
  • BS in Computer Science, Information Management or equivalent
  • 6+ years of experience working with columnar database in a clustered environment, experience with Vertica in particular is a plus
  • Experience building ETL processes for large volumes of data
  • Very strong SQL skills - the ability to write intermediate level SQL joins / group by / etc
  • Solid understanding of both relational and NoSQL database technologies
  • Basic fundamentals of Computer Science and proficiency in at least one procedural language
  • Experience with ETL frameworks like Airflow is highly desirable
  • Experience with managing database systems on the cloud is highly desirable
  • Experience writing real functional programs for data processing in PHP, Python or Bash
  • Experience / ability to debug and solve operational issues
  • Able to take individual ownership of very large complex projects from start to finish. Insatiable curiosity, high aptitude with a flair for business impact
  • Excellent communication skills and the ability to work effectively with others in a team environment
10

Principal Data Engineer Resume Examples & Samples

  • Ideally 8 - 10 years of hands on experience, preferably in a lead data engineer role
  • Ruby, Python, bash, SQL a must and knowledge of other languages like Java, Node.JS bumps up your profile in our list
  • Experience building REST, SOAP, Web services and experience building connectors to SaaS services like SFDC, Zuora, Netsuite, Xactly etc. would be nice
  • Heavy experience with cloud technologies (preferably Google Cloud Platform components), containerization (Kubernetes, Docker, Vagrant), scaling, high availability, high performance systems
11

Principal Data Engineer Resume Examples & Samples

  • Responsible for growing, deploying, and maintaining mission critical security analytics data environment that process data quickly at large scale
  • Contributes design, code, configurations, and documentation for components that manage data ingestion, real time streaming, batch processing, data extraction, transformation, and loading across a broad portion of the existing Hadoop, GreenPlum, and MPP ecosystems
  • Evaluates new and upcoming big data solutions and makes recommendations for adoption to extend our platform to meet advanced analytics use cases, such as predictive modeling and recommendation engines
  • Mentors junior and senior engineers, provides code reviews, feedback, and enables professional growth
  • Comfort and experience with the art and science of extracting insight from large structured and unstructured data sets
  • Experience building, maintaining, and improving Data Pipeline / Data routing in large scale environments
  • Fluency in query languages, API development, data transformation, and integration of data streams
  • Strong experience with large dataset platforms such as (e.g., Greenplum, Hadoop, MongDB, NoSQL, Elasticsearch, Cassandra, etc.)
  • Fluency in multiple programming languages, such as R, Perl, Python, Shell Scripting, Regex, SQL, Java, Alpine, Matlab, or similar languages and tools appropriate for large scale analysis of numerical and textual data
  • Experience with acquiring data from APIs, data queues, flat-file, and remote DB sources as well as creating and modifying APIs for downstream consumers
  • Must have basic Linux administration skills and Multi-OS familiarity (Windows / OSX)
  • Data Pipeline and Data processing experience using common platforms and environments such as Apache Nifi, Flink, Pivotal Spring,
  • Understanding of traditional DW/BI components (ETL, Staging, DW, ODS, Data Marts, BI Tools)
  • Must have a solid understanding of enterprise software development from design and architecture through production
  • Creativity to go beyond current tools to deliver best solution to the problem
  • Ability to work effectively across functions, disciplines, and levels
  • Ability to identify, prioritize, and answer the most critical areas where analytics and modeling with have a material impact
  • Familiarity with Kafka, Logstash, Syslog-ng
  • Familiarity with data visualization tools and techniques
  • Familiarity of information/network security detection tools, logging systems, network architecture, and threat intelligence a plus
  • Familiarity with IP networking concepts (TCP/IP, pcap analysis) a plus
12

Principal Data Engineer Resume Examples & Samples

  • Track the progress that you are making towards the wider goals of your team, and transparently communicate that progress to your team and Technology Manager
  • Keep to the processes, standards and designs that your team has agreed
  • Ensure you and others continuously improve by proposing and then implementing new processes, designs and engineering practices that make your team more effective
  • Contribute to the definition of product features, including in workshops with other product development teams
  • Communicate the technical work of your team to other product development teams
  • Take responsibility for supporting your changes all the way through to proving that they work well – and continue to work well – in production
  • Build and maintain the data pipelines at Just Eat
  • Ensure the architecture is production ready and can be supported by the SOC teams
  • Ensure we are able to scale to meet the ever growing demands of the business
  • Have experience working with streaming solutions
  • Have experience with technologies such as Hadoop, Kinesis, Spark, Storm etc
  • Have worked on data projects with AWS, Azure or GCP
  • Are comfortable in programming languages outside of C# such as Python, Java and Scala
  • Will become a trusted technical authority and specialist in one or more technical areas of your team. You’ll be someone the rest of the team and your Technology Manager can turn to for engineering advice and guidance
  • Collaborate with other product development teams and use their feedback as a basis for improvements
  • Write technically sound, clean, maintainable code – with practical application of established patterns and practices
  • ASP.NET MVC
  • API Development
  • JavaScript/Angular.JS
  • HTML
  • CSS & Responsive Web Design
  • SQL Server, MySQL or PostgreSQL, DynamoDB
  • Experience working in a Continuous Delivery cycle
  • Working in SOA or Microservices Architecture
  • EMR, Streaming solutions, Spark, Lambda Architecture
13

Principal Data Engineer Resume Examples & Samples

  • Become expert on HERE’s 3D tools, their use cases, and extensibility
  • Use HERE's location platform to access location services
  • Develop visualization and analytic services or libraries for accessing, assessing, and publishing map quality data
  • Build low-latency user experiences
  • Support requests from HAD workstreams for 3D visualization and editing tools
  • Display the output of analytic, statistical, predictive, and simulation models
  • Colorize and optimize LIDAR to allow efficient identification and measurement of objects
  • Create 3D visualizations integrating LIDAR, imagery, and other sensor data
  • Integrate batch or streaming data from multiple sources or tools in one visualization
  • A degree in a discipline such as Engineering, Statistics, Applied Mathematics, Computer Science, or Econometrics
  • Five or more years’ experience C++ and OpenGL programming
  • Experience in JavaScript programming on the client or server
  • Experience in Java, Node.js, or Python server programming
  • Knowledge of Database Systems (SQL and NoSQL), Data Modeling, and GIS,
  • Experience using AWS, Linux command line interface, and shell scripting
  • Experience working in an agile program management environment
  • Experience working with geospatial data using GIS tools or libraries
  • Experience building portable applications using SDL2, Emscripten, and Boost
  • Experience programming using modern C++, such as C++11 or C++14
14

Principal Data Engineer Resume Examples & Samples

  • Translating complex functional and technical requirements into detailed architecture, design, and high
  • Performing software
  • Leading Big Data analytical solutions leveraging transformational technologies
  • Driving user ideation analysis and elaboration, design and development of software applications, testing, and builds automation tools
  • Selecting data solution software and defining hardware requirements
  • Translating business requirements into system requirements
  • Building next-generation Big Data analytics framework developed on a group of core
  • Technology
  • Coding, testing, and documenting new or modified data systems to create robust and scalable applications for data analytics
  • Ensuring all automated processes preserve data by managing the alignment of data availability and integration processes
  • Collaborating with data architects, modelers and IT team members on project goals
  • Assisting with extracting standard sets of features from important datasets that will be leveraged by Data Science team
  • Translating code from native R and Python applications into Spark applications
  • Design and implement big data analytics models and solutions, as well as integrate new data management technologies and software engineering tools into existing structures
  • Research opportunities for data acquisition and new uses for existing data, and develop data set processes for data modeling, mining and production
  • Serve as consultant to data scientists, data architect and business project leads, and interprets data results to business customers
  • Develops standards and governance processes for integration projects and initiatives
  • Advanced degree in computer science, engineering, statistics, physics, mathematics, economics, or management of information systems
  • 7 years of experience in data science and data engineering
  • Working knowledge of Open Analytics Platforms (such as Hadoop ecosystem)
  • In depth knowledge of code writing in R, Python, Spark (1.6 and 2.0) for machine learning
  • Knowledge of application integration middleware
  • Fundamental understanding of RDBMS and ETL tools
  • Leadership skills in collaborating with data architects, modelers and IT team members
  • Experience in translating ideas into analytics solutions
  • Experience working with Horton works distribution of Hadoop
  • Ph.D degree in computer science, engineering, statistics, physics, mathematics, economics, or management of information systems
  • Extensive experience in large scale system implementation, and project management
  • Mastery of machine learning, algorithms and advanced mathematics
  • Knowledge of Microsoft SSRS, SSIS, SSAS, and Power BI
  • Excellent verbal and written communication skills as well as the ability to bridge the gap between data science and business management
  • Exceptional organizational skills and is detail oriented
15

Principal Data Engineer Resume Examples & Samples

  • With the number of connected Mercedes-Benz vehicles on the road increasing daily, the amount of data being generated and its significance is also increasing rapidly. The Vehicle Infrastructure & Data Analytics team is looking for passionate and versatile Data Engineers to support with the collection, storage, processing, and analysis of large amounts of vehicle data
  • This position will play a key role in the development and deployment of an innovative Big Data Platform for vehicle data processing and advanced analytics. As a Principal Data Engineer on the team, you'll be responsible for defining and building the data pipelines that combine vehicle data and third-party content and will enable better, faster, data-informed decision-making. As part of a larger global interdisciplinary team, you will also make strategic decisions that influence the platform and data warehouse architecture
  • This is a unique opportunity to join an innovative group of engineers creating the next generation big data platform that will be accessed and used by hundreds of people across domains within Mercedes-Benz
  • Architect and lead the implementation of workflows with varying demands of throughput and latency
  • Evaluate performance and take steps to optimize all parts of the big data stack and implementation
  • Experience in one or more major programming languages (e.g. Unix shell scripting, Python, Java, C++, C#, etc.)
  • Proven record building a large scale data warehouse
  • Strong background in large-scale query processing and data crunching
  • Experience working with low latency, high performance stream processing systems
  • Strong expertise with Hadoop, MapReduce, HDFS and SQL engines
  • Working knowledge of Spring Boot, Spring XD, Pig or Spark is a plus
  • Experienced with test-driven development based on Mockito, Maven & Junit
  • Familiarity with different data warehouse solutions including Cassandra, Redis, HBase, Grenplum or Gemfire
  • Experience with enterprise data platforms including Pivotal's Big Data Suite, Microsoft Azure Analytics or SAP Analytics is a plus
16

Principal Data Engineer Resume Examples & Samples

  • Advanced degree in Computer Science with 3 – 5 years of significant exposure to data architecture, data mining and distributed systems
  • Strong analytical and problem solving skills, good understanding of different data structures, algorithms and their usage in solving business problems
  • Experience with Java, C++, SQL, and Python
  • Experience in database design and integration of data from multiple data sources
  • Knowledge of various ETL techniques and frameworks
  • Hands-on expertise of NoSQL system, MongoDb, Cassandra, Hadoop/map-reduce, Spark, Hbase etc
  • Experience in cloud providers like Azure and AWS preferable
  • Experience in distributed and highly optimized data stores highly desirable
  • Ability to solve any ongoing issues with operating the cluster
  • Ability to work in a fast paced, test-driven, collaborative and iterative environment
  • Demonstrated ability to exchange ideas and convey complex information clearly and concisely
  • Proven ability lead and drive projects and assignments to completion
17

Principal Data Engineer Resume Examples & Samples

  • MSc in Computer Science/Engineering/Mathematics or equivalent
  • 8+ years of backend and web development experience
  • Proven experience in developing scalable analytics and machine learning solutions for production environments with large amounts of data
  • Sound understanding of modeling techniques and algorithms, particularly in the geospatial arena
  • Solid knowledge of a variety of data science tools and an understanding about when to use which
  • Practical experience in the engineering involved in developing a new pipeline is a must
  • Experience with various Apache stream-processing projects essential – Kafka, Flink, Spark
  • Passion for high back-end performance and scalability
  • Experience with handling cloud infrastructure deployments
  • Ability to work with people from a wide variety of disciplines
18

Principal Data Engineer Resume Examples & Samples

  • Develop, enhance, and automate processes for queuing and prioritizing data management and curation requests
  • Develop data processing pipelines to automate data ingestion
  • Identify and integrate new technologies to enhance existing infrastructure (AWS, SaaS, etc.)
  • Create and test end-to-end integration strategies with existing (and projected) scientific applications and services
  • Migrate proof-of-concept projects into robust, fully developed, widely available solutions
  • Empower scientists with tools, processes and data structures needed to support project objectives, including data integration efforts that may span multiple studies, data types and/or experiments
  • Ensure accurate, complete and timely collection, delivery and tracking of analytical information from internal or contract laboratory service providers or collaborating laboratories for curation, ingestion and delivery to computational and translational scientists
  • Assist in defining, delivering and implementing R/ED analytical data management systems, processes and procedures for collaborating and partnered laboratories
  • Work directly with R/ED study teams to develop R/ED information management plans that outline data capture, data flow, data queries, manual quality checks, and data listings needed to ensure data integrity and interpretability
  • Work with computational analytics, biostatisticians and translational scientists to resolve data quality issues
  • Make study data, including raw/interim data, available to R/ED department personnel as required to support timely interpretation for decision making
  • Acquire user feedback to inform business requirements for future data systems refinement and development
  • Actively contribute to Celgene’s R/ED Data Assets initiative to build Celgene-specific knowledge collections for cell lines, animal models and drug compounds based on internal and external clinical, genomic, proteomic, transcriptional and epigenetic assay data
  • Bachelor’s degree in a relevant discipline with at least 16 years’ experience, Master’s degree with at least 14 years’ experience or PhD with at least 8 years’ experience in high complexity data management, including some relevant biomedical assay or biospecimen data management
  • Demonstrated proficiency with molecular biology assay concepts and ability to support, develop and deploy laboratory and other research data management processes and procedures as they apply to complex, high dimensional data sets
  • Extensive practical experience in curating and working with diverse but highly-connected scientific knowledge collections and their query interfaces to enable generation of research hypotheses around compound targets, mechanisms of action, and patient response
  • Demonstrated ability to understand and translate high-level scientific datasets and results into effective data curation and management strategies
  • Working knowledge of cloud computing; preference will be given to candidates with AWS experience
  • Working knowledge of REST APIs and container strategies strongly preferred
  • Knowledge of distributed database design and implementation, LAMP/ MySQL, etc. with capability to perform/direct/assess implementation of such databases
  • Excellent skills in R programming and experience in additional computer languages such as Perl, Python, PHP, S-PLUS or Java (or C/C++/C#)
  • Experience leading technical development projects using Agile methodologies with both internal and external contracted resources is desirable
  • Working knowledge of both Windows and Linux operating systems is required
  • Along with programming proficiency must have creativity, and show a strong capacity for independent thinking and the ability to grasp underlying biological questions
19

Principal Data Engineer Resume Examples & Samples

  • A strong passion to identify and solve real business problems using data
  • Ability to work independently and with minimal or no direction
  • Entrepreneurial mindset to identify and promote innovative travel products, and solve the internal puzzles and prototype the end product
  • Ability to move seamlessly between business problems and coding
  • Good presentation skills to explain complicated analytical solutions to a non-technical group of people, internal or external
  • Creativity in using available tools and out-of-the-box thinking to provide the best solution to problems
  • Ability to rapidly prototype and evaluate innovative travel products and applications
  • Persistence and willingness to learn and apply new techniques/new tools constantly
  • Experience working with very large datasets which happen to be residing in different data stores in different formats
  • Architect, build and launch data models that analytics teams can use to solve business problems using Advanced Analytics and Machine Learning
  • Design and build efficient data pipelines to move data between different source systems
  • Own the solution end-to-end
  • Collaborate with Architecture teams in providing an analytics platform to perform advanced analytics
  • Strong Java programming skills required. Experience with Scala is preferred. Other scripting skills, such as Python or Perl a plus
  • Expert knowledge on Hadoop-related technologies, MapReduce, Spark, Kafka, Hive, Impala, etc
  • Strong knowledge on batch and streaming data processing with structured and unstructured data
  • Familiarity or experience with open source Machine learning and Artificial Intelligence tools such as Spark MLlib, H2o.ai or TensorFlow
  • Experience in scaling Machine learning solutions
  • Strong ability to drive a project from the beginning to the end
  • Experience working with teams who have highly analytical skills
  • Strong desire to learn new technologies and try implementing them
20

Principal Data Engineer Resume Examples & Samples

  • As the Principal Data Engineer for HERE IoT Tracking products you have the opportunity to develop analytics tooling and pipelines to process location data from millions of devices
  • You will work closely with the cloud infrastructure and application development teams to design and maintain scalable data models and pipelines and connect-the-dots across multiple use cases and create statistical modelling and machine learning for wide variety of projects
  • You will be contributing to technology strategy for the tracking product line and take ownership of technical product architecture of analytics and machine learning functionality
  • You will explain complex technical concepts in a simplified manner to our internal and external stakeholders on different levels of organization
  • You will also educate and mentor developers in the techniques and best-practices required to build highly-scalable, performant analytics pipelines
  • You will have your finger on the pulse for emerging technologies and asses their relevance for the product
  • Through leadership you bring passion for innovation to create delightful product experiences
  • Experience with various Apache stream-processing projects essential Kafka, Flink, Spark
21

Principal Data Engineer Resume Examples & Samples

  • Lead event stream analytics initiative and create the foundation for event data collection at Zillow Group
  • Design and build highly scalable and responsive platform to collect the data across all the brands and all devices (mobile apps, desktop …)
  • Work with teams across Zillow group to drive the new platform adoption