Lead Database Engineer

  • Dublin
  • Oclc
You have a life. We like that about you.At OCLC, we believe you'll do the best work of your life when you're living the best life possible.We work hard to build the technology that connects thousands of today's libraries. But we also work hard to make a job at OCLC a meaningful part of a balanced life- not a substitute for one.Technology with a Purpose. OCLC supports thousands of libraries in making information more accessible and more useful to people around the world. OCLC provides shared technology services, original research and community programs that help libraries meet the ever-evolving needs of their users, institutions, and communities. With office locations around the globe, OCLC employees are dedicated to offering premier services and software to help libraries.The Job Details are as follows:We are looking for people who are passionate around solving problems through innovation and engineering best practices! You will be part of the Data Platform Engineering team responsible for administration, infrastructure planning, and scaling of our Big Data Hadoop and relational database technologies within OCLC. This position will be responsible for the administration and support of multiple open-source Hadoop NoSQL environments with a secondary role supporting Snowflake and Kafka on-premise and within AWS. You will work closely with infrastructure and development teams to ensure business applications are highly available and performing within agreed on service levels.This position will work closely with the development teams to tune and troubleshoot applications to maintain high performing environments. Additionally, you will provide guidance for best practices and standards for the data platforms to your development partners.Responsibilities:Install, manage, maintain, and tune Hadoop/HBase clustersDevelop and enforce operational best practices to maintain availability and responsiveness of HDFS/Hbase cluster resourcesWork in embedded development teams to understand usage patterns and tune application and Hadoop cluster interaction to provide reliable performance and limit impact to other data usersParticipate in on-call support rotationReview existing software/hardware architecture and identify opportunities for improvement in the areas of scalability, maintainability, and performanceMonitor Hadoop cluster connectivity and securityUnderstand and review Hadoop log filesHDFS support and maintenance.QualificationsHands on experience with the Hadoop stack (HBase, Yarn, Spark/MapReduce, Kafka)Tuning and hardening of data platforms, using open source/native tools and/or internal solutionsAbility to deploy a Hadoop cluster, add and remove nodes, manage jobs, monitor critical components of the cluster, configure name-node high availability, configure backups.Strong understanding and support of the Linux operating system (Red Hat preferred)Strong knowledge of scripting and automation tools (e.g. Ansible, Bash, Python, GitLab)Ability to read, understand, and write java codeExpertise in typical system administration duties such as capacity management, performance tuning, system dump analysis, and server hardening (security)7+ years relevant IT experienceAdditional Skills/Experience desired:Experience with Snowflake and AWS technologies (EC2, S3, RDS, Aurora, DynamoDB)Experience with other database technologies (PostgreSQL, Cassandra, MongoDB, MySQL)Experience supporting a production multi-site enterprise environmentExperience working within a Scrum/Agile frameworkWorking Conditions: Normal office environment.ADA/EAA: The above statements cover what are generally believed to be principal and essential functions of this job. Specific circumstances may allow or require some people assigned to the job to perform a somewhat different combination of duties.