Position Role/Tile: Big Data / Cloudera Admin Location: Reston, VA. Independently installs and maintains Big Data (Cloudera, Horton Works, etc.) clusters in high available, load balanced configuration across multiple (Production, User Acceptance, Performance, and Development) environments.
Under general supervision, manage Big Data Administration activities, technical documentation, system performance support, and internal customer support.
May provide input into the development of Systems Architecture for mission critical corporate development projects.
The incumbent will work with Solutions Architects, Infrastructure Architects, Lead Big Data Administrator, Big Data Supplier and Developers to setup environments and support the development teams. Candidate will support the implementation of a Data Integration/Data Warehouse for the NextGen and Informatics Teams Tasks:
Responsible for all the tasks involved in administration of ETL Tool (Ab-Initio). Maintaining access, licensing and file system in the ETL server.
Provide guidance on the design and integration of ETL to ETL Developers.
Manage metadata hub, Operational Console and troubleshoot environmental issues which affect these components.
Responsible for technical Metadata management.
Work with the team to maintain data lineage and resolve data lineage issues.
Design and develop automated ETL process and architecture. Interact with the Client on daily basis to define the scope of different applications.
Work on the break fix and continuous development items and review and inspection for the production changes.
Perform the code review for the ETL code developed by the development team and guide to resolve any issues.
Work with the various other groups - DBAs, Server Engineer Team, Middleware Group, Citrix Group, Network Group, data transmission, etc.
to resolve the performance related/integration related issues. Provide update on progress of infrastructure/development and unit testing tasks to the client on weekly meetings.
Required Skills:
This position requires a BA/BS in Computer Science, Information Systems, Information Technology or related field with 3-5 years of prior experience in software development, Data Warehousing and Business Intelligence OR equivalent experience.
Administrator experience working with batch processing and tools in the Hadoop technical stack (e.g. MapReduce, Yarn, Pig, Hive, HDFS, Oozie) Administrator experience working with tools in the stream processing technical stack (e.g. Spark, Storm, Sama, Kafka, Avro) Administrator experience with NoSQL stores (e.g. ElasticSearch, Hbase, Cassandra, MongoDB, CouchDB) Expert knowledge on AD/LDAP security integration with Big Data Hands-on experience with at least one major Hadoop Distribution such as Cloudera, Horton Works, MapR or IBM Big Insights Advanced experience with SQL and at least two major RDBMSs Advanced experience as a systems integrator with Linux systems and shell scripting Advanced experience doing data related benchmarking, performance analysis and tuning, troubleshooting Excellent verbal and written communication skills
Big plus to have: Hbase, Solr, Spark, and Kafka Experience
Software & Tools:
BPM/BRE/SOA, Java/JEE, SOAP/REST, MQ, Spring, Hibernate/MyBatis, DB2, COBOL
Central Business Solutions, Inc,
37600 Central Ct.
Suite #214
Newark, CA 94560.