• Sr. Engineer (Big Data)

    Job Locations US-GA-Lawrenceville
    Job Post Information* : Posted Date 11 months ago(12/6/2017 12:02 PM)
    # of Openings
    Information Technology
  • Overview


    Our mission is to be the world’s single, most- trusted source of truth regarding healthcare institutions, providers, pharmaceuticals and devices to improve outcomes through transparency and accountability.

    We believe access to accurate, complete information is at the center of healthcare.  MedFax continually ingests data from more than 10,000 sources, creating the first truly validated, trustworthy, actionable tool


    This is a full-time position with competitive salary plus benefits.

    MedFax provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, sex, sexual orientation, gender identity or national origin, citizenship, veteran’s status, age, disability status, genetics or any other category protected by federal, state, or local law.



    The Senior Software Engineer with a focus on Big Data is responsible for joining a product team and contributing to the software design, software development, and overall product lifecycle for a product that our users love. The engineering process is highly collaborative. Software Engineers are expected to collaborate on a daily basis as they work through user stories and support products as they evolve. As a Software Engineer you will grow your skills while you create, support, and deploy production applications.

    As a Senior Software Engineer, it is a critical part of your responsibilities to mentor and guide other engineers on the team and provide both technical as well as personal leadership.

    In addition, Software Engineers may be involved in product configuration, performance tuning and testing as well as production monitoring.

    • Design, develop and implement applications that critical operations.
    • Provide innovative solutions to complex Big Data initiatives.
    • Plan, develop and implement large-scale projects from conception to completion.
    • Develop and architect lifecycle of projects working on different technologies and platforms.
    • Execute system development and maintenance activities.
    • Develop solutions to improvise performance and scalability of systems.
    • Proven technical troubleshooting skills with initiative to lead and resolve hard issues.
    • Excellent communication and documentation skills.
    • Highly motivated and to work in team & deliver without close supervision.
    • Excellent Analytical skills.

    Required Skills

    • Manage large scale Hadoop clusters including design, capacity planning, partner with Ops admin for cluster set up etc.
    • Demonstrated experience with the Hadoop ecosystem (Hive, Pig, MapReduce, Spark, Flume, Sqoop, Strom etc.)
    • Good Understanding of Data format such JSON, Avro, Carbon etc
    • Demonstrated experience with ETL, and experience with ETL tools
    • Development experience in one or more of the following languages: Java, Python and Scala
    • Demonstrated experience with SQL and no-SQL databases (SQLServer, Hbase, Cassandra, MySQL, VoltDB)
    • Experience with Cloud offerings such as MicroSoft Azure
    • Experience with DevOps/CI tools such as Git Hub, Puppet, CloudFormation and Jenkins
    • Performance Tuning of the Hadoop cluster & it’s components to meet business SLA.
    • Manage multi-tenant business users in the same cluster.
    • Set up resource management & user queues.
    • Automate monitoring & alerting.
    • Manage the security set up – Kerberos authentication & authorization using Ranger.
    • Establish/socialize access policies within the team and with business partners.
    • Identify and document best practices, standards and guidelines.
    • Evaluate and recommend software & hardware needs for Big Data Platform.
    • Contribute to evolving architecture of our Hadoop eco system with cloud considerations.
    • Partner with release engineer to automate deployments.
    • Create metrics / measurement of utilization and performance.
    • Research and recommend innovative and automated approaches for administrative tasks.
    • Provide guidance and mentor the technical teams by providing solutions, tuning recommendations and documentation of use cases for continuous improvement.
    • Work with Architecture and Development teams to understand usage patterns and work load requirements of new projects in order to ensure the Hadoop platform can effectively meet performance requirements and service levels of applications.
    • Develop and enhance platform best practices and educate developers on best practices.
    • Serve as an escalation contact for critical production issues.
    • Perform other work-related duties as assigned.
    • Consistently demonstrates regular, dependable attendance & punctuality.
    • Strong estimating, planning skills and proven ability as a problem-solver
    • Ability and desire to thrive in a proactive, high-pressure, SLA bound environment

     Communication Skills:

     Excellent verbal and written communications skills






    • Bachelor’s Degree in Computer Science or other related discipline
    • 7+ years of professional experience supporting production medium to large scale Linux environments
    • 5+ years of professional experience working with Big Data (HDFS, Sqoop, Hive, Spark) and related open source technology
    • Fluency in python, shell scripts and other languages. Selenium is plus.
    • Deep understanding of Hadoop design principles, cluster concepts, security, distributed system performance
    • SQL skills, including experience querying large, complex data sets and performance analysis
    • Experience implementing Hadoop in Cloud is a plus
    • Experience developing highly performant distributed data processing platform with Big Data technologies
    • Experience in the high paced energy of startups as well as experience in the structured world of larger companies
    • Understanding of systems design and technical concepts in order to communicate effectively with other cross-functional teams
    • Strong knowledge and experience with Agile/Scrum methodology and iterative practices in a service delivery lifecycle


    Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
    Share on your newsfeed