Job description

Job Description


    Job Title:  Data Engineer 

    Work Location: Bentonville, AR
     

    Job Description: 

    We are seeking a highly skilled Data Engineer to join our team in Bentonville, AR. The ideal candidate will have extensive experience with Google Cloud Platform (GCP), Apache Spark, and Scala. This role involves designing, building, and maintaining scalable data pipelines, ensuring optimal performance, and integrating cloud solutions. If you are passionate about data engineering and cloud technologies, we would love to hear from you.

    Key Responsibilities:

    • Data Pipeline Development: Design, develop, and maintain robust data pipelines using Spark and Scala. Ensure data is processed efficiently and made available for analysis on time.
    • Cloud Integration: Leverage GCP to build and manage scalable cloud-based data solutions. Understand cloud design considerations and limitations, and their impact on performance.
    • Performance Optimization: Monitor and optimize data pipeline performance. Identify bottlenecks and implement solutions to enhance efficiency and reduce latency.
    • CI/CD Deployment: Set up and manage Continuous Integration/Continuous Deployment (CI/CD) pipelines in Git. Ensure smooth deployment and integration of data solutions.
    • Development Tools: Utilize IntelliJ IDE for development tasks. Implement and manage workflows using Apache Airflow for cloud scheduling.
    • Collaboration: Work closely with cross-functional teams to understand data requirements and provide solutions that meet business needs. Foster a collaborative environment to drive data-driven decision-making.
    • Documentation: Create and maintain comprehensive documentation for data pipelines, cloud integrations, and best practices. Ensure knowledge sharing and continuity within the team.
       

    Qualifications:

    • Bachelor’s degree in Computer Science, Information Technology, or a related field.
    • Proven experience as a Data Engineer, with a strong focus on GCP, Spark, and Scala.
    • Proficiency in Spark and Scala programming for building and automating data pipelines.
    • Experience with setting up and managing CI/CD pipelines in Git.
    • Proficiency in using IntelliJ IDE for development and Apache Airflow for cloud scheduling.
    • Strong understanding of cloud design considerations and limitations, and their impact on performance.
    • Excellent problem-solving skills and attention to detail.
    • Strong communication and collaboration skills.
       

    Skill requirement

    1. Proficiency in Programming Languages:

    • Strong experience with programming languages commonly used in data engineering, such as Python, Java, Scala, and SQL.
    • Familiarity with shell scripting for automation tasks.
       

    2. Database and Data Warehousing Knowledge:

    • In-depth knowledge of both SQL (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra).
    • Experience with data warehousing solutions such as Amazon Redshift, Google BigQuery, Azure SQL Data Warehouse, or Snowflake.
    • Understanding of ETL (Extract, Transform, Load) processes and tools.
       

    3. Big Data Technologies:

    • Expertise in big data frameworks like Apache Hadoop, Spark, Hive, Flink, or Kafka.
    • Familiarity with distributed data storage and processing systems (e.g., HDFS, S3, Google Cloud Storage).
       

    4. Data Modeling and Data Architecture:

    • Experience in data modeling techniques, including relational, dimensional, and star schema modeling.
    • Ability to design, build, and manage data pipelines and data architectures that scale.
       

    Benefits:

    • Competitive salary and performance-based bonuses.
    • Comprehensive health, dental, and vision insurance.
    • Retirement savings plan with company match.
    • Opportunities for professional development and career growth.
    • Flexible work hours and remote work options.
       

    About Us:

    We are a forward-thinking company dedicated to leveraging data to drive innovation and business success. Our team is composed of passionate professionals who are committed to excellence and continuous improvement. We value creativity, collaboration, and a strong work ethic. Join us and be part of a dynamic team that is shaping the future of data engineering.

    If you are a skilled Data Engineer with expertise in GCP, Spark, and Scala, and a passion for building scalable data solutions, we want to hear from you! Apply now to join our team and contribute to our mission of harnessing the power of data to achieve remarkable outcomes.