Sr. Data Engineer - Mondo Staffing : Job Details

Sr. Data Engineer

Mondo Staffing

Job Location : Hughson,CA, USA

Posted on : 2025-08-05T07:51:07Z

Job Description :
Apply now: Sr. Data Engineer, Data Solutions\n \n Location: Onsite, Houston, TX (4 days/week, Mon-Thurs) \n Start Date: 2 weeks from offer for this 6-month contract position\n \n \n Job Title: Sr. Data Engineer, Data Solutions\n \n Location-Type: Onsite, Houston, TX (4 days/week, Mon-Thurs) (Remote-Fridays)\n \n Start Date Is: 2 weeks from offer\n \n Duration: 6 month contract (potential to extend)\n \n Compensation Range: $50 - $70/hr w2\n Role Overview: \n Client is seeking a Sr. Data Engineer to build and optimize data pipelines, ETL processes, and cloud-based data architectures. This role focuses on Python development, cloud data integration, and supporting advanced analytics and BI initiatives. You'll collaborate with cross-functional teams to enhance data flows and drive data product innovation. \n Day-to-Day Responsibilities: \n
    \n
  • Develop Python modules using Numpy, Pandas, and dynamic programming techniques
  • \n
  • Design, build, and optimize ELT/ETL pipelines using AWS and Snowflake
  • \n
  • Manage data orchestration, transformation, and job automation
  • \n
  • Troubleshoot complex SQL queries and optimize data queries
  • \n
  • Collaborate with Data Architects, Analysts, BI Engineers, and Product Owners to define data transformation needs
  • \n
  • Utilize cloud integration tools (Matillion, Informatica Cloud, AWS Glue, etc.)
  • \n
  • Develop data validation processes to ensure data quality
  • \n
  • Support continuous improvement and issue resolution of data processes
  • \n
\n Must-Haves: \n
    \n
  • Bachelor's degree in Computer Science, Engineering, or related field
  • \n
  • 8 years of experience in data engineering, ETL, data warehouse, and data lake development
  • \n
  • Strong SQL expertise and experience with relational databases
  • \n
  • Experience with Snowflake, Amazon Redshift, or Google BigQuery
  • \n
  • Proficiency in AWS services: EC2, S3, Lambda, SQS, SNS, etc.
  • \n
  • Experience with cloud integration tools (Matillion, Dell Boomi, Informatica Cloud, Talend, AWS Glue)
  • \n
  • GitHub version control knowledge and its integration with ETL pipelines
  • \n
  • Familiarity with Spark, Hadoop, NoSQL, APIs, and streaming data platforms
  • \n
  • Python (preferred), Java, or Scala scripting experience
  • \n
  • Agile/Scrum development experience
  • \n
\n Soft Skills: \n
    \n
  • Excellent communication and collaboration skills
  • \n
  • Strong problem-solving abilities and attention to detail
  • \n
  • Ability to multitask and meet tight deadlines
  • \n
  • Highly motivated and self-directed
  • \n
  • Comfortable working with technical and business stakeholders
  • \n
\n Nice-to-Haves: \n
    \n
  • Experience with data quality and metadata management tools
  • \n
  • Exposure to BI platforms and data visualization tools
  • \n
  • Experience in building event-driven data architectures
  • \n
Apply Now!

Similar Jobs ( 0)