We are seeking a highly skilled Scientific Data Scientist or Software Engineer with a strong academic foundation and hands-on experience developing scientific software. The ideal candidate combines deep STEM expertise with rigorous coding ability, and has a demonstrated track record of integrating, evaluating, and advancing scientific software systems. <br> Technical Skills Fluent proficiency in Python (required). Strong ability to design, implement, and maintain scientific software applications and codebases—beyond general AI/ML model development. Experience contributing to third-party or open-source scientific software projects.
<p><strong>Responsibilities:</strong></p><ul><li>Collect, process, and analyze large structured and unstructured datasets to identify meaningful trends, patterns, and opportunities for business improvement</li><li>Develop, test, and deploy predictive models, machine learning algorithms, and statistical analyses to address key business challenges </li><li>Collaborate with cross-functional teams, including business analysts, engineers, and stakeholders, to identify analytics solutions and align deliverables with strategic goals </li><li>Communicate complex findings and recommendations clearly to technical and non-technical audiences through reports, dashboards, and visualizations</li><li>Automate repetitive tasks, streamline data flows, and ensure data quality and governance throughout the analytics lifecycle</li><li>Stay updated on industry trends, emerging technologies, and best practices in data science and AI to continuously enhance solutions</li></ul><p><br></p>
<p>As our portfolio of AI-driven solutions continues to expand, we’re looking for an experienced <strong>Machine Learning Engineer</strong> to join our high-impact data science team. This role offers the opportunity to work across trading, operations, and support functions—delivering production-grade machine learning systems that solve real business problems.</p><p>You’ll collaborate with data scientists, software engineers, and commercial stakeholders to design, build, and deploy models that drive decision-making and innovation. From project scoping to model deployment, you’ll have visibility and influence across the full ML lifecycle.</p><p>🔧 Core Responsibilities</p><ul><li>Act as a thought partner to commercial teams, identifying high-value opportunities for AI/ML applications</li><li>Lead the design, development, and deployment of machine learning systems, with a focus on <strong>NLP</strong>, <strong>LLMs</strong>, and <strong>Generative AI</strong></li><li>Prioritize projects based on business impact and evolving market conditions</li><li>Collaborate with cross-functional teams to gather requirements and align solutions with strategic goals</li><li>Integrate ML solutions—including GenAI—into existing platforms to ensure seamless user experiences and scalable adoption</li><li>Participate in code reviews, experiment design, and tooling decisions to maintain high engineering standards</li><li>Share knowledge and mentor colleagues to build machine learning fluency across the organization</li></ul><p><br></p>
We are looking for a skilled Data Reporting Analyst to join our team in Cincinnati, Ohio. This long-term contract position offers an exciting opportunity to contribute to data-driven decision-making processes within the Personal Loans Marketing team. The ideal candidate will leverage their expertise in data analysis, reporting, and visualization to provide actionable insights and optimize marketing performance.<br><br>Responsibilities:<br>• Develop and maintain efficient data workflows to support marketing dashboards and reporting systems.<br>• Document data sources, definitions, and reporting logic to ensure consistency and governance.<br>• Optimize database queries and backend data processes to enhance scalability and performance.<br>• Conduct quality assurance checks to maintain data integrity and accuracy across reports.<br>• Automate recurring reports to monitor campaign metrics such as performance, conversions, and cost-per-acquisition.<br>• Collaborate with engineering and data science teams to streamline data pipelines and improve data quality.<br>• Design and implement interactive dashboards using tools like Tableau to visualize key performance indicators.<br>• Provide ad hoc data analysis and reporting to support marketing campaign decision-making.<br>• Ensure data models and workflows are scalable and align with business needs.
<p>Design and manage data pipelines, ensuring optimized performance for analytics and reporting. Support BI tools to provide actionable insights for decision-making.</p>
We are looking for a skilled Patent Agent to join our team in Hayward, California. In this role, you will play a key part in managing and expanding our intellectual property portfolio while collaborating with legal counsel, engineers, and leadership. This position offers an exciting opportunity to contribute to innovative projects in a fast-paced environment.<br><br>Responsibilities:<br>• Draft and review patent applications while coordinating with outside counsel to ensure accuracy and quality.<br>• Oversee the management of the company's patent and trademark portfolio, working closely with teams such as R& D, marketing, and business development.<br>• Maintain and update the internal patent database and docket, including performing related administrative tasks.<br>• Evaluate invention disclosures, collaborate with inventors, conduct patentability searches, and formulate effective search strategies.<br>• Conduct analysis of patents for due diligence, freedom to operate assessments, and patentability evaluations.<br>• Monitor patent alerts and watches using tools like PatBase or PatDoc to stay informed on relevant developments.
<p>Are you passionate about building AI-powered systems that automate and orchestrate complex workflows? Our team is seeking a Workflow Engineer with a focus on AI & LLM Platforms to design, implement, and scale cutting-edge solutions for next-generation agentic and data-driven applications.</p><p><br></p><p><strong>Key Responsibilities:</strong></p><ul><li>Architect and implement workflow solutions integrating Large Language Models (LLMs) and machine learning components.</li><li>Build and optimize AI-powered, agentic workflow patterns for real-world business processes.</li><li>Develop robust orchestration logic using Python or Node.js.</li><li>Collaborate with AI engineers and data scientists to design and integrate agentic systems and applied data science solutions.</li><li>Implement Retrieval Augmented Generation (RAG) pipelines to enhance LLM- and data-driven applications.</li><li>Automate and monitor machine learning workflows, ensuring performance, reliability, and scalability.</li><li>Support deployment and lifecycle management of production AI/ML workflows on modern platforms.</li><li>Document best practices and promote knowledge sharing related to workflow automation and AI/LLM system integration.</li></ul>
<p>Position Overview</p><p>We are seeking a Materials Scientist / Product Testing Engineer to support hands‑on laboratory testing, materials analysis, and preparation of technical documentation for regulatory submissions. This role is primarily onsite and will support the Advanced Technology team by conducting experiments, generating product‑representative samples, and producing high‑quality protocols and technical reports.</p><p>The ideal candidate is precise, detail‑oriented, and comfortable working in a regulated environment with cross‑functional teams, suppliers, and external partners.</p><p><br></p><p>Key Responsibilities</p><p>Laboratory Testing & Materials Analysis</p><ul><li>Conduct laboratory experiments and analyze test data for product‑representative material samples</li><li>Support mechanical and chemical testing of ceramics, metals, and biomaterials (e.g., tensile bond strength, X‑ray diffraction)</li><li>Summarize findings and provide clear, concise interpretations</li></ul><p>Regulatory Documentation & Reporting</p><ul><li>Prepare technical memos, protocols, and comprehensive reports for regulatory submissions</li><li>Document current processes, test methods, observations, and final results</li><li>Ensure all documentation meets required regulatory and internal quality standards</li></ul><p>Cross‑Functional Coordination</p><ul><li>Collaborate with internal teams and external suppliers/vendors to obtain samples and execute required testing</li><li>Coordinate project activities and ensure timely delivery of test results</li><li>Communicate findings with clarity to scientists, engineers, and project leaders</li></ul>
<p>The AI/ML Solutions Architect will lead the design, development, and deployment of advanced AI/ML solutions. This role combines deep technical expertise with strategic thinking to ensure AI/ML initiatives are successfully integrated into business operations. You will work closely with data scientists, engineers, and stakeholders to create architectures that maximize performance, scalability, and reliability.</p><p> </p><p><strong>Key Responsibilities:</strong></p><ul><li>Design end-to-end AI/ML architectures, including data pipelines, model training, deployment, and monitoring.</li><li>Collaborate with stakeholders to define AI/ML solution requirements aligned with business objectives.</li><li>Provide technical leadership and guidance to teams implementing AI/ML models and systems.</li><li>Develop scalable and secure solutions using cloud platforms (AWS, Azure, GCP) and MLOps best practices.</li><li>Ensure seamless integration of AI/ML models into existing IT systems and workflows.</li><li>Conduct feasibility studies, prototyping, and performance evaluations for new technologies and frameworks.</li><li>Stay updated on advancements in AI/ML and recommend innovative solutions to meet emerging needs.</li><li>Document technical designs, workflows, and implementation plans to ensure clarity and reproducibility.</li></ul><p><br></p>
We are looking for a skilled Data Engineer to support our organization's data initiatives in Savannah, Georgia. This Contract to permanent role focuses on managing, optimizing, and securing data systems to drive strategic decision-making and improve overall performance. The ideal candidate will work closely with technology teams, analytics departments, and business stakeholders to ensure seamless data integration, accuracy, and scalability.<br><br>Responsibilities:<br>• Design and implement robust data lake and warehouse architectures to support organizational needs.<br>• Develop efficient ETL pipelines to process and integrate data from multiple sources.<br>• Collaborate with analytics teams to create and refine data models for reporting and visualization.<br>• Monitor and maintain data systems to ensure quality, security, and availability.<br>• Troubleshoot data-related issues and perform in-depth analyses to identify solutions.<br>• Define and manage organizational data assets, including SaaS tools and platforms.<br>• Partner with IT and security teams to meet compliance and governance standards.<br>• Document workflows, pipelines, and architecture for knowledge sharing and long-term use.<br>• Translate business requirements into technical solutions that meet reporting and analytics needs.<br>• Provide guidance and mentorship to team members on data usage and best practices.
<p>We are looking for a talented Systems Integration & AI Automation Specialist to join our team in Westport, Connecticut. In this role, you will design, implement, and optimize advanced AI solutions and system integrations to drive operational efficiency and enhance business processes. You will collaborate with cross-functional teams to identify opportunities for automation and AI enablement, ensuring scalable and impactful results.</p><p><br></p><p>Responsibilities:</p><p>• Develop and implement AI-driven solutions to optimize business workflows and enhance operational efficiency.</p><p>• Design and deploy integrations across enterprise systems such as Microsoft 365, Salesforce, and other SaaS platforms.</p><p>• Identify and resolve data silos, workflow fragmentation, and system bottlenecks to streamline processes.</p><p>• Collaborate with global teams to unify identity, access, and governance processes across systems.</p><p>• Partner with departments like Finance and HR to automate manual workflows using tools such as PowerAutomate and Zapier.</p><p>• Create and maintain dashboards to monitor automation performance and its impact on business operations.</p><p>• Deploy custom AI tools, including GPT models and intelligent agents, to improve productivity across various departments.</p><p>• Ensure responsible use of AI technologies in compliance with company policies and global privacy standards.</p><p>• Evaluate and implement AI use cases to address high-impact business needs, leading the development and execution of solutions.</p>
<p>Our client is seeking a Data Scientist II – Generative AI to join a cutting-edge team focused on building scalable, production-ready AI solutions that transform business workflows and deliver measurable impact across global operations. This role is ideal for professionals passionate about leveraging Generative AI technologies, creating intelligent agents, and driving innovation at scale.</p><p><br></p><p>You will design and implement GenAI-powered agents that streamline internal processes, enhance productivity, and support business development initiatives. Responsibilities include developing robust prompt engineering frameworks, building RAG pipelines, and converting prototypes into production-ready solutions. You’ll collaborate closely with engineering and business teams to ensure solutions meet diverse client needs and are optimized for global deployment.</p><p><br></p><p>Key projects include extending the company’s GPT platform, creating AI agents that improve efficiencies for RFP development, onboarding materials, and SOW requirements. Success in this role means quickly ramping up on backlog projects, delivering high-priority initiatives, and staying ahead of emerging GenAI frameworks to continuously advance internal AI capabilities.</p>
<p>The Data Engineer role focuses on designing, building, and optimizing scalable data solutions that support diverse business needs. This position requires the ability to work independently while collaborating effectively in a fast-paced, agile environment. The individual in this role partners with cross-functional teams to gather data requirements, recommend enhancements to existing data pipelines and architectures, and ensure the reliability, performance, and efficiency of data processes.</p><p>Responsibilities</p><ul><li>Support the team’s adoption and continued evolution of the Databricks platform, leveraging features such as Delta Live Tables, workflows, and related tooling</li><li>Design, develop, and maintain data pipelines that extract data from relational sources, load it into a data lake, transform it as needed, and publish it to a Databricks-based lakehouse environment</li><li>Optimize data pipelines and processing workflows to improve performance, scalability, and overall efficiency</li><li>Implement data quality checks and validation logic to ensure data accuracy, consistency, and completeness</li><li>Create and maintain documentation including data mappings, data definitions, architectural diagrams, and data flow diagrams</li><li>Develop proof-of-concepts to evaluate and validate new technologies, tools, or data processes</li><li>Deploy, manage, and support code across non-production and production environments</li><li>Investigate, troubleshoot, and resolve data-related issues, including identifying root causes and implementing fixes</li><li>Identify performance bottlenecks and recommend optimization strategies, including database tuning and query performance improvements</li></ul>
We are looking for a highly experienced Lead Artificial Intelligence (AI) Engineer to spearhead the development and implementation of cutting-edge AI and Machine Learning solutions within our organization. This role is integral to driving innovation, optimizing operational processes, and delivering impactful business outcomes across various projects. As a senior technical expert, you will collaborate with cross-functional teams to design scalable systems, set technical standards, and mentor emerging talent in AI engineering.<br><br>Responsibilities:<br>• Design, develop, and deploy production-ready AI and machine learning solutions tailored to renewable energy planning, construction optimization, and risk management.<br>• Establish and enforce technical standards for AI/ML development, MLOps pipelines, and the management of model lifecycles.<br>• Collaborate with IT and data teams to ensure AI systems are secure, scalable, and seamlessly integrated into the enterprise environment.<br>• Work closely with business stakeholders to translate strategic goals into practical AI applications that drive measurable results.<br>• Lead initiatives to explore and implement generative AI, predictive analytics, and optimization technologies to enhance forecasting and operational efficiency.<br>• Mentor and guide technical teams, fostering knowledge-sharing and the adoption of best practices in AI development.<br>• Evaluate emerging AI tools and technologies to ensure the organization remains at the forefront of innovation.<br>• Oversee the deployment and scaling of AI models in production environments, ensuring performance and reliability.<br>• Drive the development of automation solutions using AI to streamline processes and improve productivity.<br>• Collaborate with partners and vendors to integrate AI solutions effectively into existing systems.
We are looking for an experienced Data Engineer Lead to join our team in Columbus, Ohio on a contract basis. In this role, you will be responsible for leading the development and operation of data pipelines, ensuring seamless integration and delivery of data for analytics initiatives. As a senior member of the team, you will also take on a mentorship role, guiding and developing less experienced team members while driving technical excellence.<br><br>Responsibilities:<br>• Design, build, and maintain robust data pipelines to support enterprise-wide analytics initiatives.<br>• Collaborate with data science and business teams to refine data requirements and ensure streamlined data consumption.<br>• Lead efforts to renovate and automate data management infrastructure to enhance integration and processing efficiency.<br>• Implement and enforce data quality standards to ensure accuracy, consistency, and reliability of data.<br>• Provide training and guidance to colleagues on data preparation techniques and tools.<br>• Partner with data governance teams to curate and promote reusable data content across the organization.<br>• Communicate complex data insights effectively to both technical and non-technical stakeholders.<br>• Stay informed on emerging technologies, assessing their potential impact and integrating relevant advancements.<br>• Offer leadership, coaching, and mentorship to team members, fostering a collaborative and growth-oriented environment.<br>• Work closely with stakeholders to understand business goals and align services to meet those needs.
<p>We are looking for an AI Engineer to join our team. In this role, you will contribute to the development and implementation of AI and Machine Learning solutions that optimize renewable energy projects. You will work on creating scalable models, applications, and workflows that drive data-driven decision-making across the organization while adhering to established engineering standards and practices.</p><p><br></p><p>Responsibilities:</p><p>• Develop, test, and deploy AI/ML models and applications to enhance renewable energy planning, forecasting, and construction processes.</p><p>• Build and maintain MLOps workflows, including data pipelines, model packaging, versioning, monitoring, and retraining.</p><p>• Collaborate with IT and data teams to ensure AI solutions meet security, integration, and performance requirements.</p><p>• Break down technical requirements into actionable tasks and contribute to design implementation during reviews.</p><p>• Deliver scalable and secure solutions by following established standards and reference architectures.</p><p>• Support deployment of AI/ML solutions to cloud environments, with a focus on Azure.</p><p>• Create APIs and integrate AI solutions into enterprise systems to enable seamless operations.</p><p>• Utilize advanced machine learning frameworks such as TensorFlow and Scikit-learn to develop innovative solutions.</p><p>• Analyze data engineering concepts and apply them to enhance AI workflows.</p><p>• Provide technical input and partner with stakeholders to meet project objectives</p>
We are looking for a dedicated Data Scientist to join our team in Louisville, Kentucky. In this role, you will leverage advanced data science techniques to analyze healthcare data, develop predictive models, and deliver actionable insights that enhance patient care and operational efficiency. This position offers the opportunity to make a meaningful impact by addressing health disparities and supporting innovative healthcare initiatives.<br><br>Responsibilities:<br>• Utilize machine learning, predictive analytics, and data mining techniques to uncover patterns and trends in healthcare datasets.<br>• Design and implement predictive models to anticipate patient outcomes and guide proactive clinical interventions.<br>• Create clear and impactful data visualizations to communicate analytical findings to both technical and non-technical audiences.<br>• Collaborate with Data Architects to develop scalable, cloud-based analytic solutions tailored to population health and value-based care.<br>• Establish and monitor KPIs to assess the effectiveness of data-driven strategies in improving patient outcomes and efficiency.<br>• Participate in healthcare projects that focus on chronic disease management and enhancing provider performance.<br>• Stay updated on emerging technologies and methodologies to continuously enhance data science capabilities within the organization.<br>• Ensure compliance with data security and privacy regulations while handling sensitive healthcare information.
Join our team as a Business Intelligence Software Engineer and help design, build, and maintain innovative reporting and data-driven applications that power field operations, business units, and customer solutions. This is a hands-on coding role that requires strong technical judgment and collaboration with cross-functional teams. You’ll manage the entire development lifecycle, ensuring solutions are scalable, reliable, and aligned with business priorities. Key Responsibilities: Lead the Software Development Lifecycle (SDLC): Oversee all phases of BI application development, from concept through deployment and support. Hands-on Development: Build and maintain applications using Python (PySpark), SQL, and TypeScript/JavaScript. Technical Strategy & Architecture: Apply best practices for design, performance, and scalability. Quality Assurance: Establish testing frameworks, conduct code reviews, and maintain bug-tracking processes. Continuous Improvement: Identify and implement tools and methodologies to streamline development and increase system reliability. Collaboration: Work with internal stakeholders, data scientists, analysts, and operations teams to translate business needs into software solutions. Support & Maintenance: Provide ongoing support for newly developed applications, ensuring smooth integration with existing systems.
<p>Position Overview</p><p>We are seeking a highly skilled Data Scientist with deep expertise in Artificial Intelligence, Natural Language Processing (NLP), Computer Vision (CV), and Generative AI. This role is ideal for an innovative problem‑solver who thrives in a fast‑paced environment and is passionate about developing advanced AI/ML models that drive meaningful business value.</p><p>You will design, train, and deploy cutting‑edge AI models—including large language models (LLMs) and multi‑agent systems—turning complex data into scalable, high‑impact solutions.</p><p><br></p><p>Key Responsibilities</p><ul><li>Develop, train, and optimize machine learning and deep learning models</li><li>Build advanced AI solutions leveraging LLMs, multi‑agent systems, fine‑tuning methods, and inference optimization</li><li>Translate complex data science methodologies into clear, actionable insights for business stakeholders</li><li>Collaborate with cross‑functional teams to ensure AI solutions align with business needs</li><li>Create compelling presentations, dashboards, and data stories for non‑technical audiences</li><li>Contribute to innovation initiatives involving NLP, CV, Generative AI, and predictive analytics</li></ul>
<p>We are seeking a Quantitative Researcher to join a high-performing systematic equities team focused on execution research, market microstructure modeling, and the development of predictive tools that directly impact trading performance. This role is ideal for someone who enjoys working at the intersection of data science, engineering, and automated trading, and who wants to contribute to the evolution of a sophisticated global trading platform.</p><p>In this position, you will work closely with researchers, traders, and engineers to enhance models used in portfolio optimization, execution strategy, and cost forecasting. This is a highly collaborative environment where practical problem-solving, technical rigor, and creativity are valued. This role is hybrid in Boston, MA onsite 3-4 days a week.</p><p><br></p><p><strong>Responsibilities</strong></p><p>• Build, maintain, and enhance quantitative models for forecasting trading costs, slippage, and market impact</p><p>• Monitor and actively improve execution quality across large-scale global equity portfolios</p><p>• Conduct research on market microstructure dynamics using high-resolution datasets (tick data, order book data, etc.)</p><p>• Develop and backtest short-horizon predictive signals that support execution and alpha-related initiatives</p><p>• Design analytical tools and simulations used for portfolio optimization, strategy evaluation, and execution measurement</p><p>• Collaborate with engineering teams to deploy research into production within fully automated trading systems</p><p><br></p><p><strong>What This Role Offers</strong></p><p>• High-impact opportunity where research directly influences trading performance</p><p>• Exposure to both execution research and alpha-adjacent modeling</p><p>• Collaborative, engineering-driven environment with significant autonomy</p><p>• Ability to work with cutting-edge data, tools, and fully automated trading infrastructure</p><p>• Clear growth pathway within a rapidly scaling quantitative team</p>
<p>Robert Half is hiring! We are looking for an experienced Artificial Intelligence (AI) Engineer to join our team. In this role, you will design and implement cutting-edge AI and machine learning solutions to enhance our SaaS platform. You will collaborate with cross-functional teams to optimize workflows, improve customer experiences, and drive innovation through intelligent features.</p><p><br></p><p>Responsibilities:</p><p>• Develop and deploy robust machine learning models for predictive analytics, generative AI, and other advanced capabilities within a SaaS environment.</p><p>• Create scalable data pipelines for model training, testing, and monitoring, ensuring optimal performance and reliability.</p><p>• Collaborate with product, engineering, and data teams to identify and implement AI-driven solutions that address business challenges.</p><p>• Design and integrate AI functionalities, such as recommendations and classification systems, while maintaining efficiency and accuracy.</p><p>• Incorporate AI models into cloud-based systems using APIs, microservices, and containerized infrastructure.</p><p>• Assess and implement third-party AI tools and frameworks to enhance productivity and product capabilities.</p><p>• Ensure models align with privacy, security, and fairness standards, maintaining compliance across all implementations.</p><p>• Document workflows, track experiments, and maintain reproducibility for all AI-related processes.</p><p>• Keep up-to-date with advancements in AI technologies, machine learning techniques, and SaaS architecture trends.</p><ul><li>Python, Typescript, React </li></ul>
<p>Our Enterprise Data and Analytics team is growing. We’re looking for a Lead Data Scientist to assist with building and developing our Data Science team and lead us into the next generation of banking. We are reimagining how data is used across the bank to better serve our customers support our communities and make our colleagues lives better. Our goal is to be the best performing Regional Bank in America and we need data and analytics to meet that goal.</p><p>As we advance our data science and analytics capabilities we want a Lead Data Scientist to develop experts in modeling complex business problems and discovering business insights using statistical algorithmic mining and visualization techniques. We are looking for a leader who has a passion for developing others driving change and continuously improving and evolving the application of technologies to meet todays and tomorrow’s challenges.</p><p><br></p><p><strong>Key Responsibilities:</strong></p><p>· Prioritizes analytical projects based on business value and technological readiness</p><p>· Performs large-scale experimentation and build data-driven models to answer business questions</p><p>· Conducts research on cutting-edge techniques and tools in machine learning/deep learning/artificial intelligence</p><p>· Evangelizes best practices to analytics and products teams</p><p>· Acts as the go-to resource for machine learning across a range of business needs</p><p>· Owns the entire model development process, from identifying the business requirements, data sourcing, model fitting, presenting results, and production scoring</p><p>· Provides leadership, coaching, and mentoring to team members and develops the team to work with all areas of the organization</p><p>· Works with stakeholders to ensure that business needs are clearly understood and that services meet those needs</p><p>· Anticipates and analyzes trends in technology while assessing the emerging technologies impact(s)</p>
<p>We are seeking a talented and motivated Python Data Engineer to join our global team. In this role, you will be instrumental in expanding and optimizing our data assets to enhance analytical capabilities across the organization. You will collaborate closely with traders, analysts, researchers, and data scientists to gather requirements and deliver scalable data solutions that support critical business functions.</p><p><br></p><p>Responsibilities</p><ul><li>Develop modular and reusable Python components to connect external data sources with internal systems and databases.</li><li>Work directly with business stakeholders to translate analytical requirements into technical implementations.</li><li>Ensure the integrity and maintainability of the central Python codebase by adhering to existing design standards and best practices.</li><li>Maintain and improve the in-house Python ETL toolkit, contributing to the standardization and consolidation of data engineering workflows.</li><li>Partner with global team members to ensure efficient coordination and delivery.</li><li>Actively participate in internal Python development community and support ongoing business development initiatives with technical expertise.</li></ul>
We are looking for an experienced Lead Data Engineer to oversee the design, implementation, and management of advanced data infrastructure in Houston, Texas. This role requires expertise in architecting scalable solutions, optimizing data pipelines, and ensuring data quality to support analytics, machine learning, and real-time processing. The ideal candidate will have a deep understanding of Lakehouse architecture and Medallion design principles to deliver robust and governed data solutions.<br><br>Responsibilities:<br>• Develop and implement scalable data pipelines to ingest, process, and store large datasets using tools such as Apache Spark, Hadoop, and Kafka.<br>• Utilize cloud platforms like AWS or Azure to manage data storage and processing, leveraging services such as S3, Lambda, and Azure Data Lake.<br>• Design and operationalize data architecture following Medallion patterns to ensure data usability and quality across Bronze, Silver, and Gold layers.<br>• Build and optimize data models and storage solutions, including Databricks Lakehouses, to support analytical and operational needs.<br>• Automate data workflows using tools like Apache Airflow and Fivetran to streamline integration and improve efficiency.<br>• Lead initiatives to establish best practices in data management, facilitating knowledge sharing and collaboration across technical and business teams.<br>• Collaborate with data scientists to provide infrastructure and tools for complex analytical models, using programming languages like Python or R.<br>• Implement and enforce data governance policies, including encryption, masking, and access controls, within cloud environments.<br>• Monitor and troubleshoot data pipelines for performance issues, applying tuning techniques to enhance throughput and reliability.<br>• Stay updated with emerging technologies in data engineering and advocate for improvements to the organization's data systems.
<p><strong>Data Pipeline Development</strong></p><ul><li>Design, build, and optimize scalable ETL/ELT pipelines to support analytics and operational workflows.</li><li>Ingest structured, semi-structured, and unstructured data from multiple internal and external sources.</li><li>Automate and orchestrate data workflows using tools like Airflow, Azure Data Factory, AWS Glue, or similar.</li></ul><p><strong>Data Architecture & Modeling</strong></p><ul><li>Develop and maintain data models, data marts, and data warehouses (relational, dimensional, and/or cloud-native).</li><li>Implement best practices for data partitioning, performance optimization, and storage management.</li><li>Work with BI developers, data scientists, and analysts to ensure datasets are structured to meet business needs.</li></ul><p><strong>Cloud Engineering & Storage</strong></p><ul><li>Build and maintain cloud data environments (Azure, AWS, GCP), including storage, compute, and security components.</li><li>Deploy and manage scalable data systems such as Snowflake, Databricks, BigQuery, Redshift, or Synapse.</li><li>Optimize cloud data cost, performance, and governance.</li></ul><p><strong>Data Quality & Reliability</strong></p><ul><li>Implement data validation, error handling, and monitoring to ensure accuracy, completeness, and reliability.</li><li>Troubleshoot pipeline failures, performance issues, and data discrepancies.</li><li>Maintain documentation and data lineage for transparency and auditability.</li></ul><p><strong>Collaboration & Cross‑Functional Support</strong></p><ul><li>Partner with product, engineering, and analytics teams to translate business requirements into technical solutions.</li><li>Support self-service analytics initiatives by preparing high-quality datasets and data products.</li><li>Provide technical guidance on data best practices and engineering standards.</li></ul><p><br></p><p><br></p>