Overview
Hybrid
Depends on Experience
Contract - Independent
Contract - W2
Contract - 9 Month(s)
Skills
Principal Data Engineer (Pharmaceutical Domain)
Job Details
Role: Principal Data Engineer (Pharmaceutical Domain)
Location: Piscataway, New Jersey (Hybrid) We can submit non-local candidates
Duration: 9 to 12 months
Responsibilities:
Design, develop, and implement high-quality software solutions and high-performance data platforms that can handle large data sets with both batch and real time processing
Troubleshoot and address complex technical challenges and improve existing software to fit unique organization needs and ecosystem.
Provide leadership and consultation on technical matters to a variety of development functions.
Analyze business requirements and translate to technical requirements to develop proposals outlining how the organization s products and services can meet these needs and be integrated and implemented with the organization's technical infrastructure.
Design end-to-end technical solutions that are robust, secure, resilient, scalable and maintainable which align with business requirements, architectural best practices and standards to facilitate seamless data flow and communication between systems.
Innovation mindset, looking for opportunities to enhance & improve the GCO products & tools with new tech innovation, AI/ML, LLM/NLP, Automation and helping to embed them in products
Lead the definition of scope, delivery approach, resourcing plan and cost models to deliver new technology (system, data and services as a product) or amend existing technologies to deliver business needs and objectives.
Define technical roadmap and own delivery oversight for all in-scope systems working closely with Technology Product Owners and leadership teams.
Identify technical risks, vulnerabilities, and dependencies and develop mitigation strategies to address them. Proactively assess and manage risks throughout the software development lifecycle.
Lead the integration of disparate systems, platforms, and data sources to create unified and interoperable solutions.
Provide technical leadership, guidance, and mentorship to Technology development teams sourced from Colgate Palmolive's technology delivery partners including architects, engineers, testers and analysts. Collaborate with cross-functional teams to foster a culture of innovation, continuous improvement, and knowledge sharing.
Define and evolve the data architecture strategy, including data modeling, data storage, and data processing frameworks.
Optimize the performance of systems and applications through architectural design, performance tuning, and capacity planning. Identify bottlenecks, inefficiencies, and areas for improvement, and implement solutions to enhance system performance and responsiveness.
Apply software design principles to complex work in research, design and development of new or existing products, tools and processes required for operation, maintenance and testing
Liaise with hardware, software, and systems design engineers to ensure that products and services are modified, configured and installed.
Design, develop, and implement high-quality software solutions and high-performance data platforms that can handle large data sets with both batch and real time processing
Troubleshoot and address complex technical challenges and improve existing software to fit unique organization needs and ecosystem.
Provide leadership and consultation on technical matters to a variety of development functions.
Analyze business requirements and translate to technical requirements to develop proposals outlining how the organization s products and services can meet these needs and be integrated and implemented with the organization's technical infrastructure.
Design end-to-end technical solutions that are robust, secure, resilient, scalable and maintainable which align with business requirements, architectural best practices and standards to facilitate seamless data flow and communication between systems.
Innovation mindset, looking for opportunities to enhance & improve the GCO products & tools with new tech innovation, AI/ML, LLM/NLP, Automation and helping to embed them in products
Lead the definition of scope, delivery approach, resourcing plan and cost models to deliver new technology (system, data and services as a product) or amend existing technologies to deliver business needs and objectives.
Define technical roadmap and own delivery oversight for all in-scope systems working closely with Technology Product Owners and leadership teams.
Identify technical risks, vulnerabilities, and dependencies and develop mitigation strategies to address them. Proactively assess and manage risks throughout the software development lifecycle.
Lead the integration of disparate systems, platforms, and data sources to create unified and interoperable solutions.
Provide technical leadership, guidance, and mentorship to Technology development teams sourced from Colgate Palmolive's technology delivery partners including architects, engineers, testers and analysts. Collaborate with cross-functional teams to foster a culture of innovation, continuous improvement, and knowledge sharing.
Define and evolve the data architecture strategy, including data modeling, data storage, and data processing frameworks.
Optimize the performance of systems and applications through architectural design, performance tuning, and capacity planning. Identify bottlenecks, inefficiencies, and areas for improvement, and implement solutions to enhance system performance and responsiveness.
Apply software design principles to complex work in research, design and development of new or existing products, tools and processes required for operation, maintenance and testing
Liaise with hardware, software, and systems design engineers to ensure that products and services are modified, configured and installed.
Mandatory Skills:
Over 10 years of experience, with a minimum of 5 years in software development within large organizations.
Proven ability to create production-ready technical solutions using innovative technologies.
Experience with agile/scrum methodologies.
Strong expertise in architecting complex solutions on the Microsoft Azure cloud stack (IaaS, PaaS, and SaaS).
Hands-on experience in developing scalable, reliable, and high-performance data platforms to handle both large and small datasets along with expertise in implementing advanced AI/ML architectures, including Retrieval-Augmented Generation (RAG), Agentic-AI, Federated Learning, and Transformer-based models (e.g., OpenAI, Llama, BERT), using retrieval, embedding, and LLM techniques.
Development and Maintenance of MLOps pipelines for model versioning, monitoring, and retraining.
Experienced in vector search tools like Pinecone, Azure Cognitive Services, as well as AutoML, NASNet, GANs, and VAEs for generative tasks and representation learning.
Knowledge on ML/DL frameworks such as TensorFlow, PyTorch, Scikit-learn, and Hugging Face for data curation, predictive and prescriptive modelling.
Hands-On with integration and deployment of AI models using cloud services (AWS, Azure, Google Cloud Platform) in containerized environments (Docker, Kubernetes), building APIs, and microservices.
Optimization of models for accuracy, efficiency, and scalability.
Expertise in programming languages such as Python, PySpark, and strong SQL skills for data manipulation and performance tuning. Having knowledge on Java script tools is a plus.
Strong expertise in Azure services, including ADF, ADLS2, Azure Event Hub, Azure Functions, Databricks, Databricks Unity Catalog, version control (Git), CI/CD pipelines and data warehousing solutions like Snowflake/Azure Synapse.
Experience building batch and real-time data pipelines using tools such as Apache Kafka, Apache Spark, Flink, Airflow, or similar.
Strong understanding of architecture patterns including Kappa and Lambda patterns, with best practices for building cloud-ready systems.
Knowledge in data virtualization implementation using Denodo, with an understanding of Data Mesh and Data Fabric concepts.
Personal skills:
Able to appreciate short term vs. long term goals and take both tactical and strategic decisions.
Proven ability to create production-ready technical solutions using innovative technologies.
Experience with agile/scrum methodologies.
Strong expertise in architecting complex solutions on the Microsoft Azure cloud stack (IaaS, PaaS, and SaaS).
Hands-on experience in developing scalable, reliable, and high-performance data platforms to handle both large and small datasets along with expertise in implementing advanced AI/ML architectures, including Retrieval-Augmented Generation (RAG), Agentic-AI, Federated Learning, and Transformer-based models (e.g., OpenAI, Llama, BERT), using retrieval, embedding, and LLM techniques.
Development and Maintenance of MLOps pipelines for model versioning, monitoring, and retraining.
Experienced in vector search tools like Pinecone, Azure Cognitive Services, as well as AutoML, NASNet, GANs, and VAEs for generative tasks and representation learning.
Knowledge on ML/DL frameworks such as TensorFlow, PyTorch, Scikit-learn, and Hugging Face for data curation, predictive and prescriptive modelling.
Hands-On with integration and deployment of AI models using cloud services (AWS, Azure, Google Cloud Platform) in containerized environments (Docker, Kubernetes), building APIs, and microservices.
Optimization of models for accuracy, efficiency, and scalability.
Expertise in programming languages such as Python, PySpark, and strong SQL skills for data manipulation and performance tuning. Having knowledge on Java script tools is a plus.
Strong expertise in Azure services, including ADF, ADLS2, Azure Event Hub, Azure Functions, Databricks, Databricks Unity Catalog, version control (Git), CI/CD pipelines and data warehousing solutions like Snowflake/Azure Synapse.
Experience building batch and real-time data pipelines using tools such as Apache Kafka, Apache Spark, Flink, Airflow, or similar.
Strong understanding of architecture patterns including Kappa and Lambda patterns, with best practices for building cloud-ready systems.
Knowledge in data virtualization implementation using Denodo, with an understanding of Data Mesh and Data Fabric concepts.
Personal skills:
Able to appreciate short term vs. long term goals and take both tactical and strategic decisions.
Uday Raj Manager at Onwardpath 2701 Larsen Rd #BA142, Green Bay, WI 54303 Ph: +1 |
Certified WBE & MBE |
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.