Senior Manager Data Engineering

Número de la requisición: 2282737
Categoría de la vacante: Technology
Localização da vaga: Noida, Uttar Pradesh

Man standing and writing on a white board while presenting to coworkers in a meeting room.

Optum is a global organization that delivers care, aided by technology to help millions of people live healthier lives. The work you do with our team will directly improve health outcomes by connecting people with the care, pharmacy benefits, data and resources they need to feel their best. Here, you will find a culture guided by inclusion, talented peers, comprehensive benefits and career development opportunities. Come make an impact on the communities we serve as you help us advance health optimization on a global scale. Join us to start Caring. Connecting. Growing together. 

Primary Responsibilities:

  • Participate in scrum process and deliver stories/features according to the schedule
  • Collaborate with team, architects and product stakeholders to understand the scope and design of a deliverable
  • Participate in product support activities as needed by the team
  • Understand product architecture, features being built and come up with product improvement ideas and POCs
  • Be able to learn and adpat to new data technologies
  • Help design and develop the next generation of NLP, ML & AI products, and services for healthcare
  • Develop machine learning and deep learning models and systems in domains including, but not limited to: NLP, NLU, NLG, SLU and multidimensional time series forecasting among others
  • Be able to Quantize, Optimize GenAI models
  • Manage NLP & ML models lifecycle for a suite of products
  • Run large complex proof-of-concepts for the healthcare business
  • Manage prioritization and technology work for building NLP, ML & AI solutions
  • Lead the full end-to-end machine learning development process including data ingestion and preparation, feature engineering, analysis and modeling, model deployment, performance tracking and documentation
  • Establish best practices for end-to-end deep learning and machine learning development cycle to ensure rigor in process and quality in outcome
  • Work with a great deal of autonomy to find solutions to complex problems
  • Comply with the terms and conditions of the employment contract, company policies and procedures, and any and all directives (such as, but not limited to, transfer and/or re-assignment to different work locations, change in teams and/or work shifts, policies in regards to flexibility of work benefits and/or work environment, alternative work arrangements, and other decisions that may arise due to the changing business environment). The Company may adopt, vary or rescind these policies and directives in its absolute discretion and without any limitation (implied or otherwise) on its ability to do so

Required Qualifications:

  • Undergraduate degree in applicable area of expertise or equivalent experience
  • Experience:
    • DevOps, implementation of Bigdata, Apache Spark and Azure Cloud
    • Large scale data processing using PySpark on azure ecosystem
    • Implementation of self-service analytics platform ETL framework using PySpark on Azure
    • Experience in deploying scalable solutions to complex problems, from defining the problem, implementing the solution, and launching the new product successfully
    • Skill Set; NLP, NLU, NLI
    • Architecture: Transformers, Attention
    • Models: GPT, Llama, Mistral
    • Model Quantization
    • Model Optimization
    • Retrieval & Ranking, RAG, RAGAS
    • Statistics, Machine Learning Models, Model Deployment
    • Experience in the health care industry 
    • Proven excellent communication, writing and presentation skills
  • Tools/Technologies:
    • Knowledge of Programming Languages: Python, PySpark
    • Knowledge of Cloud Technologies: Azure (ADF, Databricks, WebApp, Key vault, SQL Server, function app, logic app, Synapse, Azure Machine Learning, DevOps)
  • Hands-on data migration experience from legacy systems to new solutions, such as from on-premises clusters to Cloud
  • Hands-on experience with a large-scale data warehouse
  • Deep experience in Data Analysis, including source data analysis, data profiling and mapping
  • Good experience in building data pipelines using ADF/Azure Databricks
  • Exposure in RAG, LangChain, VectorDBs 
  • Expert skills in Azure data processing tools (Azure Data Factory, Azure Databricks)
  • Solid proficiency in SQL and complex queries
  • Proven good problem-solving skills
  • Proven good communication skills

Preferred Qualifications:

  • Knowledge/Experience in Azure Synapse and Power BI
  • Knowledge of US healthcare industry/Pharmacy data

At UnitedHealth Group, our mission is to help people live healthier lives and make the health system work better for everyone. We believe everyone-of every race, gender, sexuality, age, location and income-deserves the opportunity to live their healthiest life. Today, however, there are still far too many barriers to good health which are disproportionately experienced by people of color, historically marginalized groups and those with lower incomes. We are committed to mitigating our impact on the environment and enabling and delivering equitable care that addresses health disparities and improves health outcomes – an enterprise priority reflected in our mission.

Información adicional sobre la vacante

Número de la requisición 2282737

Segmento de negocio Optum

Disponibilidad para viajar No

País IN

Estado de horas extras Exempt

Vacante de teletrabajo No