Hybrid Data Engineer
Talascend is currently seeking a hybrid Data Engineer for a contract opportunity with our client in Glendale, Arizona.
Overview
The role involves designing, developing, and deploying industry-leading data science and big data engineering solutions, using artificial intelligence, machine learning, and big data platforms and technologies to increase efficiency in complex work processes and enable data-driven decision making.
Hybrid Shift
3 days in office each week required
Responsibilities
- Big data design and analysis, data modeling, development, deployment, and operations of big data pipelines
- Collaborate with a team of other data engineers, data scientists, and business subject matter experts to process data and prepare data sources for a variety of use cases including predictive analytics, generative AI, and computer vision
- Mentor other data engineers to develop a world class data engineering team
- Ingest, process, and model data from structured, unstructured, batch and real-time sources using the latest techniques and technology stack
Qualifications
- Bachelor’s degree or higher in Computer Science, or equivalent degree and 5+ years working experience
- In depth experience with a big data cloud platform such as Azure, AWS, Snowflake, Palantir, etc.
- Strong grasp of programming languages (Python, Scala, SQL, Panda, PySpark, or equivalent) and a willingness to learn new ones
- Strong understanding of structuring code for testability
- Experience writing database-heavy services or APIs
- Strong hands-on experience building and optimizing scalable data pipelines, complex transformations, architecture, and data sets with Databricks or Spark, Azure Data Factory, and/or Palantir Foundry for data ingestion and processing
- Proficient in distributed computing frameworks, with familiarity in handling drivers, executors, and data partitions in Hadoop or Spark
- Working knowledge of queueing, stream processing, and highly scalable data stores such as Hadoop, Delta Lake, Azure Data Lake Storage (ADLS), etc.
- Deep understanding of data governance, access control, and secure view implementation
- Experience in workflow orchestration and monitoring
- Experience working with and supporting cross-functional teams
Preferred Requirements
- Experience with schema evolution, data versioning, and Delta Lake optimization
- Exposure to data cataloging solutions in Foundry Ontology
- Professional experience implementing complex ML architectures in popular frameworks such as Tensorflow, Keras, PyTorch, Sci-kit Learn, and CNTK
- Professional experience implementing and maintaining MLOps pipelines in MLflow or AzureML
We thank all applicants for their interest. However, only those qualified individuals who closely meet the qualifications of the position will be contacted. The details of the position are only a summary, other duties may be assigned as necessary.
Background Check and Drug Screen may be required.
Talascend is an Equal Opportunity Employer that recruits and hires qualified candidates without regard to race, religion, sex, sexual orientation, gender identity, age, national origin, ancestry, citizenship, disability, or veteran status.