Overview
Skills
Job Details
Position: Data Engineer/Devops/Big data -Open Position-CA
Type: contract
Duration: 12+ months
Location: Pleasanton, CA
Job Description:
We are really looking for someone that has experience building an analytics platform and some light work with showing people how to use it. The candidate would need to know enough to get users and how to do simple submits maybe and someone who is an operational visionary in this space of analytics / bigdata that can help develop or implement new tools to better run Kubernetes platforms, storage, GPU usage, parallel processing code (CUDA) etc.
Required Skills:
Spark
Hadoop/CDH
H2O/Steam
MapR
Kubernetes
Docker
Tensorflow
Apache Airflow
Apache Toree
Jupyterhub
Rstudio
PyTorch
ELK
OpenVino
MySql
GitLab
Traefik
Prometheus, Grafana, Node Manager, Alert Manager
Vault
Machine learning methods and data mining, Including the use of 2 or more novel/emerging methods in supervised learning (Eg: Standard regression, shrinkage methods, subset selection, discriminant analysis, tree-based methods, boosting, random forests, neural networks) Unsupervised learning (clustering, Self-organizing maps) and ensemble modeling Distributed computing methodology both in terms of HW (working across multiple servers) as well as SW tools,(Hadoop, H20 etc) In-memory analytic Packages (R, Python, Scala) and Visualization (Tableau)
3 yrs of experience on Projects, programs, requiring the integration of cross-functional technology and business solutions.
Notes: Client needs someone who can bring the ecosystem together
Currently client has on prem- environment
The want someone to bring the data environment ,containerization with Kubernentes with backend and
hook all the components together like R, python, pytorch, spark, Mapr, Big data
Please do reply to me with your updated resume with the hourly rates ASAP