Overview
On Site
$$60 - $63/hr. on C2C
Accepts corp to corp applications
Contract - Independent
Contract - W2
Contract - 12 month
Skills
Python
Splunk
Node
Job Details
Role: Python Backend Engineer
Location: Snoqualmie, WA (Hybrid)
- Data Ingestion
- API Development
- Backend Engineering
- Python, Node.js, or Go
We are seeking a strong Backend Engineer to design and implement robust data pipelines that ingest logs, metrics, and telemetry data from observability tools such as Splunk, InfluxDB, and OpenSearch. The data will be processed, normalized, and stored in appropriate backend data stores (SQL/NoSQL) for downstream usage. The engineer will also develop performant and secure RESTful APIs to support frontend visualizations and dashboards.
Key Responsibilities:- Design and implement data ingestion pipelines to pull data from:
- Splunk (via REST API or SDKs)
- InfluxDB (using Flux/InfluxQL)
- OpenSearch (via query DSL or API)
- Normalize, transform, and insert collected data into backend systems such as:
- PostgreSQL / MySQL
- MongoDB / DynamoDB / TimescaleDB (optional)
- Build RESTful APIs to expose processed data for:
- Dashboards
- Alerts/Health indicators
- Metrics visualizations
- Implement data retention and archival logic for compliance and performance
- Collaborate with DevOps to integrate pipelines into CI/CD and containerized environments (Docker/Kubernetes)
- Implement basic observability (logs, metrics, alerts) for APIs and pipelines
- Work closely with frontend developers and business analysts to define data contracts and endpoint requirements
- 7+ years of backend development experience with Python, Node.js, or Go
- Hands-on experience with API frameworks (e.g., FastAPI, Flask, Express, Gin)
- Integration experience with Splunk, InfluxDB, and/or OpenSearch
- Strong knowledge of query languages:
- SPL (Splunk)
- Flux or InfluxQL (InfluxDB)
- Elasticsearch DSL (OpenSearch)
- Proficiency in SQL and data modeling
- Experience with JSON, REST, OAuth, JWT, and API security best practices
- Experience building services that process high-velocity telemetry or monitoring data
- Solid understanding of asynchronous processing (e.g., Celery, Kafka)
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.