Data Integration Specialist – Professional Services

Offer by SmartDrive Systems

sql

apache-spark

continuous-integration

About this job

Job type: Full-time



Technologies

sql, apache-spark, continuous-integration



Job description

Description:

The Data Integration Specialist will be responsible for (a) the integration of external data sources with our published APIs (e.g.: driver ID, driver master, vehicle master, vehicle-to-recorder assignment), and (b) the export of our analytical data to customers’s systems. The job is focused on addressing data-needs, addressing challenges associated with gaining accurate, efficient and effective use and access to transportation intelligence data systems from and to our customers, partners and third party vendors. This is an exciting and rewarding role requiring a smart and expedient approach to data understanding, data integration and blending to help solve Transportation Intelligence challenges (driver safety, driver behavior, fuel performance, and operational efficiencies).

The Data Integration Specialist works as a crucial team member within the Professional Services organization in the San Diego, CA office. The right candidate must be a detail-oriented, self-reliant contributor who can master and coordinate processes across a broad range of internal and external stakeholders.

Having worked in internal-and-customer-facing roles, you have a track record in educating and interacting with a diverse set of stakeholders. You also have proven expertise and confidence in the creation and execution of robust data integrations, transformations, and blending of data from large, diverse operational data sets and external data feeds into analytics-ready results.

The position requires a real passion for data and analytics, a problem-solving attitude, a zeal for continuous learning, proven skills in data manipulation and a mindset for customer success.

Responsibilities:
    Engag
    • e directly with customers and data providers to participate in design -- and to own development – of data integrations according to customer requirements
    • Facilitate discovery and analysis sessions to understand data and functional requirements
    • Deliver end-to-end integrations that fit our analytics platform or fit export/integration with customer systems.
    • Support and assist customers’ and partners’ technical teams on two-way integrations.
    • Interpret and document requirements, designs, data and stakeholder feedback.
    • Schedule, maintain and monitor the execution of data integration pipelines
    • Provide technical insight and develop estimates for early project planning and scope of work
    • Participate in creation of documentation, technical procedures and user guides.
Minimum Qualifications & Competencies

If you have passion for -- and a record of success with past responsibilities to frequently deliver -- data integrations and an interest in analytics, we would love to talk with you.

    Exper
    • tise loading and querying relational and non-relational databases with a fluency in writing complex queries and transformations.
    • Efficient manipulaton of data at the scale of multi-billion records and beyond
    • Comfort using existing API’s to extract or load data.
    • Appreciation of the differences in loading or querying data models in relational, dimensional and NoSQL data systems.
    • Operating Systems: Windows, Linux, (AS/400 a plus, not required)
    • Database: Extract and transform from SQL Server, PostGres, MS Access, Excel ,
    • Big Data: (good-to-have loading and) must-have experience querying the following
    • Tools: AWS, EMR, Hive (DDL), Presto (SQL Query Interface), S3 (Storage), good-to-have some Spark (Compute)
    • Data Formats: Avro, Parquet, JSON,
    • Programming Languages SQL, Regex, Windows command line
    • Data cleansing and testing
    • Alteryx skill is very helpful.
    • Extraction, packaging and grouping of flat files, CSV, etc.;
    • Experience with  S3, Google Drive, DropBox interfaces
The following skills/experience are useful but not essential:-
    Abili
    • ty to build custom connections and Web Services API’s to interact with data across diverse technologies, including Amazon EC2, Kafka, Cassandra.
    • Big Data: Experience building/loading in Amazon ecosystem: EMR , Hive, S3, Spark and Presto
    • Python programming
    • SHP Shape Files for spatial analytics
    • For customer data entry, ability to build data-entry web-form and back-end pipeline for data delivery.
    • Experience in transportation, telematics, and datasets: FMCSA/DOT, US Census, traffic, WX
    • Tableau skill is useful but not required.


A new version is available REFRESH