Job Details:
Primary address of work: Auckland
Vacancy number: 1
Employment type: Permanent full time
Minimum hours per week: 30 Hours
Maximum hours per week: 40 Hours
Minimum hourly rate (low salary range): $50/hour
Maximum hourly rate (high salary range): $55/hour
Pay frequent: Monthly
Company Overview
Auslink Group is a New Zealand-based logistics and supply chain service provider headquartered in Auckland. Auslink has been committed to delivering reliable, efficient, and customer- focused logistics solutions that support businesses across a
wide range of industries.
Key Responsibilities
1. Data Pipeline Development
- Design and maintain automated data pipelines to collect and move data from multiple sources (CRM, freight system, warehousing, etc.) to centralized repositories.
- Ensure efficient data flow between operational systems and analytics platforms.
2. Data Modeling and Architecture
- Develop logical and physical data models to support logistics operations and analytical applications.
- Design and maintain scalable data architecture, including data warehouses and data marts, to enable business intelligence and reporting.
3. Data Storage and Management
- Manage cloud and on-premise databases, ensuring efficient, secure, and reliable storage of structured and unstructured data.
4. Data Integration
- Integrate data across logistics functions (CRM, communication portals, warehousing, customs clearance, delivery, and mapping) into a unified system.
- Develop APIs and data connectors to support seamless system interoperability.
5. Data Quality and Governance
- Implement data quality frameworks to ensure completeness, consistency, and accuracy of business- critical data.
- Establish and enforce data governance policies aligned with privacy, regulatory, and operational standards.
6. Collaboration and Stakeholder Engagement
- Work closely with business analysts, data scientists, software engineers, and logistics teams to understand data requirements and deliver actionable solutions.
- Provide expert input on data architecture in cross- functional ICT and business transformation projects.
7. Performance Optimization
- Monitor and tune data systems to ensure optimal performance under varying loads.
- Anticipate scalability needs and proactively adapt infrastructure to support business growth.
8. Troubleshooting and Support
- Identify, diagnose, and resolve issues in data pipelines, ETL processes, and production databases.
- Respond to incidents and perform root cause analysis to prevent recurrence.
Job requirement: applicant must meet following requirement to apply for this job
Education and Experience
- Bachelor’s degree in Computer Science, Information Technology, or a related field; or
- Minimum 3 years’ experience in a data engineering, database specialist, or system architecture role—preferably in logistics, warehousing, or E-commerce sectors.
Technical Skills
- Proficiency in database management systems such as MySQL, PostgreSQL, SQL Server, or Oracle.
- Strong experience in ETL frameworks, data modeling, and warehousing technologies.
- Knowledge of cloud data platforms (e.g., AWS RDS, Redshift, Azure SQL) and data orchestration tools.
- Familiarity with Python, SQL, or similar scripting languages for data handling.
- Understanding of data governance, privacy compliance,
and security protocols.
To submit your application, click Apply Now!!!