Job ID: VA-754979 (98490113)
Hybrid/Local Data Analyst (12+/AWS) with ELT/ETL, SQL, Snowflake, and data warehousing/migration experience
Location: Richmond, VA (DBHDS)
Duration: 12 Months
ON SITE REQUIRED: 2-3 days
No Parking provided for contractors
Skills:
Proficiency in data extraction and transformation tools Required 6 Years
Strong SQL querying and scripting skills Required 7 Years
Knowledge of, and Experience with ELT/ETL processes Required 6 Years
Familiarity with data modeling and database design principles Required 7 Years
Understanding of data warehousing concepts and practices Required 7 Years
Experience with data migration in large-scale enterprise environments Required 7 Years
Understanding of data synchronization and replication techniques Required 7 Years
Strong analytical and problem-solving abilities Required 6 Years
Critical thinking and decision-making skills Required 7 Years
Ability to handle multiple tasks and prioritize effectively Required 6 Years
Expertise in AWS Desired 2 Years
Expertise in Snowflake Desired 2 Years
Description:
A Data Migration Specialist is a professional who specializes in the movement of data from one system or platform to another.
Pls note: Initial web-based interviews will be conducted; however, the team may request a second round IN PERSON with the finalists.
ABOUT THE ROLE
Main Responsibilities of a Data Migration Specialist
The following list describes the typical responsibilities and deliverables of a Data Migration Specialist:
Transform and load data into the respective data warehouse for a modernization initiative.
Transform data to meet target system requirements
Translate functional requirements into technical specifications for ETL development.
Troubleshoot and resolve data migration issues
Validate and analyze data sets to ensure data consistency and internal requirements are met.
Analyze DBDHS Legacy data, data quality rules, business rules, data dictionaries,
Analyze, prepare and execute data cleansing services.
Analyze, specify and document data migration requirements with clients and internal teams.
Articulate how Legacy data will be cleansed and adapted to fit into the new Future solution.
Assist others with the selection of approaches, methods and / or services as required.
Attend Project stand-ups and Stakeholder meetings
Champion standards and implementation process that you define with the team.
Collaborate with data architects to design data migration solutions
Collaborate with the IMS Business customers and the DBHDS EDW Team, ensuring quicker and more accurate data management.
Complete assigned tasks in an organized manner to ensure project deadlines.
Develop and execute data migration process and procedures (from analysis to post implementation) including scripts and programs
Develop and execute test plans to ensure client requirements are fulfilled.
Contribute to the release and cutover strategy.
Coordinate with stakeholders to gather requirements and system assessments as it relates to data.
Create and maintain utilities, SQL scripts as needed to accelerate the process.
Ensure adherence to compliance needs for the agency.
Implement new Mapping and ETL scripts.
Lead data cleaning and dedupe program to drive accurate and complete data.
Maintain appropriate levels of data security and privacy relating to customer data.
Ensure documentation is created and maintained to enable migration process and post implementation support.
Create and Monitor data migration progress and build dashboards to support audit , reconciliation and verification
Perform regular backups to ensure data preservation.
Propose a data migration and reconciliation strategy.
Represent DBHDS IT and the Data cleansing/migration initiative at meetups with DBHDS internal and external partners
Retrieve data from the database or electronic files as requested.
Validate post-migration results and communicate outcome with Key DBHDS Central Office Leadership and Business Stakeholders, Key External DBHDS Reporting Stakeholders, CO Security, VITA, and DBHDS IT.
Verify data by comparing it to source documents.
Write validation scripts to run over the Staging tables to report any invalid data.