146k INDIA https://mail.sattvahuman.com/home JOB RECRUITMENT SERVICES
by Moon_WALK - Monday March 24, 2025 at 11:30 AM
#1
146k INDIA https://mail.sattvahuman.com/home JOB RECRUITMENT SERVICES

Sattva Human Private Limited (Formerly Known as United Overseas HR Consultants)

DATA INCLUDE : email,first_name,last_name,contact_no,gender,dob,skills,cv_path,cv_text and more

TOTAL LINE :  146,744

PRICE : $400

ESCROW ACCEPTED ALWAYS!!

DM ONLY SERIOUS BUYER, TIME WASTER AND SCAMMER STAY OFF

DM ON SITE OR

34 LINES SAMPLE LINK : https://biteblob.com/Information/Iy5E73y...sample.csv

TOX CHAT : 3280C0ACAEB0905A8DABD12BA5AD6E718E09FCCCBB55C1AA8BE556DCF56A3C60FB615447E8B5

SAMPLE :
"email":"sasietldwh03@gmail.com","first_name":"Allam","last_name":"Sasikumar","contact_no":"9063841067","gender":null,"dob":null,"curr_company":null,"skills":"Data Modeling","curr_position":null,"pref_location":null,"source":"Naukri","exp":null,"ctc":null,"exp_ctc":null,"notice_period":null,"cv_path":"CV/Resume-Allam-9063841067-1673254488582.doc","cv_text":"  Allam Sasikumar\n  EmailConfusedasietldwh03@gmail.com\n  Contact Number: +91 - 9063841067\nPROFESSIONAL SUMMARY:\n Over 6 years of experience in all phases of Analysis, Design, Development, Data Modeling Implementation and support of Data Warehousing Applications using tools like Informatica, SnowFlake, AWS Oracle, SQL Server and DB2.\n Expertise in ETL operations using Informatica (Power Center 8/9/10)\n Proficient in Data Mart, Star Schema, Extract, Transform and Load the Data, Requirement Analysis, Design, Development, Testing. Documentation and Implementation of Business Applications.\n Expertise in developing the mappings, sessions and workflows in Informatica.\n Good knowledge in Project Management, Business Analysis, Business Modeling and Data Modeling.\n Hands-on experience with Snowflake utilities, SnowSQL, SnowPipe techniques\n Develop SQL queries in SnowSQL and optimize and fine tune queries\n Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source system which includes data load into snowflake table.\n Mentor and train junior team members and ensure coding standard is followed across the project.\n ETL pipelines in and out of data warehouse using combination Snowflakes SnowSQL Writing SQL queries against Snowflake.\n Expertise in prioritizing the tasks and decision making during the different phases of project\n Experience with Snowflake Multi - Cluster Warehouses.\n Define virtual warehouse sizing for Snowflake for different type of workloads.\n Experience in building Tasks, Streams, Materialized Views and Snowpipe.\n Experience in using Snowflake Clone and Time Travel.\n Expertise in identifying the key issues/Risk in the project and bringing them to closure\n Experienced in cloud-based snowflake to extract data from Amazon S3 and Azure for business intelligence\n Experienced in sprint planning and sprint retro activities in agile methodology as certified scrum master\n Experienced in handling multiple project and teams in the onsite-offshore model collaboratively\n Involved in doing the POC and convincing the clients to get the approvals to new projects\n Strong analytical, communication, leadership and problem solving skills\n Expertise in Project Management, Business Analysis, Business Modeling and Data Modeling\n Expertise in Performance tuning of ETL and Database routines to reduce the data load time\n Good knowledge in performance tuning in Informatica mapping level and session level\nQUALIFICATION:\n B.E ECE from Anna University in the year 2016.\nWORK EXPERIENCE:\n Working as a Senior Software Engineer in  Wipro, Bangalore from Jun2016 to till date.\nTECHNICAL SKILLS:\nETL Tools\n: Informatica Power Center 8/9/10, SAP BODS, SSIS\nOLAP Tools\n: Cognos\nData Modelling\n: Erwin4.2.2, Visio 2000, Power Designer.\nData Bases\n: Oracle/SQL Server/SnowFlake, AWS\nData Cleansing\n: Trillium 7.0.\nLanguages\n: SQL, T-SQL PL/SQL, Python\nOS\n: Windows/Unix\nOthers\n: IDQ\nProject #1\nPROJECT\n: Policy Management System\nCLIENT\n: American Insurance Group\n ROLE\n          : ETL Designer and Developer\nDescription:\nAmerica Insurance Group(AIG) is a U.S. insurer group of automobiles, homes and small businesses and also provides other insurance and financial services products. Developed Modules for Finance & Accounts, Re-insurance, Production  Premium and Claims; Commission - current, unsettled; Payouts; Collections.\nResponsibilities:\n Involved in reverse engineering of existing system to document the product functionality.\n Gathered new requirements and involved in preparation of ETL High level design documents.\n Created mappings as per business requirements using Informatica Power Center to move Data from Sybase, Flat files sources to Staging, and then to master Data Warehouse in Oracle DB.\n Used most of the transformations such as the Source Qualifier, Expression, Aggregator, connected & unconnected lookups, Filter, Router, Sequence Generator, Sorter, joiner, and Update Strategy.\n Developed SCD type2 with dynamic lookup cache.\n Involved in preparing designs, unit test cases and in code review activities.\n Designed Optimized mappings to get better performance.\n Developing the Informatica Mappings, sessions and workflows required to load the data into oracle application from flat files and staging tables\n Involved in development of UNIX scripts necessary in ETL processes\n Writing the test cases for unit and integration testing\n Expertise in unit test, integration test and system testing of Informatica components\n Interacting with developers and explaining the mapping rules from legacy to bridges application\n Analyzing the problems and providing optimum solutions to customers.\n Interacting with customers and validating existing process understanding and proposing enhancements in the process.\n Expertise in creating data models for the Bridges using Power designer.\n      Environment: Informatica PowerCenter Designer 10.1.1, Oracle 9i, SQL, PL/SQL, Unix, Flat files, DB2, SnowFlake, AWS.\nProject #2\nPROJECT\n: Best Buy\nCLIENT\n: U.S. Foodservice Inc.\nROLE\n: ETL Developer and Snowflake\nDescription:\nBest Buy Co. Inc. is an American multinational consumer electronics retailer headquartered in Richfield, Minnesota. Originally founded by Richard M. Schulze and James Wheeler in 1966 as an audio specialty store called Sound of Music, it was rebranded under its current name with an emphasis on consumer electronics in 1983. Customer want to track all the products, sales, invoices and orders information in the Dataware house for business intelligence\nResponsibilities:\n Gathering the requirements by interacting with users and documenting the requirements.\n Design the technical design specifications and program specification documents for ETL jobs\n Developing the Informatica Mappings, sessions and workflows required to load the data into oracle application from flat files and staging tables\n Worked on SnowSQL and Snow pipe\n Converted Talend Joblets to support the snowflake functionality.\n Creating Transformer Pipelines and Data Collector Pipelines\n Test, Validate and Deploy the Pipelines\n Used COPY to bulk load the data.\n Created data sharing between two snowflake accounts.\n Analyzing the problems and providing optimum solutions to customers.\n Interacting with customers and validating existing process understanding and proposing    enhancements in the process.\n Worked with DBAs in creating the indexes to Database to improve the performance of exiting    ETL and newly proposed ETL processes.\nEnvironment: Informatica PowerCenter Designer 10.1.1, Oracle 9i, SQL, PL/SQL, Unix, Flat files, DB2, SnowFlake, AWS\nProject #3\nPROJECT\n              : Sales Analysis\nCLIENT\n  : Nissan\nROLE\n  : ETL Developer\nDescription:  Nissan was theleading Japanese brand in China, Russia and Mexico. In 2014, Nissan was the largest car manufacturer in North America. As of April 2018, Nissan was the world's largest electric vehicle (EV) manufacturer, with global sales of more than 320,000 all-electric vehicles.\nResponsibilities:\n Gathering the requirements by interacting with business\n Analyzing the problems and providing optimum solutions to customers.\n Involved in conceptual, Logical and Physical data modelling using power designer.\n Worked with DBAs for creating Physical Model by providing DDL.\n Interacting with customers and validating existing process understanding and proposing enhancements in the process.\n Converted Talend Joblets to support the snowflake functionality.\n Creating Transformer Pipelines and Data Collector Pipelines\n Test, Validate and Deploy the Pipelines.\n Designing the technical design specifications and program specification documents for ETL job\n      Environment: Informatica PowerCenter Designer 10.1.1, Oracle 9i, SQL, PL/SQL, Unix, Flat files,\n        DB2, SnowFlake, AWS\n[image: image1.png]\n[image: image2.png]","qualification":null,"remarks":null,"created_by":"[{\"title\" : \"Priyanka Gaikwad\", \"emp_id\": \"SH-040\"}]","updated_by":"[{\"title\" : \"Priyanka Gaikwad\", \"emp_id\": \"SH-040\"}]","createdAt":"09-Jan-2023","can_update_dt":"2023-01-09T08:54:48.665Z","relevant_exp":null,"viewed":null,"id":22310,"cid":22310,"candidateStatus":"UploadedBySourcer","cj_source":"Naukri","cj_added_by":"[{\"title\" : \"Priyanka Gaikwad\", \"emp_id\": \"SH-040\"}]","cj_updated_by":"[{\"title\" : \"Priyanka Gaikwad\", \"emp_id\": \"SH-040\"}]","cj_dt":"2023-01-09T08:54:48.815Z","cj_update_dt":"2023-01-09T08:54:48.815Z","job_id":675,"reason":null,"comment":null,"revenue":null,"selection_date":null,"offered_date":null,"joining_date":null,"duplicate_date":null,"revenue_confirmed":null,"sourcer":"SH-040","recruiter":null,"mgr_id":null,"p_mgr_id":null,"offered_ctc":null,"drop_reason":null,"jt_status":"UploadedBySourcer","jt_dt":"2023-01-09T08:54:49.161Z","tentative_exp":null,"max_ctc":null,"job_ids":"675"}}
"email":"himanshi23aug@gmail.com","first_name":"Himanshi","last_name":"K","contact_no":"8448773557","gender":null,"dob":null,"curr_company":null,"skills":"DATA MODELING","curr_position":null,"pref_location":null,"source":"Naukri","exp":null,"ctc":null,"exp_ctc":null,"notice_period":null,"cv_path":"CV/Resume-Himanshi-8448773557-1673254515627.pdf","cv_text":"\nContact\nSkills\nWork History\nEducation\nAccomplishments\nApplication Development Senior Analyst\nHimanshi\nCommitted Application Development Senior Analyst with 4 years of experience in Data Analyst, Data Modeling,\nData Migration and Data Quality.\nAddress\nNoida, India, 201305\nPhone\n844-877-3557\nE-mail\nhimanshi23aug@gmail.com\nSQL Server,\nT-SQL\nExcellent\nRelational\nDatabase\nExcellent\nData Modeling\nVery Good\nMicrosoft\nVisual Studio\nExcellent\nErwin Data\nModeler\nApplication Development Senior Analyst\nAccenture Solutions Pvt. Ltd, Noida, India\nCreated Schema, Database, Tables, Stored\nProcedures, Views, Function, complex SQL queries &\nMerge scripts which helps in loading data from\nJSONs on multiple environments for MyIndustryX\nproject.\nCreated logical and physical models to design OLTP\nsystem for the Industry Data Model and reports\nusing Erwin tool.\nImplementing best coding practices and created\ndocuments for development.\nManaging production environment.\nSoftware Engineer\nCPA Global, Noida, UP\nManaged over 50 major projects including American\nExpress TRS, McAfree in Data Migration.\nProficient in various kind of source data to migrate.\nBachelor of Technology: Computer Science\nGalgotias University - Greater Noida, UP\nGPA: 8.06\nAccenture Extra Mile Award (Individual category)\nThis award recognized me for staying agile & resilient\nand successfully prioritizing the task to work\nefficiently.\nVery Good\nSQL Server\nManagement\nStudio Excellent\nMicrosoft\nExcel\nVery Good\n2021-05 -\nCurrent\n2018-09 -\n2021-05\n2014-09 -\n2018-04","qualification":null,"remarks":null,"created_by":"[{\"title\" : \"Neha Yadav\", \"emp_id\": \"SH-015\"}]","updated_by":"[{\"title\" : \"Neha Yadav\", \"emp_id\": \"SH-015\"}]","createdAt":"09-Jan-2023","can_update_dt":"2023-01-09T08:55:15.723Z","relevant_exp":null,"viewed":null,"id":22311,"cid":22311,"candidateStatus":"Not Interested","cj_source":"Naukri","cj_added_by":"[{\"title\" : \"Neha Yadav\", \"emp_id\": \"SH-015\"}]","cj_updated_by":"[{\"title\" : \"Madhanaaz Shaikh\", \"emp_id\": \"SH-024\"}]","cj_dt":"2023-01-09T08:55:15.771Z","cj_update_dt":"2023-01-09T09:05:36.325Z","job_id":675,"reason":"Not Looking for Change","comment":"Not interested ","revenue":null,"selection_date":null,"offered_date":null,"joining_date":null,"duplicate_date":null,"revenue_confirmed":null,"sourcer":"SH-015","recruiter":"SH-024","mgr_id":"SH-015","p_mgr_id":"SH-003","offered_ctc":null,"drop_reason":null,"jt_status":"Not Interested","jt_dt":"2023-01-09T09:05:36.354Z","tentative_exp":null,"max_ctc":null,"job_ids":"675"}}
Reply
#2
bump up, data is still available
Reply


Possibly Related Threads…
Thread Author Replies Views Last Post
  SELLING +263K Mangolia Media Subscription Services , Like Netflix r57 0 1,136 2 hours ago
Last Post: r57
  SELLING India Money Loans Releated Database r57 7 1,658 2 hours ago
Last Post: r57
  SELLING 3.1M https://www.deuse.be/en/ USER DATA Moon_WALK 1 243 7 hours ago
Last Post: Moon_WALK
  SELLING 11K VIETNAM https://gopify.tech/ Moon_WALK 1 165 7 hours ago
Last Post: Moon_WALK
  SELLING 27k VIETNAM https://bkademy.vn/ Teacher & student data Moon_WALK 1 172 7 hours ago
Last Post: Moon_WALK

Forum Jump:


 Users browsing this thread: 1 Guest(s)