label
stringclasses 12
values | text
stringlengths 2
1.22k
⌀ |
|---|---|
Skill
|
MS Office.
|
PI
|
Personal Information:
|
PI
|
Name: Ms. Shweta Mohan Landge
|
PI
|
Date of Birth: 3rd Oct, 1992
|
PI
|
Gender: Female
|
PI
|
Father's Name: Mr. Mohan P. Landge
|
PI
|
Father's Occupation: Civil Contractor
|
PI
|
Nationality: Indian
|
PI
|
Blood Group: AB +ve
|
PI
|
Languages Known: English, Marathi, Hindi
|
PI
|
Marital Status: Unmarried
|
PI
|
Email: landge.shweta92@gmail.com
|
PI
|
Mobile: +919038877058
|
PI
|
Address:
|
PI
|
C/o - Sr.no.-77, Jyotiba Nagar, Kalewadi,
|
PI
|
Pimpri, Pune-411017.
|
PI
|
Dist - Pune
|
PI
|
Pin - 411017
|
PI
|
Hobby:
|
PI
|
Drawing, Dancing, Listening to music, Chit chat, etc.
|
Exp
|
Professional Experience:
|
Exp
|
Working as a BI Developer in PowerupCloud Technologies', Bangalore currently.
|
Exp
|
Working as a Senior Software Engineer in Synchronous Technologies, Bangalore from Aug 2016 to Mar 2017.
|
Exp
|
Working as a Software Engineer in Silicon Techlab Pvt. Ltd, Bhubaneswar, from Nov 2011 to Mar 2016.
|
Exp
|
Project Details:
|
Exp
|
Project #1: Magic Software Analytics
|
Exp
|
Client: Magic Software
|
Exp
|
Duration: May'2018 to Present
|
Exp
|
Role: BI Developer
|
Exp
|
Environment: Amazon Redshift, Tableau
|
Exp
|
Responsibilities:
|
Exp
|
Working on "Most Engaging Books reports" and "Course progress report" created in Power BI Environment.
|
Exp
|
Developed Mappings and Reusable Transformations to facilitate timely Loading of Data of a star schema.
|
Exp
|
Extracted data from various sources, transformed data according to the requirement, and loaded into the data warehouse (Data Mart).
|
Exp
|
Created Data mart (Fact and Dimension Tables) as per the requirement and built OLAP reports on them.
|
Exp
|
Project #2: Volkswagen Analytics
|
Exp
|
Client: Volkswagen
|
Exp
|
Duration: Nov'2017 To May’2018
|
Exp
|
Role: BI Developer
|
Exp
|
Environment: Amazon Redshift, Power BI
|
Exp
|
Responsibilities:
|
Exp
|
Worked on "Customer Dashboard" to create a data mart which can help to create a single reporting platform covering all types of customer activities and to provide a better user experience.
|
Exp
|
Created a data mart in Amazon Redshift.
|
Exp
|
Involved in data profiling, data cleansing, and data enrichment.
|
Exp
|
Created reports in Power BI Environment.
|
Exp
|
Involved in Unit Testing.
|
Exp
|
Implemented various Performance Tuning techniques.
|
Exp
|
Project #3: JD Analytics
|
Exp
|
Client: 2AdPro
|
Exp
|
Duration: Jul'2017 To Jan’2018
|
Exp
|
Role: BI Developer
|
Exp
|
Environment: Amazon Redshift, Power BI
|
Exp
|
Responsibilities:
|
Exp
|
Worked on "JD Analytics" to create a data mart which can help to create a single reporting platform covering all types of digital commercial activity and to provide a better user experience.
|
Exp
|
Created a data mart in Amazon Redshift.
|
Exp
|
Involved in data profiling, data cleansing, and data enrichment.
|
Exp
|
Created reports in Power BI Environment.
|
Exp
|
Involved in Unit Testing.
|
Exp
|
Implemented various Performance Tuning techniques.
|
Exp
|
Project #4: FundsIndia Analytics POC
|
Exp
|
Client: FundsIndia
|
Exp
|
Duration: Jul'2017 To Nov'2017
|
Exp
|
Role: SQL Developer
|
Exp
|
Environment: Amazon Redshift, Power BI
|
Exp
|
Responsibilities:
|
Exp
|
Converted the existing SQL server queries into Redshift and do optimizations and check the performance.
|
Exp
|
Involved in data profiling, data cleansing, and data enrichment.
|
Exp
|
Involved in Unit Testing.
|
Exp
|
Implemented various Performance Tuning techniques.
|
Exp
|
Project #5: Network Analytics
|
Exp
|
Client: AT & T, VERIZON
|
Exp
|
Duration: Aug'2016 To Mar 2017
|
Exp
|
Role: Data Warehouse Engineer
|
Exp
|
Environment: Apache Drill, Spark SQL, JSON, Netezza
|
Exp
|
Responsibilities:
|
Exp
|
Worked on "Network Analytics" product to migrate into big data technologies to process the huge volume of unstructured data into the Data Lake.
|
Exp
|
Convert the queries into (Spark SQL)/(Apache Drill) who were previously written in Netezza/SQL SERVER and do optimization.
|
Exp
|
Create transformation/Loader using JSON/JEXL.
|
Exp
|
Involved in data profiling, data cleansing, and data enrichment.
|
Exp
|
Evaluated and tuned the SQL queries and physical database environments for efficient use.
|
Exp
|
Involved in Unit Testing.
|
Exp
|
Implemented various Performance Tuning techniques.
|
Exp
|
Project #6: HIPPO Analytics
|
Exp
|
Client: HIPPO CAMPUS, Bangalore
|
Exp
|
Duration: June'2014 To Mar '2016
|
Exp
|
Role: ETL Developer, BI Developer
|
Exp
|
Environment: Pentaho Data Integration, MySQL (Source), PostgreSQL (Target), BMBI (Reporting Tool)
|
Exp
|
Responsibilities:
|
Exp
|
Used Pentaho Data Integration for (ETL) extraction, transformation, and loading data from heterogeneous source systems into the target database.
|
Exp
|
Extracted data from various sources, transformed data according to the requirement.
|
Exp
|
Involved in extracting the data from the Flat Files and Relational databases into the staging area.
|
Exp
|
Mappings, Sessions, Workflows from Development to Test and then to UAT environment.
|
Exp
|
Developed Mappings and Reusable Transformations to facilitate timely Loading of Data of a star schema.
|
Exp
|
Extracted data from various sources, transformed data according to the requirement, and loaded into the data warehouse (Data Mart).
|
Exp
|
Created Data mart (Fact and Dimension Tables) as per the requirement and built OLAP reports on them.
|
Exp
|
Scheduling the Transformation and Jobs in UNIX environment using CronTab.
|
Exp
|
Prepared Technical Design documents and Test cases.
|
Exp
|
Involved in Unit Testing.
|
Exp
|
Implemented various Performance Tuning techniques.
|
Exp
|
Project #7: BPUT Education Analytics
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.