label stringclasses 12
values | text stringlengths 2 1.22k ⌀ |
|---|---|
Skill | MS Office. |
PI | Personal Information: |
PI | Name: Ms. Shweta Mohan Landge |
PI | Date of Birth: 3rd Oct, 1992 |
PI | Gender: Female |
PI | Father's Name: Mr. Mohan P. Landge |
PI | Father's Occupation: Civil Contractor |
PI | Nationality: Indian |
PI | Blood Group: AB +ve |
PI | Languages Known: English, Marathi, Hindi |
PI | Marital Status: Unmarried |
PI | Email: landge.shweta92@gmail.com |
PI | Mobile: +919038877058 |
PI | Address: |
PI | C/o - Sr.no.-77, Jyotiba Nagar, Kalewadi, |
PI | Pimpri, Pune-411017. |
PI | Dist - Pune |
PI | Pin - 411017 |
PI | Hobby: |
PI | Drawing, Dancing, Listening to music, Chit chat, etc. |
Exp | Professional Experience: |
Exp | Working as a BI Developer in PowerupCloud Technologies', Bangalore currently. |
Exp | Working as a Senior Software Engineer in Synchronous Technologies, Bangalore from Aug 2016 to Mar 2017. |
Exp | Working as a Software Engineer in Silicon Techlab Pvt. Ltd, Bhubaneswar, from Nov 2011 to Mar 2016. |
Exp | Project Details: |
Exp | Project #1: Magic Software Analytics |
Exp | Client: Magic Software |
Exp | Duration: May'2018 to Present |
Exp | Role: BI Developer |
Exp | Environment: Amazon Redshift, Tableau |
Exp | Responsibilities: |
Exp | Working on "Most Engaging Books reports" and "Course progress report" created in Power BI Environment. |
Exp | Developed Mappings and Reusable Transformations to facilitate timely Loading of Data of a star schema. |
Exp | Extracted data from various sources, transformed data according to the requirement, and loaded into the data warehouse (Data Mart). |
Exp | Created Data mart (Fact and Dimension Tables) as per the requirement and built OLAP reports on them. |
Exp | Project #2: Volkswagen Analytics |
Exp | Client: Volkswagen |
Exp | Duration: Nov'2017 To May’2018 |
Exp | Role: BI Developer |
Exp | Environment: Amazon Redshift, Power BI |
Exp | Responsibilities: |
Exp | Worked on "Customer Dashboard" to create a data mart which can help to create a single reporting platform covering all types of customer activities and to provide a better user experience. |
Exp | Created a data mart in Amazon Redshift. |
Exp | Involved in data profiling, data cleansing, and data enrichment. |
Exp | Created reports in Power BI Environment. |
Exp | Involved in Unit Testing. |
Exp | Implemented various Performance Tuning techniques. |
Exp | Project #3: JD Analytics |
Exp | Client: 2AdPro |
Exp | Duration: Jul'2017 To Jan’2018 |
Exp | Role: BI Developer |
Exp | Environment: Amazon Redshift, Power BI |
Exp | Responsibilities: |
Exp | Worked on "JD Analytics" to create a data mart which can help to create a single reporting platform covering all types of digital commercial activity and to provide a better user experience. |
Exp | Created a data mart in Amazon Redshift. |
Exp | Involved in data profiling, data cleansing, and data enrichment. |
Exp | Created reports in Power BI Environment. |
Exp | Involved in Unit Testing. |
Exp | Implemented various Performance Tuning techniques. |
Exp | Project #4: FundsIndia Analytics POC |
Exp | Client: FundsIndia |
Exp | Duration: Jul'2017 To Nov'2017 |
Exp | Role: SQL Developer |
Exp | Environment: Amazon Redshift, Power BI |
Exp | Responsibilities: |
Exp | Converted the existing SQL server queries into Redshift and do optimizations and check the performance. |
Exp | Involved in data profiling, data cleansing, and data enrichment. |
Exp | Involved in Unit Testing. |
Exp | Implemented various Performance Tuning techniques. |
Exp | Project #5: Network Analytics |
Exp | Client: AT & T, VERIZON |
Exp | Duration: Aug'2016 To Mar 2017 |
Exp | Role: Data Warehouse Engineer |
Exp | Environment: Apache Drill, Spark SQL, JSON, Netezza |
Exp | Responsibilities: |
Exp | Worked on "Network Analytics" product to migrate into big data technologies to process the huge volume of unstructured data into the Data Lake. |
Exp | Convert the queries into (Spark SQL)/(Apache Drill) who were previously written in Netezza/SQL SERVER and do optimization. |
Exp | Create transformation/Loader using JSON/JEXL. |
Exp | Involved in data profiling, data cleansing, and data enrichment. |
Exp | Evaluated and tuned the SQL queries and physical database environments for efficient use. |
Exp | Involved in Unit Testing. |
Exp | Implemented various Performance Tuning techniques. |
Exp | Project #6: HIPPO Analytics |
Exp | Client: HIPPO CAMPUS, Bangalore |
Exp | Duration: June'2014 To Mar '2016 |
Exp | Role: ETL Developer, BI Developer |
Exp | Environment: Pentaho Data Integration, MySQL (Source), PostgreSQL (Target), BMBI (Reporting Tool) |
Exp | Responsibilities: |
Exp | Used Pentaho Data Integration for (ETL) extraction, transformation, and loading data from heterogeneous source systems into the target database. |
Exp | Extracted data from various sources, transformed data according to the requirement. |
Exp | Involved in extracting the data from the Flat Files and Relational databases into the staging area. |
Exp | Mappings, Sessions, Workflows from Development to Test and then to UAT environment. |
Exp | Developed Mappings and Reusable Transformations to facilitate timely Loading of Data of a star schema. |
Exp | Extracted data from various sources, transformed data according to the requirement, and loaded into the data warehouse (Data Mart). |
Exp | Created Data mart (Fact and Dimension Tables) as per the requirement and built OLAP reports on them. |
Exp | Scheduling the Transformation and Jobs in UNIX environment using CronTab. |
Exp | Prepared Technical Design documents and Test cases. |
Exp | Involved in Unit Testing. |
Exp | Implemented various Performance Tuning techniques. |
Exp | Project #7: BPUT Education Analytics |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.