Goldman Sachs senior data analyst interview asked questions
SQL
1 find avg of salaries department wise from table
2 Write a SQL query to see employee name and manager name using a self-join on 'employees' table with columns 'emp_id', 'name', and 'manager_id'.
3 newest joinee for every department (solved using lead lag)
POWER BI
1. What does Filter context in DAX mean?
2. Explain how to implement Row-Level Security (RLS) in Power BI.
3. Describe different types of filters in Power BI.
4. Explain the difference between 'ALL' and 'ALLSELECTED' in DAX.
5. How do you calculate the total sales for a specific product using DAX?
PYTHON
1. Create a dictionary, add elements to it, modify an element, and then print the dictionary in alphabetical order of keys.
2. Find unique values in a list of assorted numbers and print the count of how many times each value is repeated.
3. Find and print duplicate values in a list of assorted numbers, along with the number of times each value is repeated.
I have curated best 80+ top-notch Data Analytics Resources ๐๐
https://t.me/DataSimplifier
Hope this helps you ๐
SQL
1 find avg of salaries department wise from table
2 Write a SQL query to see employee name and manager name using a self-join on 'employees' table with columns 'emp_id', 'name', and 'manager_id'.
3 newest joinee for every department (solved using lead lag)
POWER BI
1. What does Filter context in DAX mean?
2. Explain how to implement Row-Level Security (RLS) in Power BI.
3. Describe different types of filters in Power BI.
4. Explain the difference between 'ALL' and 'ALLSELECTED' in DAX.
5. How do you calculate the total sales for a specific product using DAX?
PYTHON
1. Create a dictionary, add elements to it, modify an element, and then print the dictionary in alphabetical order of keys.
2. Find unique values in a list of assorted numbers and print the count of how many times each value is repeated.
3. Find and print duplicate values in a list of assorted numbers, along with the number of times each value is repeated.
I have curated best 80+ top-notch Data Analytics Resources ๐๐
https://t.me/DataSimplifier
Hope this helps you ๐
โค1๐1
Data Analyst Interview Questions ๐
1.How to create filters in Power BI?
Filters are an integral part of Power BI reports. They are used to slice and dice the data as per the dimensions we want. Filters are created in a couple of ways.
Using Slicers: A slicer is a visual under Visualization Pane. This can be added to the design view to filter our reports. When a slicer is added to the design view, it requires a field to be added to it. For example- Slicer can be added for Country fields. Then the data can be filtered based on countries.
Using Filter Pane: The Power BI team has added a filter pane to the reports, which is a single space where we can add different fields as filters. And these fields can be added depending on whether you want to filter only one visual(Visual level filter), or all the visuals in the report page(Page level filters), or applicable to all the pages of the report(report level filters)
2.How to sort data in Power BI?
Sorting is available in multiple formats. In the data view, a common sorting option of alphabetical order is there. Apart from that, we have the option of Sort by column, where one can sort a column based on another column. The sorting option is available in visuals as well. Sort by ascending and descending option by the fields and measure present in the visual is also available.
3.How to convert pdf to excel?
Open the PDF document you want to convert in XLSX format in Acrobat DC.
Go to the right pane and click on the โExport PDFโ option.
Choose spreadsheet as the Export format.
Select โMicrosoft Excel Workbook.โ
Now click โExport.โ
Download the converted file or share it.
4. How to enable macros in excel?
Click the file tab and then click โOptions.โ
A dialog box will appear. In the โExcel Optionsโ dialog box, click on the โTrust Centerโ and then โTrust Center Settings.โ
Go to the โMacro Settingsโ and select โenable all macros.โ
Click OK to apply the macro settings.
1.How to create filters in Power BI?
Filters are an integral part of Power BI reports. They are used to slice and dice the data as per the dimensions we want. Filters are created in a couple of ways.
Using Slicers: A slicer is a visual under Visualization Pane. This can be added to the design view to filter our reports. When a slicer is added to the design view, it requires a field to be added to it. For example- Slicer can be added for Country fields. Then the data can be filtered based on countries.
Using Filter Pane: The Power BI team has added a filter pane to the reports, which is a single space where we can add different fields as filters. And these fields can be added depending on whether you want to filter only one visual(Visual level filter), or all the visuals in the report page(Page level filters), or applicable to all the pages of the report(report level filters)
2.How to sort data in Power BI?
Sorting is available in multiple formats. In the data view, a common sorting option of alphabetical order is there. Apart from that, we have the option of Sort by column, where one can sort a column based on another column. The sorting option is available in visuals as well. Sort by ascending and descending option by the fields and measure present in the visual is also available.
3.How to convert pdf to excel?
Open the PDF document you want to convert in XLSX format in Acrobat DC.
Go to the right pane and click on the โExport PDFโ option.
Choose spreadsheet as the Export format.
Select โMicrosoft Excel Workbook.โ
Now click โExport.โ
Download the converted file or share it.
4. How to enable macros in excel?
Click the file tab and then click โOptions.โ
A dialog box will appear. In the โExcel Optionsโ dialog box, click on the โTrust Centerโ and then โTrust Center Settings.โ
Go to the โMacro Settingsโ and select โenable all macros.โ
Click OK to apply the macro settings.
โค2๐1
Expand your job search to increase your chances of becoming a data analyst.
Here are alternative roles to explore:
1. ๐๐๐๐ถ๐ป๐ฒ๐๐ ๐๐ป๐ฎ๐น๐๐๐: Focuses on using data to improve business processes and decision-making.
2. ๐ข๐ฝ๐ฒ๐ฟ๐ฎ๐๐ถ๐ผ๐ป๐ ๐๐ป๐ฎ๐น๐๐๐: Specializes in analyzing operational data to optimize efficiency and performance.
3. ๐ ๐ฎ๐ฟ๐ธ๐ฒ๐๐ถ๐ป๐ด ๐๐ป๐ฎ๐น๐๐๐: Uses data to drive marketing strategies and measure campaign effectiveness.
4. ๐๐ถ๐ป๐ฎ๐ป๐ฐ๐ถ๐ฎ๐น ๐๐ป๐ฎ๐น๐๐๐: Analyzes financial data to support investment decisions and financial planning.
5. ๐ฃ๐ฟ๐ผ๐ฑ๐๐ฐ๐ ๐๐ป๐ฎ๐น๐๐๐: Evaluates product performance and user data to help product development.
6. ๐ฅ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต ๐๐ป๐ฎ๐น๐๐๐: Conducts data-driven research to support strategic decisions and policy development.
7. ๐๐ ๐๐ป๐ฎ๐น๐๐๐: Transforms data into actionable business insights through reporting and visualization.
8. ๐ค๐๐ฎ๐ป๐๐ถ๐๐ฎ๐๐ถ๐๐ฒ ๐๐ป๐ฎ๐น๐๐๐: Utilizes statistical and mathematical models to analyze large datasets, often in finance.
9. ๐๐๐๐๐ผ๐บ๐ฒ๐ฟ ๐๐ป๐๐ถ๐ด๐ต๐๐ ๐๐ป๐ฎ๐น๐๐๐: Analyzes customer data to improve customer experience and drive retention.
10. ๐๐ฎ๐๐ฎ ๐๐ผ๐ป๐๐๐น๐๐ฎ๐ป๐: Provides expert advice on data strategies, data management, and analytics to organizations.
11. ๐ฆ๐๐ฝ๐ฝ๐น๐ ๐๐ต๐ฎ๐ถ๐ป ๐๐ป๐ฎ๐น๐๐๐: Analyzes supply chain data to optimize logistics, reduce costs, and improve efficiency.
12. ๐๐ฅ ๐๐ป๐ฎ๐น๐๐๐: Uses data to improve human resources processes, from recruitment to employee retention and performance management.
Data Analyst Roadmap ๐๐
https://whatsapp.com/channel/0029VaGgzAk72WTmQFERKh02
Hope this helps you ๐
Here are alternative roles to explore:
1. ๐๐๐๐ถ๐ป๐ฒ๐๐ ๐๐ป๐ฎ๐น๐๐๐: Focuses on using data to improve business processes and decision-making.
2. ๐ข๐ฝ๐ฒ๐ฟ๐ฎ๐๐ถ๐ผ๐ป๐ ๐๐ป๐ฎ๐น๐๐๐: Specializes in analyzing operational data to optimize efficiency and performance.
3. ๐ ๐ฎ๐ฟ๐ธ๐ฒ๐๐ถ๐ป๐ด ๐๐ป๐ฎ๐น๐๐๐: Uses data to drive marketing strategies and measure campaign effectiveness.
4. ๐๐ถ๐ป๐ฎ๐ป๐ฐ๐ถ๐ฎ๐น ๐๐ป๐ฎ๐น๐๐๐: Analyzes financial data to support investment decisions and financial planning.
5. ๐ฃ๐ฟ๐ผ๐ฑ๐๐ฐ๐ ๐๐ป๐ฎ๐น๐๐๐: Evaluates product performance and user data to help product development.
6. ๐ฅ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต ๐๐ป๐ฎ๐น๐๐๐: Conducts data-driven research to support strategic decisions and policy development.
7. ๐๐ ๐๐ป๐ฎ๐น๐๐๐: Transforms data into actionable business insights through reporting and visualization.
8. ๐ค๐๐ฎ๐ป๐๐ถ๐๐ฎ๐๐ถ๐๐ฒ ๐๐ป๐ฎ๐น๐๐๐: Utilizes statistical and mathematical models to analyze large datasets, often in finance.
9. ๐๐๐๐๐ผ๐บ๐ฒ๐ฟ ๐๐ป๐๐ถ๐ด๐ต๐๐ ๐๐ป๐ฎ๐น๐๐๐: Analyzes customer data to improve customer experience and drive retention.
10. ๐๐ฎ๐๐ฎ ๐๐ผ๐ป๐๐๐น๐๐ฎ๐ป๐: Provides expert advice on data strategies, data management, and analytics to organizations.
11. ๐ฆ๐๐ฝ๐ฝ๐น๐ ๐๐ต๐ฎ๐ถ๐ป ๐๐ป๐ฎ๐น๐๐๐: Analyzes supply chain data to optimize logistics, reduce costs, and improve efficiency.
12. ๐๐ฅ ๐๐ป๐ฎ๐น๐๐๐: Uses data to improve human resources processes, from recruitment to employee retention and performance management.
Data Analyst Roadmap ๐๐
https://whatsapp.com/channel/0029VaGgzAk72WTmQFERKh02
Hope this helps you ๐
โค2
Here's a list of commonly asked data analyst interview questions:
1. Tell me about yourself : This is often the opener, allowing you to summarize your background, skills, and experiences.
2. What is the difference between data analytics and data science?: Be ready to explain these terms and how they differ.
3. Describe a typical data analysis process you follow: Walk through steps like data collection, cleaning, analysis, and interpretation.
4. What programming languages are you proficient in?: Typically SQL, Python, R are common; mention any others you're familiar with.
5. How do you handle missing or incomplete data?: Discuss methods like imputation or excluding records based on criteria.
6. Explain a time when you used data to solve a problem: Provide a detailed example showcasing your analytical skills.
7. What data visualization tools have you used?: Tableau, Power BI, or others; discuss your experience.
8. How do you ensure the quality and accuracy of your analytical work?: Mention techniques like validation, peer reviews, or data audits.
9. What is your approach to presenting complex data findings to non-technical stakeholders?: Highlight your communication skills and ability to simplify complex information.
10. Describe a challenging data project you've worked on: Explain the project, challenges faced, and how you overcame them.
11. How do you stay updated with the latest trends in data analytics?: Talk about blogs, courses, or communities you follow.
12. What statistical techniques are you familiar with?: Regression, clustering, hypothesis testing, etc.; explain when you've used them.
13. How would you assess the effectiveness of a new data model?: Discuss metrics like accuracy, precision, recall, etc.
14. Give an example of a time when you dealt with a large dataset: Explain how you managed and processed the data efficiently.
15. Why do you want to work for this company?: Tailor your response to highlight why their industry or culture appeals to you
1. Tell me about yourself : This is often the opener, allowing you to summarize your background, skills, and experiences.
2. What is the difference between data analytics and data science?: Be ready to explain these terms and how they differ.
3. Describe a typical data analysis process you follow: Walk through steps like data collection, cleaning, analysis, and interpretation.
4. What programming languages are you proficient in?: Typically SQL, Python, R are common; mention any others you're familiar with.
5. How do you handle missing or incomplete data?: Discuss methods like imputation or excluding records based on criteria.
6. Explain a time when you used data to solve a problem: Provide a detailed example showcasing your analytical skills.
7. What data visualization tools have you used?: Tableau, Power BI, or others; discuss your experience.
8. How do you ensure the quality and accuracy of your analytical work?: Mention techniques like validation, peer reviews, or data audits.
9. What is your approach to presenting complex data findings to non-technical stakeholders?: Highlight your communication skills and ability to simplify complex information.
10. Describe a challenging data project you've worked on: Explain the project, challenges faced, and how you overcame them.
11. How do you stay updated with the latest trends in data analytics?: Talk about blogs, courses, or communities you follow.
12. What statistical techniques are you familiar with?: Regression, clustering, hypothesis testing, etc.; explain when you've used them.
13. How would you assess the effectiveness of a new data model?: Discuss metrics like accuracy, precision, recall, etc.
14. Give an example of a time when you dealt with a large dataset: Explain how you managed and processed the data efficiently.
15. Why do you want to work for this company?: Tailor your response to highlight why their industry or culture appeals to you
โค1
SQL Interview Questions with Answers
1. How to change a table name in SQL?
This is the command to change a table name in SQL:
ALTER TABLE table_name
RENAME TO new_table_name;
We will start off by giving the keywords ALTER TABLE, then we will follow it up by giving the original name of the table, after that, we will give in the keywords RENAME TO and finally, we will give the new table name.
2. How to use LIKE in SQL?
The LIKE operator checks if an attribute value matches a given string pattern. Here is an example of LIKE operator
SELECT * FROM employees WHERE first_name like โStevenโ;
With this command, we will be able to extract all the records where the first name is like โStevenโ.
3. If we drop a table, does it also drop related objects like constraints, indexes, columns, default, views and sorted procedures?
Yes, SQL server drops all related objects, which exists inside a table like constraints, indexes, columns, defaults etc. But dropping a table will not drop views and sorted procedures as they exist outside the table.
4. Explain SQL Constraints.
SQL Constraints are used to specify the rules of data type in a table. They can be specified while creating and altering the table. The following are the constraints in SQL: NOT NULL CHECK DEFAULT UNIQUE PRIMARY KEY FOREIGN KEY
React โค๏ธ for more
1. How to change a table name in SQL?
This is the command to change a table name in SQL:
ALTER TABLE table_name
RENAME TO new_table_name;
We will start off by giving the keywords ALTER TABLE, then we will follow it up by giving the original name of the table, after that, we will give in the keywords RENAME TO and finally, we will give the new table name.
2. How to use LIKE in SQL?
The LIKE operator checks if an attribute value matches a given string pattern. Here is an example of LIKE operator
SELECT * FROM employees WHERE first_name like โStevenโ;
With this command, we will be able to extract all the records where the first name is like โStevenโ.
3. If we drop a table, does it also drop related objects like constraints, indexes, columns, default, views and sorted procedures?
Yes, SQL server drops all related objects, which exists inside a table like constraints, indexes, columns, defaults etc. But dropping a table will not drop views and sorted procedures as they exist outside the table.
4. Explain SQL Constraints.
SQL Constraints are used to specify the rules of data type in a table. They can be specified while creating and altering the table. The following are the constraints in SQL: NOT NULL CHECK DEFAULT UNIQUE PRIMARY KEY FOREIGN KEY
React โค๏ธ for more
โค9
Here are 5 key Python libraries/ concepts that are particularly important for data analysts:
1. Pandas: Pandas is a powerful library for data manipulation and analysis in Python. It provides data structures like DataFrames and Series that make it easy to work with structured data. Pandas offers functions for reading and writing data, cleaning and transforming data, and performing data analysis tasks like filtering, grouping, and aggregating.
2. NumPy: NumPy is a fundamental package for scientific computing in Python. It provides support for large, multi-dimensional arrays and matrices, along with a collection of mathematical functions to operate on these arrays efficiently. NumPy is often used in conjunction with Pandas for numerical computations and data manipulation.
3. Matplotlib and Seaborn: Matplotlib is a popular plotting library in Python that allows you to create a wide variety of static, interactive, and animated visualizations. Seaborn is built on top of Matplotlib and provides a higher-level interface for creating attractive and informative statistical graphics. These libraries are essential for data visualization in data analysis projects.
4. Scikit-learn: Scikit-learn is a machine learning library in Python that provides simple and efficient tools for data mining and data analysis tasks. It includes a wide range of algorithms for classification, regression, clustering, dimensionality reduction, and more. Scikit-learn also offers tools for model evaluation, hyperparameter tuning, and model selection.
5. Data Cleaning and Preprocessing: Data cleaning and preprocessing are crucial steps in any data analysis project. Python offers libraries like Pandas and NumPy for handling missing values, removing duplicates, standardizing data types, scaling numerical features, encoding categorical variables, and more. Understanding how to clean and preprocess data effectively is essential for accurate analysis and modeling.
By mastering these Python concepts and libraries, data analysts can efficiently manipulate and analyze data, create insightful visualizations, apply machine learning techniques, and derive valuable insights from their datasets.
Credits: https://t.me/free4unow_backup
ENJOY LEARNING ๐๐
1. Pandas: Pandas is a powerful library for data manipulation and analysis in Python. It provides data structures like DataFrames and Series that make it easy to work with structured data. Pandas offers functions for reading and writing data, cleaning and transforming data, and performing data analysis tasks like filtering, grouping, and aggregating.
2. NumPy: NumPy is a fundamental package for scientific computing in Python. It provides support for large, multi-dimensional arrays and matrices, along with a collection of mathematical functions to operate on these arrays efficiently. NumPy is often used in conjunction with Pandas for numerical computations and data manipulation.
3. Matplotlib and Seaborn: Matplotlib is a popular plotting library in Python that allows you to create a wide variety of static, interactive, and animated visualizations. Seaborn is built on top of Matplotlib and provides a higher-level interface for creating attractive and informative statistical graphics. These libraries are essential for data visualization in data analysis projects.
4. Scikit-learn: Scikit-learn is a machine learning library in Python that provides simple and efficient tools for data mining and data analysis tasks. It includes a wide range of algorithms for classification, regression, clustering, dimensionality reduction, and more. Scikit-learn also offers tools for model evaluation, hyperparameter tuning, and model selection.
5. Data Cleaning and Preprocessing: Data cleaning and preprocessing are crucial steps in any data analysis project. Python offers libraries like Pandas and NumPy for handling missing values, removing duplicates, standardizing data types, scaling numerical features, encoding categorical variables, and more. Understanding how to clean and preprocess data effectively is essential for accurate analysis and modeling.
By mastering these Python concepts and libraries, data analysts can efficiently manipulate and analyze data, create insightful visualizations, apply machine learning techniques, and derive valuable insights from their datasets.
Credits: https://t.me/free4unow_backup
ENJOY LEARNING ๐๐
๐2โค1
7 Essential Power BI Tips for Efficient Report Design
Use DAX Measures Over Calculated Columns
DAX measures are generally more efficient and flexible than calculated columns. They calculate results dynamically and improve report performance.
Take Advantage of Drillthrough and Tooltips
Drillthrough allows users to zoom into a specific data point for deeper insights, while tooltips provide additional information when hovering over visuals.
Keep Data Models Simple
Focus on a clean, simple data model. Overcomplicating it can make maintenance harder and lead to performance issues. Stick to the essential tables and relationships.
Design for User Experience
Prioritize user-friendly reports. A clean and intuitive design with interactive filters, slicers, and clearly labeled visuals enhances user experience.
Limit the Number of Visuals
Avoid overwhelming your report with too many visuals. Stick to key performance indicators (KPIs) and keep visuals focused to tell a clear story.
Use Power Query for Data Transformation
Power Query is your go-to tool for cleaning, transforming, and shaping your data before importing it into Power BI. It ensures a cleaner, more efficient dataset.
Implement Date Tables for Time Intelligence
If you need to perform time-based analysis, always create or use a date table. Power BI requires a dedicated date table to correctly perform time-based calculations like YTD, MTD, and QTD.
Power BI Learning Series: https://whatsapp.com/channel/0029Vai1xKf1dAvuk6s1v22c
Use DAX Measures Over Calculated Columns
DAX measures are generally more efficient and flexible than calculated columns. They calculate results dynamically and improve report performance.
Take Advantage of Drillthrough and Tooltips
Drillthrough allows users to zoom into a specific data point for deeper insights, while tooltips provide additional information when hovering over visuals.
Keep Data Models Simple
Focus on a clean, simple data model. Overcomplicating it can make maintenance harder and lead to performance issues. Stick to the essential tables and relationships.
Design for User Experience
Prioritize user-friendly reports. A clean and intuitive design with interactive filters, slicers, and clearly labeled visuals enhances user experience.
Limit the Number of Visuals
Avoid overwhelming your report with too many visuals. Stick to key performance indicators (KPIs) and keep visuals focused to tell a clear story.
Use Power Query for Data Transformation
Power Query is your go-to tool for cleaning, transforming, and shaping your data before importing it into Power BI. It ensures a cleaner, more efficient dataset.
Implement Date Tables for Time Intelligence
If you need to perform time-based analysis, always create or use a date table. Power BI requires a dedicated date table to correctly perform time-based calculations like YTD, MTD, and QTD.
Power BI Learning Series: https://whatsapp.com/channel/0029Vai1xKf1dAvuk6s1v22c
โค1
How to Become a Data Analyst from Scratch! ๐
Whether you're starting fresh or upskilling, here's your roadmap:
โ Master Excel and SQL - solve SQL problems from leetcode & hackerank
โ Get the hang of either Power BI or Tableau - do some hands-on projects
โ learn what the heck ATS is and how to get around it
โ learn to be ready for any interview question
โ Build projects for a data portfolio
โ And you don't need to do it all at once!
โ Fail and learn to pick yourself up whenever required
Whether it's acing interviews or building an impressive portfolio, give yourself the space to learn, fail, and grow. Good things take time โ
Like if it helps โค๏ธ
I have curated best 80+ top-notch Data Analytics Resources ๐๐
https://topmate.io/analyst/861634
Hope it helps :)
Whether you're starting fresh or upskilling, here's your roadmap:
โ Master Excel and SQL - solve SQL problems from leetcode & hackerank
โ Get the hang of either Power BI or Tableau - do some hands-on projects
โ learn what the heck ATS is and how to get around it
โ learn to be ready for any interview question
โ Build projects for a data portfolio
โ And you don't need to do it all at once!
โ Fail and learn to pick yourself up whenever required
Whether it's acing interviews or building an impressive portfolio, give yourself the space to learn, fail, and grow. Good things take time โ
Like if it helps โค๏ธ
I have curated best 80+ top-notch Data Analytics Resources ๐๐
https://topmate.io/analyst/861634
Hope it helps :)
โค1