When you perform kubectl apply, Kubernetes executes a series of steps to manage the desired state of the resources defined in the provided configuration files. Here’s on what happens:
Please open Telegram to view this post
VIEW IN TELEGRAM
When you're dealing with an instance in an Amazon Web Services (AWS) environment that is connected via a NAT (Network Address Translation) Gateway, it's important to understand the specific roles and configurations involved, which affect how network traffic is managed. A NAT Gateway in AWS primarily allows instances within a private subnet to connect to the Internet or other AWS services while preventing the Internet from initiating a connection with those instances. Here’s how it works:
A NAT Gateway enables instances in a private subnet to send outbound traffic to the internet, allowing for updates, downloads, and other internet-dependent activities. It also allows the instances to receive the responses from this outbound traffic.
However, the NAT Gateway does not enable inbound connections from the internet to the instances behind it. This is a security feature designed to protect instances in private subnets from unwanted external access.
Instances in the private subnet do not have public IP addresses. Instead, they are assigned private IP addresses that are not routable on the internet.
When an instance in a private subnet communicates with the internet, the NAT Gateway translates the private IP address of the instance to the public IP address of the NAT Gateway. This translation is part of why the process is called Network Address Translation.
The translation setup of the NAT Gateway only maintains the state of active connections initiated from the private subnet. Since the NAT Gateway maps multiple private IPs to a single public IP, it uses a combination of the port number and the source IP to distinguish between different connections.
When a connection is initiated from outside (the internet) without a prior corresponding internal request, the NAT Gateway has no rules or states to match this incoming connection to an internal private IP; thus, it blocks/drops such requests.
Please open Telegram to view this post
VIEW IN TELEGRAM
As a DevOps engineer, every day brings a unique blend of challenges and opportunities to drive innovation while ensuring the stability of our systems.Here’s a glimpse into what a typical day looks like
1. 𝐂𝐨𝐧𝐭𝐢𝐧𝐮𝐨𝐮𝐬 𝐈𝐧𝐭𝐞𝐠𝐫𝐚𝐭𝐢𝐨𝐧 & 𝐃𝐞𝐩𝐥𝐨𝐲𝐦𝐞𝐧𝐭 (𝐂𝐈/𝐂𝐃): Mornings often start with reviewing and enhancing our CI/CD pipelines. Automating builds, tests, and deployments not only accelerates our development cycles but also improves overall software quality
2. 𝐈𝐧𝐟𝐫𝐚𝐬𝐭𝐫𝐮𝐜𝐭𝐮𝐫𝐞 𝐚𝐬 𝐂𝐨𝐝𝐞 (𝐈𝐚𝐂): Crafting infrastructure using tools like Terraform or CloudFormation ensures consistency and scalability.
3. 𝐌𝐨𝐧𝐢𝐭𝐨𝐫𝐢𝐧𝐠 𝐚𝐧𝐝 𝐈𝐧𝐜𝐢𝐝𝐞𝐧𝐭 𝐑𝐞𝐬𝐩𝐨𝐧𝐬𝐞: Monitoring our systems is crucial. Rapid incident response is key to maintaining high availability and minimizing downtime.
4. 𝐂𝐨𝐥𝐥𝐚𝐛𝐨𝐫𝐚𝐭𝐢𝐨𝐧 & 𝐊𝐧𝐨𝐰𝐥𝐞𝐝𝐠𝐞 𝐒𝐡𝐚𝐫𝐢𝐧𝐠: DevOps thrives on collaboration. Whether it’s troubleshooting with developers, sharing best practices with teams, or participating in cross-functional meetings, fostering a culture of continuous learning is essential
5. 𝐒𝐞𝐜𝐮𝐫𝐢𝐭𝐲 𝐚𝐧𝐝 𝐂𝐨𝐦𝐩𝐥𝐢𝐚𝐧𝐜𝐞: Integrating security into every stage of our pipeline is non-negotiable.
6. 𝐂𝐨𝐧𝐭𝐢𝐧𝐮𝐨𝐮𝐬 𝐈𝐦𝐩𝐫𝐨𝐯𝐞𝐦𝐞𝐧𝐭: At the heart of DevOps is continuous improvement. Reflecting on metrics, gathering feedback, and planning optimizations are ongoing processes.
Please open Telegram to view this post
VIEW IN TELEGRAM
Docker has revolutionized the world of containerization, enabling scalable and efficient application deployment.
To make the most of this powerful tool, here are 10 essential Docker best practices:
Please open Telegram to view this post
VIEW IN TELEGRAM
Docker Documentation
Writing a Dockerfile
This concept page will teach you how to create image using Dockerfile.
A Dockerfile 🐬 is a text-based document that provides instructions for creating a container image. Let's walk through the basics of writing one:
1. Choose a Base Image:
Start by specifying the base image you want to use. It serves as the foundation for your custom image. For example:
2. Set the Working Directory:
Use the
3. Copy Files:
Use
4. Install Dependencies:
Run any necessary commands to install dependencies (e.g., using
5. Expose Ports:
Specify which ports your application will listen on using
6. Define Startup Command:
Finally, set the command that runs when the container starts:
For a hands-on tutorial, check out this Dockerfile tutorial from Docker's official documentation. [1]
➡️ Reference links: [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14]
📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
1. Choose a Base Image:
Start by specifying the base image you want to use. It serves as the foundation for your custom image. For example:
FROM node:14
2. Set the Working Directory:
Use the
WORKDIR instruction to define the working directory inside the container:WORKDIR /usr/src/app
3. Copy Files:
Use
COPY or ADD to copy files from your local machine into the image:COPY package\.json package-lock\.json \./
4. Install Dependencies:
Run any necessary commands to install dependencies (e.g., using
RUN npm install for Node.js):RUN npm install
5. Expose Ports:
Specify which ports your application will listen on using
EXPOSE:EXPOSE 3000
6. Define Startup Command:
Finally, set the command that runs when the container starts:
CMD ["npm", "start"]
Remember, this is just a basic example. You can customize your Dockerfile based on your specific application and requirements.
For a hands-on tutorial, check out this Dockerfile tutorial from Docker's official documentation. [1]
Please open Telegram to view this post
VIEW IN TELEGRAM
In this project, I will walk you through the process of deploying a Petshop Java-Based Application using Jenkins as a CI/CD tool. This deployment utilizes Docker for containerization, Kubernetes for container orchestration, and incorporates various security measures and automation tools like Terraform, SonarQube, Trivy, and Ansible. This project showcases a comprehensive approach to modern application deployment, emphasizing automation, security, and scalability.
This project was an incredible learning experience, providing hands-on practice with a variety of tools and technologies critical for modern DevOps practices.
📣 Note: Fork this Repository🧑💻 for upcoming future projects, Every week releases new Project.
Please open Telegram to view this post
VIEW IN TELEGRAM
DevOps engineers play a critical role in driving collaboration, automation, and efficiency across development and operations teams, ultimately enabling organizations to deliver high-quality software products and services more rapidly and reliably.
Please open Telegram to view this post
VIEW IN TELEGRAM
1. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗽𝗼𝗱𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀: Check the status of all pods across namespaces to identify failures.
2. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝘀𝗰𝗿𝗶𝗯𝗲 𝗽𝗼𝗱 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲: Gather detailed information about a failed pod.
3. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗹𝗼𝗴𝘀 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 -𝗰 𝗰𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿_𝗻𝗮𝗺𝗲: View logs of a specific container inside a pod to troubleshoot issues.
4. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗲𝘃𝗲𝗻𝘁𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀 --𝘀𝗼𝗿𝘁-𝗯𝘆='.𝗺𝗲𝘁𝗮𝗱𝗮𝘁𝗮.𝗰𝗿𝗲𝗮𝘁𝗶𝗼𝗻𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽': Review recent events for clues on crashes and errors.
5. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗻𝗼𝗱𝗲𝘀: Verify the status of nodes in the cluster, checking for node failures.
6. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗿𝗮𝗶𝗻 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲 --𝗶𝗴𝗻𝗼𝗿𝗲-𝗱𝗮𝗲𝗺𝗼𝗻𝘀𝗲𝘁𝘀: Safely evacuate and cordon a node for recovery operations.
7. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗰𝗼𝗿𝗱𝗼𝗻 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲: Mark a node as unschedulable to prevent new pods from being scheduled during recovery.
8. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝗹𝗲𝘁𝗲 𝗽𝗼𝗱 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 --𝗴𝗿𝗮𝗰𝗲-𝗽𝗲𝗿𝗶𝗼𝗱=0 --𝗳𝗼𝗿𝗰𝗲: Forcefully delete a crashed pod to restart it or clear it for recovery.
9. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗿𝗼𝗹𝗹𝗼𝘂𝘁 𝘂𝗻𝗱𝗼 𝗱𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁 𝗱𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁_𝗻𝗮𝗺𝗲: Roll back a deployment in case a new rollout causes crashes.
10. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗲𝘅𝗲𝗰 -𝗶𝘁 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 -- /𝗯𝗶𝗻/𝘀𝗵: Access a container to debug and resolve application issues directly inside the pod.
11. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗰𝗼𝗺𝗽𝗼𝗻𝗲𝗻𝘁𝘀𝘁𝗮𝘁𝘂𝘀𝗲𝘀: Check the health of core cluster components like etcd, kube-apiserver, and more.
12. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗼𝗽 𝗻𝗼𝗱𝗲𝘀: Monitor node resource usage to detect resource exhaustion causing crashes.
13. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗼𝗽 𝗽𝗼𝗱𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀: Check pod resource usage across namespaces, identifying bottlenecks leading to crashes.
14. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝗹𝗲𝘁𝗲 𝗻𝗼𝗱𝗲 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲: Remove a failed node from the cluster to allow recovery operations.
15. 𝗲𝘁𝗰𝗱𝗰𝘁𝗹 --𝗲𝗻𝗱𝗽𝗼𝗶𝗻𝘁𝘀=𝗵𝘁𝘁𝗽𝘀://𝗲𝘁𝗰𝗱-𝘀𝗲𝗿𝘃𝗲𝗿:2379 𝘀𝗻𝗮𝗽𝘀𝗵𝗼𝘁 𝗿𝗲𝘀𝘁𝗼𝗿𝗲 𝗯𝗮𝗰𝗸𝘂𝗽.𝗱𝗯: Restore etcd from a snapshot in case of etcd failure..
𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗮𝗽𝗽𝗹𝘆 -𝗳 𝗯𝗮𝗰𝗸𝘂𝗽.𝘆𝗮𝗺𝗹: Reapply configurations from a backup manifest during recovery.
17. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗮𝗶𝗻𝘁 𝗻𝗼𝗱𝗲𝘀 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲 𝗸𝗲𝘆=𝘃𝗮𝗹𝘂𝗲:𝗡𝗼𝗦𝗰𝗵𝗲𝗱𝘂𝗹𝗲: Prevent scheduling on a node experiencing issues during recovery.
18. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗲𝗻𝗱𝗽𝗼𝗶𝗻𝘁𝘀 𝘀𝗲𝗿𝘃𝗶𝗰𝗲_𝗻𝗮𝗺𝗲: Verify service endpoints during recovery to ensure services are resolving correctly.
Please open Telegram to view this post
VIEW IN TELEGRAM
𝑰𝒇 𝒚𝒐𝒖'𝒓𝒆 𝒖𝒏𝒔𝒖𝒓𝒆 𝒂𝒃𝒐𝒖𝒕 𝒕𝒉𝒆 𝒔𝒑𝒆𝒄𝒊𝒇𝒊𝒄 𝒓𝒆𝒔𝒑𝒐𝒏𝒔𝒊𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔 𝒐𝒇 𝒕𝒉𝒆𝒔𝒆 𝒓𝒐𝒍𝒆𝒔, 𝒅𝒆𝒕𝒂𝒊𝒍𝒆𝒅 𝒆𝒙𝒑𝒍𝒂𝒏𝒂𝒕𝒊𝒐𝒏𝒔 𝒂𝒓𝒆 𝒑𝒓𝒐𝒗𝒊𝒅𝒆𝒅 𝒃𝒆𝒍𝒐𝒘
1. 𝐃𝐞𝐯𝐎𝐩𝐬 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Bridges the gap between development and operations teams.
Automates build, test, and deployment processes.
Implements continuous integration and continuous delivery (CI/CD) pipelines.
Manages infrastructure as code (IaC) using tools like Terraform or Ansible.
Ensures system availability, performance, and scalability.
2. 𝐒𝐑𝐄 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫 (𝐒𝐢𝐭𝐞 𝐑𝐞𝐥𝐢𝐚𝐛𝐢𝐥𝐢𝐭𝐲 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫)
Focuses on reliability and performance of systems.
Builds and maintains scalable and efficient infrastructure.
Automates routine tasks and creates self-service tools.
Defines and tracks service level objectives (SLOs) and error budgets.
Handles incidents and performs root cause analysis.
3. 𝐂𝐥𝐨𝐮𝐝 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Manages and maintains cloud infrastructure (AWS, Azure, GCP).
Optimizes cloud costs and resource utilization.
Ensures cloud security and compliance.
Migrates workloads to the cloud.
Automates cloud provisioning and management.
4. 𝐏𝐥𝐚𝐭𝐟𝐨𝐫𝐦 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Builds and maintains the platform used by development teams.
Provides self-service tools and APIs for developers.
Ensures platform stability, performance, and scalability.
Collaborates with developers and infrastructure teams.
Automates platform provisioning and management.
5. 𝐃𝐞𝐯𝐒𝐞𝐜𝐎𝐩𝐬 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Integrates security into the DevOps pipeline.
Conducts security assessments and vulnerability scanning.
Implements security controls and best practices.
Develops secure coding standards and guidelines.
📱 𝐅𝐨𝐥𝐥𝐨𝐰 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
1. 𝐃𝐞𝐯𝐎𝐩𝐬 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Bridges the gap between development and operations teams.
Automates build, test, and deployment processes.
Implements continuous integration and continuous delivery (CI/CD) pipelines.
Manages infrastructure as code (IaC) using tools like Terraform or Ansible.
Ensures system availability, performance, and scalability.
2. 𝐒𝐑𝐄 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫 (𝐒𝐢𝐭𝐞 𝐑𝐞𝐥𝐢𝐚𝐛𝐢𝐥𝐢𝐭𝐲 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫)
Focuses on reliability and performance of systems.
Builds and maintains scalable and efficient infrastructure.
Automates routine tasks and creates self-service tools.
Defines and tracks service level objectives (SLOs) and error budgets.
Handles incidents and performs root cause analysis.
3. 𝐂𝐥𝐨𝐮𝐝 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Manages and maintains cloud infrastructure (AWS, Azure, GCP).
Optimizes cloud costs and resource utilization.
Ensures cloud security and compliance.
Migrates workloads to the cloud.
Automates cloud provisioning and management.
4. 𝐏𝐥𝐚𝐭𝐟𝐨𝐫𝐦 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Builds and maintains the platform used by development teams.
Provides self-service tools and APIs for developers.
Ensures platform stability, performance, and scalability.
Collaborates with developers and infrastructure teams.
Automates platform provisioning and management.
5. 𝐃𝐞𝐯𝐒𝐞𝐜𝐎𝐩𝐬 𝐄𝐧𝐠𝐢𝐧𝐞𝐞𝐫
Integrates security into the DevOps pipeline.
Conducts security assessments and vulnerability scanning.
Implements security controls and best practices.
Develops secure coding standards and guidelines.
Please open Telegram to view this post
VIEW IN TELEGRAM
DevOps & Cloud (AWS, AZURE, GCP) Tech Free Learning
Photo
1. In your current project, could you describe the overall architecture of your CI/CD pipeline that you have designed for cloud applications in Azure DevOps?
2. Can you explain how you handled the integration of infrastructure-as-code (IaC) into your Azure DevOps pipeline? Did you use tools like Azure Resource Manager templates, Terraform, or others to manage resources, and how did it integrate with your CI/CD pipeline?
3. How do you manage different deployment strategies like Blue-Green Deployment or Canary Releases using Azure DevOps and Azure Cloud?
4. In your project, how do you handle the automation of your build pipelines using Azure DevOps?
5. Can you provide examples of scripts or commands you’ve used in the release pipeline for deploying to multiple environments
6. You mentioned using GitHub Actions for CI/CD automation. Can you provide a practical example of a custom script you created using GitHub Actions for automated testing or build tasks?
7. In Azure DevOps, you can use Azure CLI or PowerShell commands to automate tasks. Can you give an example of how you utilized these tools in your CI/CD pipeline to interact with Azure resources, such as creating or updating Azure VMs, storage accounts, or App Services?
8. In the context of your deployment pipeline, can you explain how you wrote a script that triggers the deployment process after successful completion of build steps? How do you implement a rollback strategy if something goes wrong during deployment?
9. Tell me the deployment process of a web application to Azure App Services using Azure DevOps pipelines. What steps and commands do you include in the pipeline, from building the artifact to testing and deploying to production?
10. How did you implement continuous monitoring during the deployment process? Could you give an example of how you track deployments in real-time, and how do you handle failed deployments?
11. In your current project, how did you handle the containerization of applications using Docker? Can you walk us through the process of creating a Dockerfile for a web application and how you integrated it into your Azure DevOps pipeline?
12. Once you containerized an application, how did you manage the deployment to Azure Kubernetes Service (AKS)? What steps did you follow to push your Docker images to Azure Container Registry (ACR), and how did you create and deploy Kubernetes manifests (YAML)?
13. Let’s say during a deployment, your build pipeline has passed successfully, but the deployment to a pre-prod environment fails. What steps would you take to debug the issue, and which logs or commands would you check first in Azure DevOps?
14. In your CI/CD pipeline, how do you handle automated testing? Can you explain how you integrated unit tests, into your pipeline using Azure DevOps?
Please open Telegram to view this post
VIEW IN TELEGRAM
1. How would you ensure that a specific package is installed on multiple servers?
Answer: You can use the package module in a playbook to ensure that a specific package is installed across multiple servers.
2. How do you handle different environments (development, testing, production) with Ansible?
Answer: You can manage different environments by using inventory files and group variables. Create separate inventory files for each environment and use group variables to specify environment-specific configurations. Each hosts file would define the servers for that specific environment, and you can create a group_vars directory for each environment.
3. How would you restart a service after updating a configuration file?
Answer: You can use the notify feature in Ansible to restart a service after a configuration file is updated.
4. How can you ensure idempotency in your Ansible playbook?
Answer: Ansible modules are designed to be idempotent, meaning they can be run multiple times without changing the result beyond the initial application. For instance, if you use the file module to create a file, Ansible will check if the file already exists before trying to create it.
5. How do you handle secrets or sensitive data in Ansible?
Answer: You can handle sensitive data using Ansible Vault, which allows you to encrypt files or variables.
6. Can you explain how you would deploy an application using Ansible?
Answer: Define Inventory: Create an inventory file with the target hosts.
Create a Playbook: Write a playbook that includes tasks for pulling the application code from a repository, installing dependencies, configuring files, and starting services.
7. How would you handle task failures and retries in Ansible?
Answer: You can use the retry and when directives to handle task failures in Ansible. The retries and delay parameters can be specified for tasks that might need to be retried.
8. How would you roll back a deployment if the new version fails?
Answer: To roll back a deployment, you can maintain a previous version of the application and use a playbook that checks the health of the new version before deciding to switch back.
9. How can you manage firewall rules across multiple servers using Ansible?
Answer: You can use the firewalld or iptables modules to manage firewall rules.
10. How do you implement a continuous deployment pipeline using Ansible?
Answer: To implement a continuous deployment pipeline, you can integrate Ansible with a CI/CD tool like Jenkins, GitLab CI, or GitHub Actions.
11. How can you check if a file exists and create it if it doesn't?
Answer: You can use the stat module to check if a file exists and then use the copy or template module to create it if it doesn’t.
12. How can you execute a command on remote hosts and capture its output?
Answer: You can use the command or shell module to run commands on remote hosts and register the output
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
⏩ 1. Microsoft Azure Fundamentals
- Course AZ-900T00
- 24-Hour Course
- Course Link:
https://lnkd.in/dtYedpnZ
⏩ 2. Developing Solutions for Microsoft Azure
- Course AZ-204T00
- 120-Hour Course
- Course Link:
https://lnkd.in/dzVWhp7u
⏩ 3. Microsoft Azure Administrator
- Course AZ-104T00
- 96-Hour Course
- Course Link:
https://lnkd.in/djrYxCqW
⏩ 4. Configuring and Operating Microsoft Azure Virtual Desktop
- Course AZ-140
- 96-Hour Course
- Course Link:
https://lnkd.in/dsd5CPJy
⏩ 5. Designing Microsoft Azure Infrastructure Solutions
- Course AZ-305T00
- 96-Hour Course
- Course Link:
https://lnkd.in/dq28keX9
⏩ 7. Microsoft Azure Data Fundamentals
- Course DP-900T00
- 24-Hour Course
- Course Link:
https://lnkd.in/dmtfCKHM
⏩ 8. Microsoft Azure AI Fundamentals
- Course AI-900T00
- 24-Hour Course
- Course Link
https://lnkd.in/drnFx6qF
⏩ 9. Designing and Implementing a Microsoft Azure AI Solution
- Course AI-102T00
- 96-Hour Course
- Course Link:
https://lnkd.in/dt_rFFgK
⏩ 10. Develop Generative AI Solutions with Azure OpenAI Service
- Course AI-050T00
- 24-Hour Course
- Course Link:
https://lnkd.in/dKNN3mph
⏩ 11. Microsoft Security, Compliance, and Identity Fundamentals
- Course SC-900T00
- 24-Hour Course
- Course Link:
https://lnkd.in/dVWxqa_E
⏩ 12. Data Engineering on Microsoft Azure
- Course DP-203T00
- 96-Hour Course
- Course Link:
https://lnkd.in/duKTsYMa
⏩ 13. Microsoft Security Operations Analyst
- Course SC-200T00
- 96-Hour Course
- Course Link:
https://lnkd.in/du3d55NG
⏩ 14. Designing and Implementing Microsoft Azure Networking Solutions
- Course AZ-700T00
- 72-Hour Course
- Course Link:
https://lnkd.in/dgmBzYDS
⏩ 15. Designing and implementing a data science solution on Azure
- Course DP-100T01
- 96-Hour Course
- Course Link:
https://lnkd.in/dZ8WXxYx
Please open Telegram to view this post
VIEW IN TELEGRAM
DEV Community
Real-Time Resume Ready DevOps Projects
1. Jenkins CI/CD with GitHub Integration Author: DevOps-4u Link: Read More Credits: ©...
Looking to boost your DevOps skills and add impressive projects to your resume? Check out our latest article on Dev.to where I share a curated list of hands-on DevOps projects that are perfect for showcasing your expertise.
Stay ahead in the DevOps game with these projects and take your career to the next level!💼 ✨
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
Project Overview:
Check for full details
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM