DevOps & Cloud (AWS, AZURE, GCP) Tech Free Learning
16K subscribers
1.33K photos
14 videos
501 files
1.28K links
https://projects.prodevopsguytech.com // https://blog.prodevopsguytech.com

• We post Daily Trending DevOps/Cloud content
• All DevOps related Code & Scripts uploaded
• DevOps/Cloud Job Related Posts
• Real-time Interview questions & preparation guides
Download Telegram
As a DevOps engineer, we don’t deploy to cloud or servers daily, but most of our time is spent on troubleshooting. Here are some important HTTP status codes every DevOps engineer should know when troubleshooting:

🚫 Status code 403 Forbidden:
This usually occurs due to user permission issues, like incorrect IAM role or policy assignments.

🔍 Status code 404 Resource Not Found:
This typically happens when a route or method is unavailable on API Gateway or the backend server.

⚠️ Status code 502 Bad Gateway:
Occurs when a server acting as a gateway or proxy (e.g., Nginx, Apache, AWS API Gateway) receives an invalid or no response from the upstream server it's trying to communicate with.

⛔️ Status code 503 Service Unavailable:
Indicates that the server is temporarily unable to handle the request, often due to overload or maintenance.


✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🚀 DevOps Project - 29 🚀


🔗 Project Link: HERE

➡️ CI/CD Project: Deploy a 3-tier Microservice Voting App using ArgoCD and Azure DevOps Pipeline


I’m excited to share new latest project 29: CI/CD Project: Deploy a 3-tier Microservice Voting App using ArgoCD and Azure DevOps Pipeline.

🔧 This project demonstrates:
- Setting up a CI/CD pipeline using Azure DevOps
- Automated deployment with ArgoCD
- Scaling a 3-tier Microservice architecture for a Voting App

If you’re into DevOps, Kubernetes, or CI/CD Pipelines, this is for you! 🔥

Check out the repo, explore the code, and feel free to reach out if you have any questions. Let's discuss and collaborate! 🙌

❤️‍🔥 Share with friends and learning aspirants ❤️‍🔥

📣 Note: Fork this Repository 🧑‍💻 for upcoming future projects, Every week releases new Project.



📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
1727106322516.gif
682 KB
🚨 What makes AWS Lambda so fast?


⚡️There are 4 main pillars:

1️⃣- Function Invocation
AWS Lambda supports synchronous and asynchronous invocation.
- In synchronous invocation, the caller directly calls the Lambda function using AWS CLI, SDK, or other services.
- In an asynchronous invocation, the caller doesn’t wait for the function’s response. The request is authorized and an event is placed in an internal SQS queue. Pollers read messages from the queue and send them for processing.

2️⃣ - Assignment Service
The Assignment Service manages the execution environments.
- The service is written in Rust for high performance and is divided into multiple partitions with a leader-follower approach for high availability.
- The state of execution environments is written to an external journal log.

3️⃣ - Firecracker MicroVM
Firecracker is a lightweight virtual machine manager designed for running serverless workloads such as AWS Lambda and AWS Fargate.
- It uses Linux’s Kernel-based virtual machine to create and manage secure, fast-booting microVMs.

4️⃣ - Component Storage
AWS Lambda also has to manage the state consisting of input data and function code.

⚡️To make it efficient, it uses multiple techniques:
- Chunking to store the container images more efficiently.
- Using convergent encryption to secure the shared data. This involves appending additional data to the chunk to compute a more robust hash.
- SnapStart feature to reduce cold start latency by pre-initializing the execution environment


📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🚨 Top Kubernetes crash recovery commands that used 99% of the time:


1. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗽𝗼𝗱𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀: Check the status of all pods across namespaces to identify failures.
2. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝘀𝗰𝗿𝗶𝗯𝗲 𝗽𝗼𝗱 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲: Gather detailed information about a failed pod.
3. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗹𝗼𝗴𝘀 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 -𝗰 𝗰𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿_𝗻𝗮𝗺𝗲: View logs of a specific container inside a pod to troubleshoot issues.
4. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗲𝘃𝗲𝗻𝘁𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀 --𝘀𝗼𝗿𝘁-𝗯𝘆='.𝗺𝗲𝘁𝗮𝗱𝗮𝘁𝗮.𝗰𝗿𝗲𝗮𝘁𝗶𝗼𝗻𝗧𝗶𝗺𝗲𝘀𝘁𝗮𝗺𝗽': Review recent events for clues on crashes and errors.
5. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗻𝗼𝗱𝗲𝘀: Verify the status of nodes in the cluster, checking for node failures.
6. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗿𝗮𝗶𝗻 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲 --𝗶𝗴𝗻𝗼𝗿𝗲-𝗱𝗮𝗲𝗺𝗼𝗻𝘀𝗲𝘁𝘀: Safely evacuate and cordon a node for recovery operations.
7. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗰𝗼𝗿𝗱𝗼𝗻 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲: Mark a node as unschedulable to prevent new pods from being scheduled during recovery.
8. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝗹𝗲𝘁𝗲 𝗽𝗼𝗱 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 --𝗴𝗿𝗮𝗰𝗲-𝗽𝗲𝗿𝗶𝗼𝗱=0 --𝗳𝗼𝗿𝗰𝗲: Forcefully delete a crashed pod to restart it or clear it for recovery.
9. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗿𝗼𝗹𝗹𝗼𝘂𝘁 𝘂𝗻𝗱𝗼 𝗱𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁 𝗱𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁_𝗻𝗮𝗺𝗲: Roll back a deployment in case a new rollout causes crashes.
10. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗲𝘅𝗲𝗰 -𝗶𝘁 𝗽𝗼𝗱_𝗻𝗮𝗺𝗲 -- /𝗯𝗶𝗻/𝘀𝗵: Access a container to debug and resolve application issues directly inside the pod.
11. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗰𝗼𝗺𝗽𝗼𝗻𝗲𝗻𝘁𝘀𝘁𝗮𝘁𝘂𝘀𝗲𝘀: Check the health of core cluster components like etcd, kube-apiserver, and more.
12. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗼𝗽 𝗻𝗼𝗱𝗲𝘀: Monitor node resource usage to detect resource exhaustion causing crashes.
13. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗼𝗽 𝗽𝗼𝗱𝘀 --𝗮𝗹𝗹-𝗻𝗮𝗺𝗲𝘀𝗽𝗮𝗰𝗲𝘀: Check pod resource usage across namespaces, identifying bottlenecks leading to crashes.
14. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗱𝗲𝗹𝗲𝘁𝗲 𝗻𝗼𝗱𝗲 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲: Remove a failed node from the cluster to allow recovery operations.
15. 𝗲𝘁𝗰𝗱𝗰𝘁𝗹 --𝗲𝗻𝗱𝗽𝗼𝗶𝗻𝘁𝘀=𝗵𝘁𝘁𝗽𝘀://𝗲𝘁𝗰𝗱-𝘀𝗲𝗿𝘃𝗲𝗿:2379 𝘀𝗻𝗮𝗽𝘀𝗵𝗼𝘁 𝗿𝗲𝘀𝘁𝗼𝗿𝗲 𝗯𝗮𝗰𝗸𝘂𝗽.𝗱𝗯: Restore etcd from a snapshot in case of etcd failure.
16. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗮𝗽𝗽𝗹𝘆 -𝗳 𝗯𝗮𝗰𝗸𝘂𝗽.𝘆𝗮𝗺𝗹: Reapply configurations from a backup manifest during recovery.
17. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝘁𝗮𝗶𝗻𝘁 𝗻𝗼𝗱𝗲𝘀 𝗻𝗼𝗱𝗲_𝗻𝗮𝗺𝗲 𝗸𝗲𝘆=𝘃𝗮𝗹𝘂𝗲:𝗡𝗼𝗦𝗰𝗵𝗲𝗱𝘂𝗹𝗲: Prevent scheduling on a node experiencing issues during recovery.
18. 𝗸𝘂𝗯𝗲𝗰𝘁𝗹 𝗴𝗲𝘁 𝗲𝗻𝗱𝗽𝗼𝗶𝗻𝘁𝘀 𝘀𝗲𝗿𝘃𝗶𝗰𝗲_𝗻𝗮𝗺𝗲: Verify service endpoints during recovery to ensure services are resolving correctly.


📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
⚡️AWS DEVOPS REAL-TIME DEPLOYMENT

Development → Pre-PROD → Production


▶️ Detailed Project Explanation with Screenshots : https://blog.prodevopsguytech.com/aws-devops-real-time-deployment-dev-pre-prod-production

📱 Project Source code: https://github.com/NotHarshhaa/AWS-DevOps_Real-Time_Deployment


📱 𝐅𝐨𝐥𝐥𝐨𝐰 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
Master DevOps from Beginner to Expert


⚡️ Start with foundational skills like Version Control (Git), Automating Builds (CI), and Infrastructure as Code (IaC). Get hands-on with Docker, cloud basics, and deployment automation, while also covering basic monitoring and documentation.

⚡️ Progress to intermediate activities by mastering CI/CD pipelines, advanced scripting, and container orchestration with Kubernetes. Dive into advanced monitoring tools like ELK and Grafana, and focus on security, performance optimization, database automation, and knowledge sharing.

⚡️ Achieve advanced proficiency with enterprise CI/CD, advanced security measures, and Kubernetes optimization. Enhance your skills in advanced monitoring & analytics, cloud strategies, disaster recovery planning, and cross-team collaboration.

⚡️ Reach expert status with a focus on architecture design, AI/ML automation, and regulatory compliance. Lead with continuous improvement, strategic planning, vendor management, and executive communication. Develop your expertise in mentorship, leadership, and thought leadership.


📱 𝐅𝐨𝐥𝐥𝐨𝐰 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🚀 New Article Alert: End-to-End AWS DevOps Project 🚀


🖥 Read the full project here: End-to-End AWS DevOps Project: CI/CD Pipeline for ECS Fargate with ECR and RDS

We just published an in-depth AWS DevOps project guide that walks you through setting up a CI/CD pipeline for a containerized Node.js application deployed on AWS ECS Fargate. This tutorial covers everything from creating Docker images, pushing to Amazon ECR, setting up a managed database with Amazon RDS, and automating deployments using AWS CodePipeline.

👨‍💻 What you'll learn:
1. Containerizing a Node.js app with Docker 🐳
2. Deploying to AWS ECS Fargate without worrying about servers 🌐
3. Automating your CI/CD pipeline with AWS CodePipeline and CodeBuild 🤖
4. Managing your database with Amazon RDS 🗄
5. Real-time monitoring and logging with AWS CloudWatch 📊

💡 If you're looking to enhance your DevOps skills or want a practical project to showcase, this is for you! Packed with step-by-step instructions, installation commands, and diagrams to guide you through the entire process.

Don't forget to ⭐️ and share! Your feedback is much appreciated.


📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
📢 Free Resources
📢 Microsoft Certified Azure Solutions Architect Pathway

Connect 🤝 Like 👍 Share 🤟

➡️ Foundations
✔️ Exam AZ-900: Microsoft Azure Fundamentals
🔗 https://lnkd.in/g4NRXQNF

➡️ Administrator
✔️ Exam AZ-104: Microsoft Azure Administrator
🔗 https://lnkd.in/gfpEkQCn

➡️ Expert
✔️ Exam AZ-305: Designing Microsoft Azure Infrastructure Solutions
🔗 https://lnkd.in/gGUD3Jps


✉️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
▶️ Real-time interview questions and answers 💬 related to Ansible:-

1. How would you ensure that a specific package is installed on multiple servers?
Answer: You can use the package module in a playbook to ensure that a specific package is installed across multiple servers.

2. How do you handle different environments (development, testing, production) with Ansible?
Answer: You can manage different environments by using inventory files and group variables. Create separate inventory files for each environment and use group variables to specify environment-specific configurations. Each hosts file would define the servers for that specific environment, and you can create a group_vars directory for each environment.

3. How would you restart a service after updating a configuration file?
Answer: You can use the notify feature in Ansible to restart a service after a configuration file is updated.

4. How can you ensure idempotency in your Ansible playbook?
Answer: Ansible modules are designed to be idempotent, meaning they can be run multiple times without changing the result beyond the initial application. For instance, if you use the file module to create a file, Ansible will check if the file already exists before trying to create it.

5. How do you handle secrets or sensitive data in Ansible?
Answer: You can handle sensitive data using Ansible Vault, which allows you to encrypt files or variables.

6. Can you explain how you would deploy an application using Ansible?
Answer: Define Inventory: Create an inventory file with the target hosts.
Create a Playbook: Write a playbook that includes tasks for pulling the application code from a repository, installing dependencies, configuring files, and starting services.

7. How would you handle task failures and retries in Ansible?
Answer: You can use the retry and when directives to handle task failures in Ansible. The retries and delay parameters can be specified for tasks that might need to be retried.

8. How would you roll back a deployment if the new version fails?
Answer: To roll back a deployment, you can maintain a previous version of the application and use a playbook that checks the health of the new version before deciding to switch back.

9. How can you manage firewall rules across multiple servers using Ansible?
Answer: You can use the firewalld or iptables modules to manage firewall rules.

10. How do you implement a continuous deployment pipeline using Ansible?
Answer: To implement a continuous deployment pipeline, you can integrate Ansible with a CI/CD tool like Jenkins, GitLab CI, or GitHub Actions.

11. How can you check if a file exists and create it if it doesn't?
Answer: You can use the stat module to check if a file exists and then use the copy or template module to create it if it doesn’t.

12. How can you execute a command on remote hosts and capture its output?
Answer: You can use the command or shell module to run commands on remote hosts and register the output


📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
Kubernetes Architecture: Key Components 📱

Master Node: Manages the cluster, handling the control plane components.
💎 API Server: Frontend for Kubernetes, handling all communication.
💎 Scheduler: Assigns pods to nodes based on resource availability.
💎 Controller Manager: Manages cluster state, scaling, and node health.
💎 etcd: Distributed key-value store for all cluster data.

Worker Nodes: Run application workloads.
💎 Kubelet: Ensures containers are running as defined in Pod specs.
💎 Kube-Proxy: Manages network rules, allowing communication inside/outside the cluster.
💎 Container Runtime: Runs containers (e.g., Docker, containerd).
💎 Pods: Smallest deployable unit, encapsulating containers.
💎 Services: Stable endpoint for connecting to Pods.
💎 Namespaces: Logical partitioning of resources for isolation.

This architecture ensures efficient scaling, fault tolerance, and high availability for cloud-native applications.



📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🔣 List of Kubernetes commands: 🔣

➡️ ReplicaSets Management
kubectl create -f <replicaset-definition.yaml>: Create a ReplicaSet.
kubectl get replicasets: List all ReplicaSets.
kubectl describe replicaset <replicaset-name>: Describe a specific ReplicaSet.
kubectl scale replicaset <replicaset-name> –replicas=<replica-count>: Scale a ReplicaSet.

➡️ Service Management
kubectl create service <service-type> <service-name> –tcp=<port>: Create a service.
kubectl get services: List all services.
kubectl expose deployment <deployment-name> –port=<port>: Expose a deployment as a service.
kubectl describe service <service-name>: Describe a specific service.
kubectl delete service <service-name>: Delete a service.
kubectl get endpoints <service-name>: Get information about a service.

➡️ Config Maps and Secrets
kubectl create configmap <config-map-name> –from-file=<path-to-file>: Create a config map from a file.
kubectl create secret <secret-type> <secret-name> –from-literal=<key>=<value>: Create a secret.
kubectl get configmaps: List all config maps.
kubectl get secrets: List all secrets.
kubectl describe configmap <config-map-name>: Describe a specific config map.
kubectl describe secret <secret-name>: Describe a specific secret.
kubectl delete secret <secret_name>: Delete a specific secret.
kubectl delete configmap <config-map-name>: Delete a specific config map.

➡️ Networking
kubectl port-forward <pod-name> <local-port>:<pod-port>: Port forward to a pod.
kubectl expose deployment <deployment-name> –type=NodePort –port=<port>: Expose a deployment as a NodePort service.
kubectl create ingress <ingress-name> –rule=<host>/<path>=<service-name> –<service-port>: Create an Ingress resource.
kubectl describe ingress <ingress-name>: Get information about an Ingress.
kubectl get ingress <ingress-name> -o jsonpath='{.spec.rules[0].host}’: Retrieves the most value from the first rule of the specified Ingress resource.

➡️ Storage
kubectl create -f <persistent-volume-definition.yaml>: Create a PersistentVolume.
kubectl get pv: List all PersistentVolumes.
kubectl describe pv <pv-name>: Describe a specific PersistentVolume.
kubectl create -f <persistent-volume-claim-definition.yaml>: Create a PersistentVolumeClaim.
kubectl get pvc: List all PersistentVolumeClaims.
kubectl describe pvc <pvc-name>: Describe a specific PersistentVolumeClaim.

➡️ StatefulSets
kubectl create -f <statefulset-definition.yaml>: Create a StatefulSet.
kubectl get statefulsets: List all StatefulSets.
kubectl describe statefulset <statefulset-name>: Describe a specific StatefulSet.
kubectl scale statefulset <statefulset-name> –replicas=<replica-count>: Scale a StatefulSet.

➡️ Monitoring and Troubleshooting
kubectl get events: Check cluster events.
kubectl get component statuses: Get cluster component statuses.
kubectl top nodes: Get resource utilization of nodes.
kubectl top pods: Get resource utilization of pods.
kubectl debug <pod-name> -it –image=<debugging-image>: Enable container shell access debugging.


😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
▶️ 𝗔𝗻𝘀𝗶𝗯𝗹𝗲 𝗕𝗲𝘀𝘁 𝗣𝗿𝗮𝗰𝘁𝗶𝗰𝗲𝘀 𝗳𝗼𝗿 𝗔𝘂𝘁𝗼𝗺𝗮𝘁𝗶𝗼𝗻 𝗦𝘂𝗰𝗰𝗲𝘀𝘀! 📱

Are you ready to unlock the power of Ansible and boost your DevOps workflows? Here's a quick breakdown of the core concepts, tips, and tricks to get you on the right path to efficient automation. 👇

⭐️ 𝗖𝗼𝗿𝗲 𝗖𝗼𝗻𝗰𝗲𝗽𝘁𝘀
𝗣𝗹𝗮𝘆𝗯𝗼𝗼𝗸𝘀 📜
YAML files where automation lives! Write them to describe the desired state of your infrastructure.
𝗧𝗶𝗽: Keep them simple and modular for readability.

𝗜𝗻𝘃𝗲𝗻𝘁𝗼𝗿𝗶𝗲𝘀 📂
Define your hosts and groups of servers here.
𝗧𝗶𝗽: Use dynamic inventory scripts for cloud platforms like AWS to stay updated.

𝗠𝗼𝗱𝘂𝗹𝗲𝘀 🧩
Predefined functions that automate tasks like installing packages, copying files, etc.
𝗧𝗶𝗽: Make use of idempotent modules to ensure consistent results!

𝗥𝗼𝗹𝗲𝘀 🏗️
Group related tasks, variables, and handlers in roles to keep things organized.
𝗧𝗶𝗽: Share your roles with others through Ansible Galaxy.

𝗛𝗮𝗻𝗱𝗹𝗲𝗿𝘀 🚦
Respond to changes and only run tasks when necessary.
𝗧𝗶𝗽: Use handlers to restart services or trigger additional tasks, minimizing downtime.

💡 𝗧𝗶𝗽𝘀 & 𝗧𝗿𝗶𝗰𝗸𝘀
𝗔𝘃𝗼𝗶𝗱 𝗛𝗮𝗿𝗱𝗰𝗼𝗱𝗶𝗻𝗴 : Use variables and parameterize your playbooks to keep them flexible.
𝗨𝘀𝗲 𝗧𝗮𝗴𝘀 🏷️: Assign tags to tasks and run specific parts of playbooks without executing everything.
𝗗𝗿𝘆 𝗥𝘂𝗻 (𝗖𝗵𝗲𝗰𝗸 𝗠𝗼𝗱𝗲) 🔍: Always test before execution with --check mode to ensure no surprises.
𝗬𝗔𝗠𝗟 𝗙𝗼𝗿𝗺𝗮𝘁𝘁𝗶𝗻𝗴 🖋️: Stick to best YAML practices for indentation and structure—Ansible is strict about it!

🛠 Example
𝗣𝗿𝗼𝗯𝗹𝗲𝗺: Configuring 100 EC2 instances with different setups manually is tedious and error-prone. 😓
𝗔𝗰𝘁𝗶𝗼𝗻: Create a dynamic inventory, use roles to define common configurations, and execute your playbook across all instances. 🚀
𝗥𝗲𝘀𝘂𝗹𝘁: Successfully 𝗰𝗼𝗻𝗳𝗶𝗴𝘂𝗿𝗲𝗱 𝗮𝗹𝗹 𝟭𝟬𝟬 𝗶𝗻𝘀𝘁𝗮𝗻𝗰𝗲𝘀 𝗶𝗻 𝗺𝗶𝗻𝘂𝘁𝗲𝘀 𝘄𝗶𝘁𝗵 𝘇𝗲𝗿𝗼 𝗺𝗮𝗻𝘂𝗮𝗹 𝗲𝗿𝗿𝗼𝗿𝘀. 🎉

🔥 𝗪𝗵𝘆 𝗖𝗵𝗼𝗼𝘀𝗲 𝗔𝗻𝘀𝗶𝗯𝗹𝗲?
𝗔𝗴𝗲𝗻𝘁𝗹𝗲𝘀𝘀: No need to install agents on nodes.
𝗜𝗱𝗲𝗺𝗽𝗼𝘁𝗲𝗻𝗰𝘆: Ensures tasks are executed exactly as intended.
𝗦𝗰𝗮𝗹𝗮𝗯𝗶𝗹𝗶𝘁𝘆: Perfect for small to large infrastructures.

🔔 Action Time! If you want to streamline your DevOps automation and master Ansible, start now!



📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
1711255043413.gif
2.3 MB
👾 CICD vs. CICD? You might be saying it wrong...

While CICD gets thrown around a lot, it actually refers to two separate practices that work together in the software development lifecycle: Continuous Integration (CI) and Continuous Delivery/Deployment (CD).

Here's a quick breakdown:

➡️𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗼𝘂𝘀 𝗜𝗻𝘁𝗲𝗴𝗿𝗮𝘁𝗶𝗼𝗻 (𝗖𝗜): Automates the process of merging code changes from developers into a shared mainline frequently. This means every push triggers builds, tests, and catches bugs early on.

➡️𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗼𝘂𝘀 𝗗𝗲𝗹𝗶𝘃𝗲𝗿𝘆/𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁 (𝗖𝗗): Takes CI a step further. It automates the entire delivery pipeline, allowing you to release new features or bug fixes to production environments quickly and reliably. CD can include manual approval gates before deployment, while Continuous Deployment (CD) automates deployments entirely.

Here's the key difference:

➡️𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗼𝘂𝘀 𝗗𝗲𝗹𝗶𝘃𝗲𝗿𝘆: Gives you a "deploy button" - the code is ready to be pushed to production with a manual go-ahead.

➡️𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗼𝘂𝘀 𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁: Takes it a step further. If all automated tests pass in CD, the code is automatically deployed to production without human intervention.


✈️ 𝐅𝐨𝐥𝐥𝐨𝐰 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
1708866864365.gif
706.6 KB
𝐇𝐨𝐰 𝐭𝐨 𝐛𝐞𝐜𝐨𝐦𝐞 𝐚 𝐜𝐥𝐨𝐮𝐝 𝐞𝐧𝐠𝐢𝐧𝐞𝐞𝐫: 𝐀 𝐫𝐨𝐚𝐝𝐦𝐚𝐩

🔹 Skills Required: Master cloud fundamentals, networking, programming, infrastructure as code (IaC), containerization, monitoring, automation, database management, and cloud security. Strong communication and collaboration skills are essential for teamwork.

📚 Education & Training: While a formal degree isn't always necessary, consider a Computer Science background and pursue cloud certifications. Explore online courses to keep your skills sharp.

⭐️ Career Prospects: Cloud Engineer, Cloud Architect, DevOps Engineer, Cloud Consultant - the opportunities are limitless in this dynamic field!

⚙️ Tools & Technologies: Get comfortable with major cloud service providers (AWS, Azure, GCP), IaC tools (Terraform, CloudFormation), containerization (Docker, Kubernetes), CI/CD tools (Jenkins, GitLab CI/CD), and monitoring solutions.

⬆️ Cloud Engineering is your ticket to a future-proof career. Stay curious, adapt to new tech, and be part of the cloud revolution!

Here's your step-by-step guide:

1️⃣ Master the Basics of Cloud Computing
2️⃣ Dive into Virtualization and Containerization
3️⃣ Choose Your Preferred Cloud Platform
4️⃣ Build a Strong Foundation in Networking
5️⃣ Explore Security and Identity Management
6️⃣ Learn Infrastructure as Code (IaC)
7️⃣ Embrace DevOps Practices
8️⃣ Understand Containers and Orchestration
9️⃣ Explore Serverless Computing
😀 Focus on Cloud Security and Compliance
📜 Earn Valuable Certifications


😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
DevOps & Cloud (AWS, AZURE, GCP) Tech Free Learning
Photo
🚀 Top Docker Errors and Their Fixes for DevOps Engineers 🛠

Encountering Docker errors can be frustrating, but fear not! Here are some common Docker errors and their quick fixes to help you keep your containers running smoothly. 🐳

1. Cannot Connect to the Docker Daemon

Error:
Cannot connect to the Docker daemon at unix:///var/run/docker.sock. Is the docker daemon running?


Fix:
- Ensure Docker service is running: sudo systemctl start docker
- Add your user to the Docker group: sudo usermod -aG docker $USER and then restart your terminal or log out and back in.


2. Image Pull Failed

Error:
Error response from daemon: pull access denied for [image], repository does not exist or may require 'docker login'


Fix:
- Verify the image name and tag are correct.
- Log in to Docker Hub if the image is private: docker login


3. Container Exits Immediately

Error:
Exited (0) or Exited (1)


Fix:
- Check the container logs: docker logs [container_id]
- Ensure the command in your Dockerfile or docker run command is correct and doesn’t immediately exit.


4. Port Already in Use

Error:
Error starting userland proxy: listen tcp 0.0.0.0:[port]: bind: address already in use


Fix:
- Find the process using the port: sudo lsof -i :[port]
- Stop the process or use a different port in your Docker command.


5. No Space Left on Device

Error:
no space left on device


Fix:
- Remove unused containers, images, and volumes: docker system prune -a --volumes
- Ensure you have enough disk space on your Docker host.


6. Build Fails Due to Missing Files

Error:
COPY failed: file not found in build context or excluded by .dockerignore: stat [file]: file does not exist


Fix:
- Verify the path in your Dockerfile and ensure the file exists in the build context.
- Check your .dockerignore to ensure necessary files are not being ignored.


7. Permission Denied Errors

Error:
permission denied while trying to connect to the Docker daemon socket


Fix:
- Use sudo if you're not in the Docker group: sudo docker [command]
- Add your user to the Docker group: sudo usermod -aG docker $USER and then restart your terminal or log out and back in.


Keep these handy tips in your toolbox, and Docker errors won’t slow you down! Happy containerizing! 🐳


📱 𝐅𝐨𝐥𝐥𝐨𝐰 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
💥 New DevOps Guide: 100 Common Docker Errors & Their Solutions 💥


🖥 Read the full article here: https://dev.to/prodevopsguytech/100-common-docker-errors-solutions-4le0

Ever faced frustrating Docker issues like failed builds, daemon errors, or container networking problems? 🐳 Say no more! In my latest article, I’ve compiled 100 of the most common Docker errors along with clear, actionable solutions to get you back on track fast! 🚀

Whether you're a beginner or an experienced DevOps engineer, this guide will help you troubleshoot everything from:

🔹 Docker push failures with "no basic auth credentials"
🔹 Docker daemon connection issues
🔹 Mount path errors
🔹 OCI runtime create failures
🔹 And many more common problems!

⚙️ If you’re working with containers, this resource is a must-have in your toolkit. Bookmark it, share it with your team, and save hours of debugging!

Stay ahead in your DevOps journey with practical insights and real-world fixes!


📱 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝐟𝐨𝐫 𝐦𝐨𝐫𝐞 𝐬𝐮𝐜𝐡 𝐜𝐨𝐧𝐭𝐞𝐧𝐭 𝐚𝐫𝐨𝐮𝐧𝐝 𝐜𝐥𝐨𝐮𝐝 & 𝐃𝐞𝐯𝐎𝐩𝐬!!! // 𝐉𝐨𝐢𝐧 𝐟𝐨𝐫 𝐃𝐞𝐯𝐎𝐩𝐬 𝐃𝐎𝐂𝐬: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM