DevOps & Cloud (AWS, AZURE, GCP) Tech Free Learning
15.9K subscribers
1.33K photos
14 videos
501 files
1.28K links
https://projects.prodevopsguytech.com // https://blog.prodevopsguytech.com

• We post Daily Trending DevOps/Cloud content
• All DevOps related Code & Scripts uploaded
• DevOps/Cloud Job Related Posts
• Real-time Interview questions & preparation guides
Download Telegram
📍 What strategies would you employ to optimize the performance of your Jenkins pipeline? 📍

Based on the above question, you could ask the interviewer:
Can you specify which aspects of the Jenkins pipeline are most critical to optimize, such as build times, resource allocation, or dependency management?

The design factors for optimizing Jenkins Pipeline performance and reducing build times would include:
🔢 Parallel Execution:
Leverage parallel execution in Jenkins pipelines. This means designing the pipeline to allow multiple stages or steps to run concurrently rather than sequentially, significantly reducing total execution time for independent tasks.
🔢 Agent and Workspace Efficiency:
Focus on optimizing agent and workspace efficiency. This involves configuring pipelines to use lightweight executors, like Docker agents, and implementing practices to reuse workspaces effectively, which minimizes setup and teardown times.
🔢 Optimize Build Environment:
Ensure the build environment is optimized. This includes selecting high-performance hardware, minimizing network latency, particularly in distributed setups, and choosing efficient build tools and compilers.
🔢 Efficient Retrieval Methods for Source Code:
To minimize checkout times, implement efficient source code retrieval methods, such as local shallow cloning and caching repositories, reducing the time spent fetching code from remote sources.
🔢 Artifact Management:
Effective artifact management is another key area. Utilize artifact repositories and optimize artifact storage and retrieval strategies, such as uploading only deltas or employing parallel downloads.
🔢 Pipeline Caching:
Incorporate pipeline caching to avoid redoing work. By caching dependencies or build outputs at certain stages, the pipeline can reuse previously computed results, which is especially beneficial for dependency-heavy builds.
🔢 Use of Plugins and Tools:
Utilizing Jenkins plugins and external tools effectively is crucial. Employ plugins like Pipeline Utility Steps and Timestamper to optimize performance and manage the pipeline more efficiently.
🔢 Review and Refine Regularly:
Believe in continuous improvement. Regularly reviewing build times and performance metrics helps identify bottlenecks, allowing for the ongoing refinement of pipelines.

➡️ Interviewer expectation:
🔢 Show that you understand the core concepts of Jenkins Pipeline and CI/CD processes.
🔢 Explain how each point contributes to the optimization and efficiency of the pipeline.
🔢 Provide examples from your experiences applying these strategies to solve real-world problems. If you haven't had direct experience, discuss how you would implement these strategies in a hypothetical scenario.


✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
➡️ 𝗠𝗮𝘀𝘁𝗲𝗿𝗶𝗻𝗴 𝗗𝗲𝘃𝗢𝗽𝘀: 𝗔 𝗩𝗶𝘀𝘂𝗮𝗹 𝗚𝘂𝗶𝗱𝗲

This roadmap provides a great introduction to DevOps and the different technologies and concepts you'll need to master. Whether you're a seasoned developer or just starting out in IT operations, this roadmap can help you on your DevOps journey.

➡️𝗧𝗵𝗲 𝗕𝗮𝘀𝗶𝗰𝘀

The roadmap starts with the foundational concepts of DevOps, including the DevOps culture, continuous delivery and deployment, and blue-green deployments. It's important to understand these core principles before you dive into the more technical aspects of DevOps.

➡️𝗘𝘀𝘀𝗲𝗻𝘁𝗶𝗮𝗹 𝗧𝗲𝗰𝗵𝗻𝗼𝗹𝗼𝗴𝗶𝗲𝘀

Next, the roadmap dives into some of the essential technologies used in DevOps, such as:

Version control systems: Git is the most popular version control system, and it's essential for tracking changes to code and infrastructure.

Containerization and orchestration: Docker and Kubernetes are the leading containerization and orchestration technologies, respectively. They allow you to package applications into portable containers and then deploy and manage them at scale.

Infrastructure monitoring: Prometheus and Elasticsearch are popular tools for monitoring infrastructure performance and health.

Cloud platforms: AWS, Azure, and Google Cloud Platform are the major cloud providers, and they offer a wide range of DevOps tools and services.

Configuration management: Ansible and Puppet are popular tools for managing infrastructure configuration. They allow you to automate the provisioning and configuration of servers and other infrastructure components.

Continuous integration and continuous delivery (CI/CD): Jenkins is a popular CI/CD tool that helps you automate the software build, test, and deployment process.

➡️𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗼𝘂𝘀 𝗜𝗺𝗽𝗿𝗼𝘃𝗲𝗺𝗲𝗻𝘁

The roadmap also emphasizes the importance of continuous improvement and learning. DevOps is a journey, not a destination, so it's important to stay up-to-date on the latest trends and technologies.

➡️𝗚𝗲𝘁𝘁𝗶𝗻𝗴 𝗦𝘁𝗮𝗿𝘁𝗲𝗱

If you're interested in learning more about DevOps, there are many resources available online and in libraries. You can also start by taking a DevOps course or attending a DevOps meet-up.


🎄 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
1709007585362.gif
2.3 MB
📣 Kubernetes Services and Ingress Demystified.

Understanding how to expose Kubernetes services is key for building robust applications.
In this post,We'll walk through the main service types and ingress in Kubernetes.


Services in Kubernetes allow pods to communicate with each other and provide a stable endpoint that doesn't change as pods are created or deleted. There are several types of services:

➡️ClusterIP: Exposes the service on a cluster-internal IP only. This makes the service only reachable from within the cluster.

➡️NodePort: Exposes the service on each Node's IP at a static port. You can contact the NodePort service from outside the cluster by requesting <NodeIP>:<NodePort>.

➡️LoadBalancer: Creates an external load balancer and assigns a fixed, external IP to the service. The load balancer routes to NodePorts of cluster nodes.

➡️ExternalName: Maps the service to the contents of the externalName field (e.g. foo.bar.example.com), by returning a CNAME record.

Ingress is another Kubernetes resource that allows you to route external traffic to services based on HTTP rules. Ingress exposes HTTP and HTTPS routes from outside the cluster to services within the cluster. Traffic routing is controlled by rules defined on the Ingress resource.


😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
1709468181112.gif
4.3 MB
#️⃣ 𝐄𝐱𝐜𝐢𝐭𝐞𝐝 𝐚𝐛𝐨𝐮𝐭 𝐃𝐞𝐯𝐎𝐩𝐬?

🔣DevOps has revolutionised the way modern software is built, deployed, and managed. It's a culture, a set of practices, and a mindset that emphasizes collaboration, automation, and continuous feedback.

Whether you're a seasoned professional or just starting out, mastering DevOps principles and tools is essential for driving efficiency, reliability, and innovation in your projects.



✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🤦‍♀️ Knowing this beforehand , rolling updates would have been a cake walk to understand & not to surprise why most preferred deployment type by Big Tech Giants 🚀🚀


🔎 Why are different types of deployment strategies used ?

Deployments types are decided based on various factors. including client unique requirements , flexibility , scalability , managing risk and much more !

🪝𝙇𝙚𝙩’𝙨 𝙪𝙣𝙙𝙚𝙧𝙨𝙩𝙖𝙣𝙙 𝙖𝙗𝙤𝙪𝙩 𝙍𝙤𝙡𝙡𝙞𝙣𝙜 𝙐𝙥𝙙𝙖𝙩𝙚𝙨 𝙖𝙨 𝙙𝙚𝙥𝙡𝙤𝙮𝙢𝙚𝙣𝙩 𝙨𝙩𝙧𝙖𝙩𝙚𝙜𝙮

Let’s picturize it as you have multiple servers , but you need to upgrade their versions , at the same time make sure user experience is not hampered by, rolling update comes as saviour , one by one you can upgrade to required/ new version test it and bring it online , similarly you follow it for remaining ones

You can incrementally bring all your servers to required state

📌 It proves to be a game changer as when one of your server is updating/ upgrading rest of the services are in places and users experience no downtime and gradually same follows till whole system is updated


🎄 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
Tech News ‼️

➡️ Accenture Opens Tech Centre In Odisha's Bhubaneswar.

➡️ Apple creating 2nm chip for next iPhones, possibly most advanced SoC ever.

➡️ Google Chrome now shows search suggestions even if you are on poor internet.

➡️ WhatsApp will now let Android users search chats, photos by date.

➡️ India ranks amongst the top five most breached countries in 2023.

➡️ Harpic launches new app to help locate public toilets in India.

➡️Daily Tech News WhatsApp Group
https://chat.whatsapp.com/FndTPrJEkbq5RlKQE3jGp3
Please open Telegram to view this post
VIEW IN TELEGRAM
➡️ Understanding Container Runtimes in Kubernetes 👈

A container runtime in Kubernetes is the software component responsible for managing the lifecycle of individual containers within a pod. It's the engine that executes the commands and manages the processes within the container environment.

➡️ What it does:

➡️Creates and starts containers: Based on instructions from the kubelet (the Kubernetes agent on each node), the container runtime pulls the container image, sets up the necessary resources, and fires up the container process.

➡️Manages container resources: It allocates CPU, memory, and other resources as specified in the pod definition, ensuring each container gets its fair share.

➡️Monitors and manages container health: It keeps an eye on the container's health and restarts it if it crashes or becomes unresponsive.

➡️Stops and removes containers: When a container is no longer needed, the runtime gracefully stops it and cleans up its resources.

➡️ Why it's important:

➡️Isolation: Container runtimes create isolated environments for each container, ensuring applications don't interfere with each other or the host system.

➡️Security: They enforce security policies and resource limitations, providing a more secure environment for containerized applications.

➡️Portability: Container runtimes adhere to industry standards, allowing containers to be easily moved between different platforms and cloud providers.

➡️Common container runtimes in Kubernetes.

- containerd
- CRI-O
- Docker Engine
- Mirantis Container Runtime


✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
1707395410428.gif
1.3 MB
🔖 𝐇𝐨𝐰 𝐭𝐨 𝐀𝐮𝐭𝐨𝐦𝐚𝐭𝐞 𝐃𝐞𝐯𝐎𝐩𝐬 𝐏𝐢𝐩𝐞𝐥𝐢𝐧𝐞 𝐔𝐬𝐢𝐧𝐠 𝐉𝐞𝐧𝐤𝐢𝐧𝐬!

Jenkins is a popular automation server that can be used to automate the CI/CD pipeline. In this post we will learn how to use Jenkins to automate the following steps:

Checkout code from version control: Jenkins can be configured to automatically checkout code from a version control system, such as GitHub.
Compile code: Jenkins can use a variety of build tools, such as Maven or Gradle, to compile the code.
Run unit tests: Jenkins can run unit tests to ensure that the code is working properly.
Build a Docker image: Jenkins can build a Docker image from the compiled code.
Push Docker image to registry: Jenkins can push the Docker image to a Docker registry, such as Docker Hub.
Deploy to Kubernetes: Jenkins can notify Kubernetes of the new Docker image and deploy it to a Kubernetes cluster.


🎄 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy & @devopsdocs 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
DevOps Tools - Setup, Installations, Guides ⚙️

🔗Link: https://github.com/NotHarshhaa/DevOps_Setup-Installations

We add daily Tools Setup, Installations, Guides with each and every commands with clear explanation

💎 Now added : Kubernetes, Jenkins, Ansible, Terraform, AWS, Azure, Linux
More added daily so fork the repository for updates

🐱 Follow me on GitHub for more DevOps/Cloud related sources


✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
DevOps & Cloud (AWS, AZURE, GCP) Tech Free Learning
DevOps Tools - Setup, Installations, Guides ⚙️ 🔗Link: https://github.com/NotHarshhaa/DevOps_Setup-Installations We add daily Tools Setup, Installations, Guides with each and every commands with clear explanation 💎 Now added : Kubernetes, Jenkins, Ansible…
➡️ Repository Update 📣


➡️Added Jenkins installation and setup and integrating for various types of tools to support your continuous integration and continuous deployment (CI/CD) pipelines

➡️ Here are tools content:

🔢. Version Control & Build Tools
🔢. Testing Frameworks Tools
🔢. Containerization Tools
🔢. Infrastructure as Code (IaC) Tools
🔢. Code Quality and Analysis Tools
🔢. Deployment Tools
🔢. Notification and Collaboration Tools

✈️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
💡 Kubernetes vs Docker: What's The Difference?

➡️Docker and Kubernetes are the most common names that one might hear in the field of container technology.

➡️Docker is a runtime and containerization platform that was first introduced in 2013 and brought about a microservices-based computing model.

➡️Kubernetes is a platform that manages and runs containers from multiple container runtimes and supports various container runtimes, including Docker.


❤️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
📢 Node Affinity

🔣 In Kubernetes, node affinity is a feature that allows you to control which nodes your pods are scheduled on based on labels on the node. This can be useful for various reasons such as optimizing performance, ensuring certain workloads are placed on specific nodes, or ensuring high availability by spreading pods across different nodes.

There are two types of node affinity:
🔎 Required Node Affinity: This type specifies that a pod should only be scheduled on nodes that meet the specified node selector requirements.

🔎 Preferred Node Affinity: This type specifies that a pod prefers to be scheduled on nodes that meet the specified node selector requirements, but it can still be scheduled on nodes that do not meet those requirements.


🌐 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
🔴 A nice cheat sheet of different cloud services.

➡️ What’s included?
- AWS, Azure, Google Cloud, Oracle Cloud, Alibaba Cloud
- Cloud servers
- Databases
- Message queues and streaming platforms
- Load balancing, DNS routing software
- Security
- Monitoring



😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
https://blog.prodevopsguy.xyz/the-ultimate-devops-bootcamp-2024-pack-by-prodevopsguy

⚠️ Note: Anyone Interested, can open the Blog 🌐, share it to your friends and colleagues.


🔵 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
1709611305572.gif
8.2 MB
🔣 Docker is a powerful tool that allows you to package your application and its dependencies into a standardized unit called a container.

👉 𝗛𝗲𝗿𝗲'𝘀 𝗮 𝗯𝗿𝗲𝗮𝗸𝗱𝗼𝘄𝗻 𝗼𝗳 𝗵𝗼𝘄 𝗗𝗼𝗰𝗸𝗲𝗿 𝘄𝗼𝗿𝗸𝘀:

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝗰𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿𝘀 are self-contained units that package your application code, runtime, system tools, settings, and libraries. They're like tiny virtual machines, but they share the underlying operating system kernel with other containers, making them much more lightweight and efficient.

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝗶𝗺𝗮𝗴𝗲𝘀 are blueprints that contain instructions for creating containers. They're kind of like recipes that tell Docker how to build a container with all the necessary ingredients. You can find and share images on public registries like Docker Hub, or create your custom images.

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝗱𝗮𝗲𝗺𝗼𝗻 (𝗱𝗼𝗰𝗸𝗲𝗿𝗱) is the engine that builds, runs, and manages Docker containers. It's the behind-the-scenes workhorse that makes everything tick.

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝗰𝗹𝗶𝗲𝗻𝘁 (𝗱𝗼𝗰𝗸𝗲𝗿 𝗖𝗟𝗜) is the tool you use to interact with the Docker daemon. It allows you to build, run, stop, and manage your containers using commands or a graphical user interface (GUI).

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝗿𝗲𝗴𝗶𝘀𝘁𝗿𝗶𝗲𝘀 are repositories that store and share Docker images. Think of them as libraries for Docker images. Docker Hub is the most popular public registry, but there are also private registries that organizations can use to store their custom images.

➡️𝗗𝗼𝗰𝗸𝗲𝗿 𝘀𝘁𝗼𝗿𝗮𝗴𝗲 𝗱𝗿𝗶𝘃𝗲𝗿𝘀 manage how data is stored and persisted within containers. The default storage driver is overlay2, but there are other options available, such as aufs, zfs, and btrfs.

➡️𝗖𝗼𝗻𝘁𝗮𝗶𝗻𝗲𝗿 𝗼𝗿𝗰𝗵𝗲𝘀𝘁𝗿𝗮𝘁𝗼𝗿𝘀 like Kubernetes and Swarm are tools that help you manage and scale large deployments of Docker containers. They take care of provisioning, scheduling, and healing your containers, making it easier to run complex applications in production.


😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
📌 https://harshhaa.hashnode.dev/real-time-devops-project-deploy-to-kubernetes-using-jenkins-end-to-end-devops-project-cicd

🔗 More DevOps Projects : HERE

🔥🔥🔥🔥🔥🔥🔥🔥

Follow 🍩 Like 👍 Share 👍 Comment Your thoughts 💬

⭐️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy & @devopsdocs 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM
👍 Multi-Stage Build Images used in CICD

➡️ Every microservice should be its own separate container. If you only use a single-stage Docker build, you’re probably missing out on some powerful features of the build process. On the other hand, a multi-stage Docker build has many advantages over a single-stage build for deploying microservices.

➡️ Some Advantages are :
- Optimizes the overall size of the Docker image
- Removes the burden of creating multiple Dockerfiles for different stages
- Easy to debug a particular build stage
- Able to use the previous stage as a new stage in the new environment
- Ability to use the cached image to make the overall process quicker
- Reduces the risk of vulnerabilities found as the image size becomes smaller with multi-stage builds



😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
1708757232781.gif
285.8 KB
🐬 Docker Container Commands for a DevOps Engineer

The below illustration shows some common container commands and their syntax 👇

1. 🏗 Creates a container from an image
docker run -it --name nginx nginx


2. 🚀 Begins a Docker container
docker start nginx


3. 🔄 Restarts a Docker container.
docker restart nginx


4. Temporarily halts a container.
docker pause nginx


5. ▶️ Resumes a paused container.
docker unpause nginx


6. 🛑 Ends a running Docker container.
docker stop nginx


7. Forcefully stops a running container
docker kill nginx


8. 📊 Lists Docker containers.
docker ps


9. 🖥 Accesses a container's shell.
docker exec -it nginx /bin/bash


10. 📝 Connects to a running container.
docker attach nginx


11. 📜 Views container logs.
docker logs nginx


12. 🔄 Change a container's name.
docker rename old-name new-name


13. 🔍 Retrieves container info.
docker inspect nginx


14. 📂 Copies files to/from a container.
docker cp nginx:/container-path/file.txt /local-path


15. 🗑 Deletes a container.
docker rm nginx


These container commands are essential for managing containerized applications, whether for development, testing, or production deployment, as they enable efficient control and manipulation of container instances.


😎 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!! // Join for DevOps DOCs: @devopsdocs
Please open Telegram to view this post
VIEW IN TELEGRAM
From 'It works on my machine!' to 'It works on my container!' but hey at least it works!


✔️ 𝗙𝗼𝗹𝗹𝗼𝘄 @prodevopsguy 𝗳𝗼𝗿 𝗺𝗼𝗿𝗲 𝘀𝘂𝗰𝗵 𝗰𝗼𝗻𝘁𝗲𝗻𝘁 𝗮𝗿𝗼𝘂𝗻𝗱 𝗰𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀!!!
Please open Telegram to view this post
VIEW IN TELEGRAM