Introduction to Dockers and Containers
Introduction to Dockers and Containers
Unravel the mysteries of Docker and see how containers make creating and deploying applications easier for developers.
Join the DZone community and get the full member experience.Join For Free
We have been hearing a lot about Dockers and containers; they are said to be the next big things in the world of technology. I decided to explore the enigma that was Docker and I must say it's really impressive.
What is Docker?
Whenever we have to install software, we have to take care of a lot of things. There are lots of different versions of software available for different operating systems and their different versions. You have to go through the documentation and choose the correct fit for your needs and then run the executive file. Even after that, you may need to complete some other steps before you are able to use that software. Docker runs containers, which contain the software plus the dependencies that the software requires to run. Just use a
docker run command with the name of the image that you want to install, and your software runs in its own container, using its own resources. You do not have to worry which version of the software suits your operating system. I will demonstrate this with an example of a MongoDB installation.
What Is the Importance of Docker for Developers?
Developers can now simply write their code and create an image. This image will contain all the tools needed for the application to run. This image simply needs to be deployed on a production machine which has no prior software installed and the application will run exactly as in the development machine.
So How Do We Use Docker?
Once we have installed Docker on our systems, we go to Docker Hub or some other registry. Search for the software that you want to install. You can then run a PowerShell command
docker run imageName and the software is ready for our use.
Difference Between Containers and Docker
These two terms are used interchangeably a lot, but they mean different things. Containers are self-contained processes which include the running software along with its dependencies. Containers have existed since Linux, but they were not much used then. According to the official Docker website "Docker is a platform for developers and sysadmins to develop, deploy, and run applications with containers."
Advantages of Containers
As mentioned above, running containers simplifies the process of running software and applications. Suppose you have an ASP.NET application. A developer can create an image of the working application. This image will contain the application, the ASP.NET framework, and the dependencies. Now this image can be deployed as a container on a prod machine that needs no other software installed. Whatever is needed for the application to run will be present in the container. The container will run the same on all the systems, so you will no longer have issues like an application that is running on one machine but failing on another.
Containers and Virtual Machines
Containers and virtual machines might look the same, but they are quite different. Containers will contain only the tools that the application needs and it will share the host operating system kernel with other containers. But virtual machines, on the other hand, will have their own fully independent operating systems. Since containers do not have their own full-fledged operating system, they are lighter than virtual machines.
The Docker website explains a Docker engine with the below diagram.
A Docker engine consists of a client and a server. We interact with the server using the Docker CLI which is also the client. The client interacts with the server through the Docker Rest API. The server or Docker daemon is responsible for running the containers. When the user types in a command from the Docker CLI, like
docker run imagename , the request is received by the Docker daemon. The daemon will search for the image locally, and if found, it will run it as a container. Think of an image as an executive file. If the image is not found locally, the daemon will search for it in a registry and then run it as a container.
Now let's start exploring Docker practically. You need Windows 10 Professional or Enterprise version with at least 4 GB RAM to install Docker. Since I didn't have Windows 10 Professional, I created a virtual machine in Azure. Here are the steps:
Go to the Azure portal and click on virtual machine.
Choose a Windows 10 professional machine. Not all VMs support nested virtualization, so I went and selected a VM of size D2s_v3. Selecting a virtual machine with a size that supports nested virtualization is important to run Docker.
Also make sure in inbound and outbound port rules, all connections are allowed through RDP, or else you might not be able to connect through RDP.
If you are trying to access Azure from your office you might run into issues. You might need to contact your system administrator to open up these ports. Once our VM is up and running, we need to install Docker. Go here to install Docker for Windows.
Once you have installed the above software, your system will restart and Docker will ask you to enable Hyper-V. Click "Ok" and restart the system.
By default, Linux containers will be enabled. You can switch containers by clicking on the whale icon.
Go to PowerShell and type
docker run hello-world and press enter. You should see a message, "Hello from Docker" which means Docker has installed correctly.
Read the steps that are mentioned in the screenshot above. This is what we talked about earlier.
It is possible that if you are trying to run Docker from your workplace, you might face some proxy-related issues. You can set your proxy by navigating to Settings.
Demo: Running Your ASP.NET Application as A Container
Create a new ASP.NET MVC core project in Visual Studio 2017. While creating it, make sure you have the "Enable Docker Support" option checked.
I call my app "aspnetapp." Once you enable Docker support, a file called a Dockerfile is created in the solution explorer.
Replace the existing code in this file with the following piece of code:
FROM microsoft/dotnet:sdk AS build-env WORKDIR /app # Copy csproj and restore as distinct layers COPY *.csproj ./ RUN dotnet restore # Copy everything else and build COPY . ./ RUN dotnet publish -c Release -o out # Build runtime image FROM microsoft/dotnet:aspnetcore-runtime WORKDIR /app COPY --from=build-env /app/out . ENTRYPOINT ["dotnet", "aspnetapp.dll"]
Go to PowerShell and navigate to the project directory. Once there run the command:
docker build -t aspnetapp .
Once the project is built, run the following command:
docker run -d -p 8080:80 --name myaspnetapp aspnetapp
Once this is successful, go to localhost:8080 to navigate to the app:
So what happened here? The Dockerfile gave information which is needed for creating an image. For example, it says that the image should be created with the base image as Microsoft/dotnet:aspnetcore. An image for application is created when you run the
build command. If you write the following the command in PowerShell, you see the images listed.
When you run the image using the
docker run command, it runs this image as a container where "myaspnetapp" is the container name and "aspnetapp" is the image name. The
run command instructs it to run on port 8080. So, when you navigate to localhost:8080, you can find your containerized application running. You can check out all the running containers using the commander
docker ps .
For further information on this demo refer the official Docker website here.
Developers can create their images and upload it to a repository. This image can then be run on production machines when the application needs to be activated.
Now let's check out how we can run MongoDB as a container.
Demo: Installing MongoDB
Traditional Way of Installing Software
Now let's install Mongo using traditional methods. If we head over to its documentation, it will list the steps that are required for installing MongoDB including running the executive, setting it up through the installer, etc. Installing MongoDB is a lengthy process.
Now let's see how Docker simplifies this process.
Running MongoDB as A Container
Go to Docker Hub and search for MongoDB.
Before we run the command, click on the whale icon, go to Settings -> Daemon and set the experimental flag as "True."
Once you are done, Docker will restart. You can then type in the following command in PowerShell or command prompt.
docker run --name some-mongo -d mongo:4.1
Here, Docker installs a container with the name "some-mongo." You can give some other name as you please. "Mongo" is the image name and "4.1" is its version or tag.
It says that a newer image has been downloaded.
Let's use the following command to run the downloaded image:
docker run some-mongo
We will get a message "Waiting for connections on port 27017," which means our server is up and running.
So open another instance of PowerShell and run the following command:
docker exec -it some-mongo mongo
Then type in the command:
This shows that there are no databases created yet in our server. We can now proceed with other MongoDB commands here.
So we see Docker simplifies the process of installing the software.
What Happens Behind the Screens?
Your operating system can be divided into two major portions: kernel and userspace. The kernel has control over the hardware. Everything other than the kernel like our applications, OS apps, and libraries that are required by this application fall under userspace. Userspace accesses the hardware through the kernel.
Traditionally, when we install software we simply install the application and use the drivers and the libraries already present in the user space. But now with the containerization approach, when an image is created it will contain the application plus other drivers required for it to run. Thus, an application will be independent of the resources that the operating system provides.
docker run imagename: This will run the specified image. This is equivalent to running an executive in traditional software.
docker -help: This will list down all the docker commands available to you.
docker ps: This will list down the currently running containers. If we type this command in our PowerShell we get the following output:
docker ps -a: This will list all the running and the exited containers.
docker stop containername: This will stop the software.
docker ps -awill list the container as stopped, but
docker pswill not list it.
docker rm containername: This will remove the container. This is like uninstalling software in the traditional sense. Both
docker ps -aand
docker pswill not list it, since the container is removed.
docker images: This will list the downloaded images. Images are like the executives as far as traditional methods of installing software are concerned.
For the more Docker commands visit here.
Upload An Image To Docker Hub
In the first demo above, we created an image for ASP.NET Core which was stored locally. We will now have a look at how to upload images on Docker Hub. First, you need to create a free account on Docker Hub. Then create a repository.
Login from PowerShell as below:
Then run the following commands:
docker tag command, aspnetapp is the image name. Then comes my username/respository:tag.
Once this image is pushed on Docker Hub, you can log in there and you will be able to see the image in a browser.
I hope that this article has taken you one step closer to unraveling the mysteries of Docker. Feel free to reach out to me in case you want to discuss further.
This article was originally published on my website.
Published at DZone with permission of Tuba Mansoor . See the original article here.
Opinions expressed by DZone contributors are their own.