I've experienced a similar issue, which started to accure after I've updated my PC using windows update. I've then found this thread in the intel community forum. As it seems there is a bug in a driver for Intel UHD Graphics 4600, which is used by at least some of the CPUs which are used inside you notebook model according to my research.

For my CPU there have been no newer drivers solving this issue, but as a workaround I've been able to find a older version of the driver without the bug on the Intel support site. Note that you may should disable automatic driver updates as described here to avoid having the same issue again if this proves to be the problem.

Problem installing docker on windows 10

I've just tried installing docker on my windows pc, however after the installation, my monitors are not able to work properly anymore. Normally I use 2 monitors that extend each other so I can use both at the same time. After the installtion of docker they started to just duplicate and also had issues with the resolution. When I remove docker and disable hyper-v the problems stop, I do need to use docker however for a current project, so I need to find a workaround to solve these issues. I am using a Lenovo Y50-70 notebook.

I found out that it is truly impossible to have multiple docker files in one folder, but there is a workaround. I have created a separate subfolders with each having its own dockerfile for either windows or linux in it and nothing else. In the YAML file for the pipeline that creates those docker images, I have added a step before docker compose, that moves the dockerfile out of the subfolder and into the main folder where the relevant code files are. It is possible to have only one YAML file and create an IF statement that will select the correct dockerfile based on parameters of the pipeline run.


multiple dockerfiles for one code

I have a python application that I need to put into a docker. Since I need the docker to be able to run on windows servers and linux servers, its best to create two separate docker images, one for each operating system. Since docker is composed using docker files, and each docker image is a derivation of prebuilt docker images and they are already build with different operating systems in them, I need to have different docker files for each pipeline. However, docker does not allow to have multiple docker files and only allows one docker file named "Dockerfile" in a folder with code. How can I create multiple docker files for multiple pipelines?

it seems that the reason why Docker uses much more memory is that it does not allow the operating system to use virtual memory (swap memory). What I didn't see in the original debugging, is that the python program consumes memory of several GB, but windows has automatically moved another several GB of memory to virtual memory to save RAM space. This is however not enabled in docker environment, so the only RAM available to the docker is the one that is specified in the set up phase.



Docker uses way more memory than python program on its own

I have created a program in python that simulates business processes. For extremely large processes, the simulation can run for hours and the memory requirements are in GB. However, if I create a docker container, the docker image consumes cca 3x more memory than the original python program and if I limit the memory of the docker to even 2x the original python memory, the python program inside the docker will fail with out of memory error.


When starting docker a bridge network is automatically started, this network has its own ip and you have to check what is the ip of that network. If you are running it locally and used localhost, the bridge network's adress might be different from localhost and that might be why you weren't able to connect. However, using the containers' name is a much safer way to communicate between servers, docker will choose the correct address.

Docker compose versions are not compatible, because of that they don't support all functions, you have to read the documentation of the version.


First of all, while there is more or less one stable option for Java, there exist many different libraries for JavaScript and some of them don't interact very well with Docker. As such, I recommend using the client library to implement the socket client.

Docker actually allows you to address your components via the names of their containers. So instead of using localhost or other IP addresses to connect to your container, you should specify the name of the container, for example use the address ws://socketserver:4001 if the name of the container that hosts the server is socketserver and it runs on port 4001. Furthermore, in your docker-compose file, you should make sure that there are no conflicts between exposed ports, and that the relevant port on the server is exposed. After that, your components should be able to interact with each other without any issues.



Websocket interactions in Docker

I am trying to get my Websocket server (Java) and client (JS) to interact with each other when they are both running in separate Docker containers. However, no matter what I do, I cannot get the client to even establish a connection to the server. What can be done here?


Subscribe to Docker