That is because windows filesystem is mounted to WSL through NFS and while transferring large files through that is ok, transferring huge amounts of small files is really slow.
That is because windows filesystem is mounted to WSL through NFS and while transferring large files through that is ok, transferring huge amounts of small files is really slow.
You mentioned you changed firewall rules for that device. Any chance you have set outbound rule instead of inbound rule?
Anyway, what’s the output of ip route
?
Ah yes, perfect data format, where markup takes more space than the actual data.
Mainly GTG response time and latency. For watching movies it’s generally not a problem, but when it comes to playing games with a mouse, latency can be a huge issue, and bad GTG response time leads to smearing.
But yeah, 4x the price is ridiculous.
Am I too 1Gb/s fiber connected to understand that?
Better in which way? WSL2 is a VM running ALONGSIDE Windows, not inside. Its performance is basically bare metal. If you have enough RAM, there is no reason to use cygwin instead of WSL2.
If you know how to use git, you will know how to use docker (provided you know what you want to do). They are completely different programs, yet you can quickly grasp the other instinctively.
Now, Photoshop and Blender - they are also different programs, but if you know Photoshop, you still need to relearn Blender’s interface completely.
This is why I prefer terminal programs in general. Unless it’s more convenient to use GUi, i.e. Drag&Drop file manager, some git tools etc.
Learn it first.
I almost exclusively use it with my own Dockerfiles, which gives me the same flexibility I would have by just using VM, with all the benefits of being containerized and reproducible. The exceptions are images of utility stuff, like databases, reverse proxy (I use caddy btw) etc.
Without docker, hosting everything was a mess. After a month I would forget about important things I did, and if I had to do that again, I would need to basically relearn what I found out then.
If you write a Dockerfile, every configuration you did is either reflected by the bash command or adding files from the project directory to the image. You can just look at the Dockerfile and see all the configurations made to base Debian image.
Additionally with docker-compose you can use multiple containers per project with proper networking and DNS resolution between containers by their service names. Quite useful if your project sets up a few different services that communicate with each other.
Thanks to that it’s trivial to host multiple projects using for example different PHP versions for each of them.
And I haven’t even mentioned yet the best thing about docker - if you’re a developer, you can be sure that the app will run exactly the same on your machine and on the server. You can have development versions of images that extend the production image by using Dockerfile stages. You can develop a dev version with full debug/tooling support and then use a clean prod image on the server.
Or - you know - for consistency? In physics kilo, mega etc. are always 10^(3n), but then for some bizarre reason, unit of information uses the same prefixes, but as 2^(10n).
General rule of thumb: Comments say why is it here, not what it does. Code itself should describe what it does.
English is not my first not language. When I write something down in my first language (polish), it feels more like I’m transcribing things I silently say to myself, while with english I’m actually thinking about every word I type.
The funny thing is, the better I am getting at English, making those types of mistakes is getting easier for me.
But idk, this is just my experience.
Because they learned that from hearing, not reading so that makes sense.
Why do you need Windows VM for developing GUI apps? Last time I used Visual Studio to make GUI app I almost gave up programming, because of how code-generation dependent it was.
For C# you have AvaloniaUI. For cpp you have countless multi-platform GUI toolkits, same for rust, Java has its own toolkits (multi-platform), and finally you can make an Electron/Tauri app.
The way for your desktop to communicate with the hardware.
It used to be X11 - A server-client architecture, which meant your desktop was effectively just a client that told the server what to do. The server was the one doing the drawing
Wayland is just a protocol, defining how programs and desktop should communicate with each other - without a middleman that was X11 server. The desktop does the actual drawing here.
The main argument to use password managers to prevent password leaks to all of your services (that you use with the same login/email). You can’t trust any service to store your password securely, therefore you should use different ones everywhere.
Using a password manager gives you the convenience of using one, strong password that’s being used very securely, and mitigating risk of password leaks spreading further.
If you abstract it that way, it by no means eliminates the risk of someone breaking into your database, but makes it harder and from a single entry point, instead of any service that uses your password.
Plus many of those password managers give you an option to use YubiKey for additional security.
Oh and also you won’t ever need to press “forgot password” ever again due to the arbitrary requirements that your password doesn’t pass, so you modify it slightly so it would.
It’s a go to solution for scientific papers.
Seems like CPU-intensive game, so it makes sense.
Simple fails when complex problem arrives. Declarative approach of systemd daemons allows for more versatile solutions in unified format.
Not only OS - written using 3D APIs closed source available only for your OS.
24.04 won’t have Plasma 6, but 24.10 will. In other words, fall 2024.
Or you can use KDE Neon, which is basically Ubuntu LTS, but with the newest Plasma.