Mainly GTG response time and latency. For watching movies it’s generally not a problem, but when it comes to playing games with a mouse, latency can be a huge issue, and bad GTG response time leads to smearing.
But yeah, 4x the price is ridiculous.
Mainly GTG response time and latency. For watching movies it’s generally not a problem, but when it comes to playing games with a mouse, latency can be a huge issue, and bad GTG response time leads to smearing.
But yeah, 4x the price is ridiculous.
Am I too 1Gb/s fiber connected to understand that?
If you know how to use git, you will know how to use docker (provided you know what you want to do). They are completely different programs, yet you can quickly grasp the other instinctively.
Now, Photoshop and Blender - they are also different programs, but if you know Photoshop, you still need to relearn Blender’s interface completely.
This is why I prefer terminal programs in general. Unless it’s more convenient to use GUi, i.e. Drag&Drop file manager, some git tools etc.
Learn it first.
I almost exclusively use it with my own Dockerfiles, which gives me the same flexibility I would have by just using VM, with all the benefits of being containerized and reproducible. The exceptions are images of utility stuff, like databases, reverse proxy (I use caddy btw) etc.
Without docker, hosting everything was a mess. After a month I would forget about important things I did, and if I had to do that again, I would need to basically relearn what I found out then.
If you write a Dockerfile, every configuration you did is either reflected by the bash command or adding files from the project directory to the image. You can just look at the Dockerfile and see all the configurations made to base Debian image.
Additionally with docker-compose you can use multiple containers per project with proper networking and DNS resolution between containers by their service names. Quite useful if your project sets up a few different services that communicate with each other.
Thanks to that it’s trivial to host multiple projects using for example different PHP versions for each of them.
And I haven’t even mentioned yet the best thing about docker - if you’re a developer, you can be sure that the app will run exactly the same on your machine and on the server. You can have development versions of images that extend the production image by using Dockerfile stages. You can develop a dev version with full debug/tooling support and then use a clean prod image on the server.
I mean, GPL guarantees code remains open and free. If they release an app based on the source code licensed under GPL, they have to give a source code along with essential build instructions to anyone who is using it, and then you can do anything with that code, including sharing, compiling, and distributing that app, provided it’s under GPL license.
Edit: I see it’s licensed under GPL 3.0, so no worries.
I think ARM is their end goal, it’s really the only option for a handheld console, as today ARM is the only way you’ll get enough performance/power rate to make it both good on battery with good enough performance.
Win-win for everyone if they invest in an open source x86 to ARM project, similar like they did with Wine.
Why do you need Windows VM for developing GUI apps? Last time I used Visual Studio to make GUI app I almost gave up programming, because of how code-generation dependent it was.
For C# you have AvaloniaUI. For cpp you have countless multi-platform GUI toolkits, same for rust, Java has its own toolkits (multi-platform), and finally you can make an Electron/Tauri app.
The way for your desktop to communicate with the hardware.
It used to be X11 - A server-client architecture, which meant your desktop was effectively just a client that told the server what to do. The server was the one doing the drawing
Wayland is just a protocol, defining how programs and desktop should communicate with each other - without a middleman that was X11 server. The desktop does the actual drawing here.
The main argument to use password managers to prevent password leaks to all of your services (that you use with the same login/email). You can’t trust any service to store your password securely, therefore you should use different ones everywhere.
Using a password manager gives you the convenience of using one, strong password that’s being used very securely, and mitigating risk of password leaks spreading further.
If you abstract it that way, it by no means eliminates the risk of someone breaking into your database, but makes it harder and from a single entry point, instead of any service that uses your password.
Plus many of those password managers give you an option to use YubiKey for additional security.
Oh and also you won’t ever need to press “forgot password” ever again due to the arbitrary requirements that your password doesn’t pass, so you modify it slightly so it would.
It’s a go to solution for scientific papers.
Have an actual sane developer experience? There is a reason why almost every developer that uses Windows actually uses WSL.
Nah. Nvidia is still Nvidia, but 2 years ago or so they finally gave up and started supporting GBM and even opened part of their driver stack.
Some things, like hardware encoder are even easier to set up than AMD’s counterpart. (Mainly because Nvidia proprietary driver being supported better than AMD’s proprietary driver)
If you’re using Excel so much to the point that certain things can only be done in Excel on Windows, just learn Python.
Is SMTP out of question?
Actually ACL on Windows is very bad. Recursively changing owner of directory can take minutes, same operation on any UNIX-like OS takes seconds.
What? Pipewire has been working out of the box for like 2 years
Yup, if you have SSH service open on port 22, you’re automatically spammed by bots trying to brute force their way onto the system.
AFAIK it’s now 100, so basically unlimited for personal use.
To add to an answer, caddy is better fit for reverse proxy in my opinion. It’s like easier to configure version of nginx. With nginx proxy_pass you also have to configure other headers like x_forwarded_to, and you will also need to do some magic to get websockets working.
And also caddy automatically generates certs using ACME, by default.
You mentioned you changed firewall rules for that device. Any chance you have set outbound rule instead of inbound rule?
Anyway, what’s the output of
ip route
?