![](/static/253f0d9b/assets/icons/icon-96x96.png)
![](https://lemmy.ml/pictrs/image/h1ChnLuBHr.png)
Many people have a warped understanding of what “two factor” means.
They conflate it with devices and they think it means that one of the factors (why one? which one? who knows) needs to be restricted to exactly one device.
What “two factor” really means is that you should have more than one required factor of authentication so that if one is compromised the attackers still can’t get in.
Ideally the factors should be spread across the “something you know” / “something you own” / “something you are” categories to complicate the manner in which they can be compromised.
We can only reliably rememeber a limited amount of passwords, so like it or not we have to use some devices at least some of the time.
The trouble with “something you own” is that it can be lost or damaged or stolen, and if you only have one of it then you’re fucked. So adding some redundancy is not a bad idea.
The larger issue is that everybody is stuck into extremely rigid and outdated mindsets that date back decades. “Two factors” don’t have to be exactly two, and they don’t have to include exactly one password, and so on. It should be fine if you wanted to secure your account with 3 passwords, and should be up to you if one of those password is a barcode tattooed on your taint so you need a mirror and to bend upside down to scan it.
Bottom line, use whatever you want and use your best judgment as to how secure is each factor. If you want to use something that syncs to multiple devices, go ahead. What you should consider is who has access to those devices and how it would affect you if they’re lost or stolen.
Do we really want that?
We have it pretty good right now. I would actually say we’re living in a golden age of desktop Linux: there’s constant innovation, good support, you get to do pretty much everything you need, while flying under the radar.
Linux has won the majority of the industry (servers, mobile etc.) so it’s not like it has anything left to prove.
If it starts getting noticeable on the desktop I fear we’re just gonna get negative attention. Users who take and not contribute, because Windows had taught them to be entitled. Unwanted attention from Microsoft, who I bet are not going to be doing nice things once they start getting paranoid about it.
I really don’t think that large companies like Adobe will care about Linux even at 10% and even if they did, they are a super toxic company nowadays, the least we get to interact with them the better.