Slack trains machine-learning models on user messages, files and other content without explicit permission. The training is opt-out, meaning your private data will be leeched by default.
it's funny how the conventional wisdom at the end of the last decade was that slack was preferred over other simpler/free alternatives because of its UX. People were hailing it for how simple and intuitive it was to use, etc.
5, 6 years later, it has become a bloated piece of crap riddled with bugs. And the UI changes which come unannounced... it should be a criminal offense to change UI through automated updates.
Anyway, here we are, companies have handed their data to this monster and we'll see how they react when the data gets misused. Hopefully that would be the beginning of the end for it
I miss Slack, though circa several years back. “Just worked,” on most any platform, without the BS or “help”.
Wouldn’t like it now, I’m sure, but haven’t had a chance to use it since I started working for a co who is “all in” on MS, including foisting AI on us.
I am capable of drafting an email or message, bitches. If I am concerned about tone, etc., I’d prefer to employ an actual human I have a close relationship with to review the same.
I have zero desire to be constantly corrected, and there are certain niche scenarios where very minor errors are actually endearing, and indicate enthusiasm.
“Bob, I saw the posting for your role, can you tell me about your avg day?” is effective because it’s honest, coherent, and just excited enough that you made a minor error that slipped through.
When Bob gets 25 of those emails and they all look the same because AI, it’s much harder to make the connection.
It was a the comma splice, wasn't it? Depending on Bob's cohort, he may never notice.
.. and if I was receiving notes and questions about a role, an error like "emails" would earn relegation for sure; so be careful which error you leave in.