Stafford197 wrote: »ZOS said they would reply here this week,
Since it’s pretty late in the day on a Friday, we probably won’t have any feedback until earlier next week.
In that case I'd like to bring up something someone else did above me. Why wait going on 11 years to start handing out actions against accounts then? I will admit I've had my share of spicy RP. I've had my share of gruesome RP. No characters I've ever played use explicit language, or if they do, a lot of the time it's Old English like arse (which is used in-game) or something like "sod" in place of the f-word.ESO_player123 wrote: »@manukartofanu @jaws343
Are you guys being purposefully obtuse? You're saying that we should all be at risk of warning, suspension, or bans because of a computer program that can't tell context deciding private conversation between adults needs auto-moderating?
I'm sure, then, you'd be totally fine if your phone provider was monitoring every single private call you had and threatening to call the police on you if you ever said a bad word or something lewd?
I'm truly struggling to understand why y'all think it's cool for people to have accounts actioned for the HORRIBLE OFFENSE of using colorful language among friends in private instances and settings. So like...what if someone opens up about trauma they've dealt with and say they were S/A'd but use the word grape without the g, and the bot that can't tell the context that word was used in decides it must be a threat or an insult?
What about the issue that's been brought up multiple times with words in different languages being able to have innocent meanings in their native tongue but being offensive in another?
The report feature exists for a reason. No actions should be taken against accounts by a bot. Ever. And nothing anyone can say will change that.
I think your phone provider example is not a good one here. While subscribing to a phone service you do not sign any agreements about what you are allowed to say over the phone. People in this thread posted parts of the ToS that describe the topics/language that is not going to be tolerated. And we all signed it.
If this has always been in the ToS why have I never once, in like 9-ish years of RP, ever been hit by this bot/AI/whatever? Why is it only now being an issue? There are a few potential reasons.
This is all happening and it's a recent addition, thus it only really recently seeming to be such a huge issue.
This is all happening and it's because that computer program is glitching badly somehow. Maybe it existed this whole time and properly flagged things for human review, maybe it was recently added, but either way it's bugging out.
It's not actually happening and people are just stirring the pot trying to create drama where none exists. Though I feel if this were the case, the devs should have been able to quickly be like "Oh nah this isn't actually happening people are just out here spreading worry and chaos all's good" by now.
The fact that we don't have any sort of comment saying it's not an issue is very concerning. It likely means the devs are having to check things, maybe check other things depending on what they found, and potentially check even more branching stuff. Which means something either did go wrong or this was a deliberate decision and now it's a matter of them trying to work on a damage control response.
And I think we all know how long THAT can take.
Also I know we don't sign user agreements for phone service, and honestly I didn't phrase it right. I should have said they monitor all your private calls with an automatic system that can potentially flag you for use of "bad language". It was meant to highlight how ridiculous it is for anyone to think actions made by machines that can't consider context is justifiable. Do those people not realize they could be hit by this same system as anyone else, especially considering the whole accidentally triggering the filter thing? Nearly all of us should realize it really is too sensitive and nothing that trips that easily should be responsible for actioning accounts.
SilverBride wrote: »ESO_player123 wrote: »People in this thread posted parts of the ToS that describe the topics/language that is not going to be tolerated. And we all signed it.
This is true. However this does not mean that monitoring our private conversations is a good idea, and it is certainly not good customer service.
We need some communication about this.
SilverBride wrote: »ESO_player123 wrote: »People in this thread posted parts of the ToS that describe the topics/language that is not going to be tolerated. And we all signed it.
This is true. However this does not mean that monitoring our private conversations is a good idea, and it is certainly not good customer service.
We need some communication about this.
SeaGtGruff wrote: »Stafford197 wrote: »ZOS said they would reply here this week,
Technically, no they didn't.Since it’s pretty late in the day on a Friday, we probably won’t have any feedback until earlier next week.
"Probably" does not mean "definitely."
Basically, Kevin said that, it being late Friday, the soonest they'd be able to give any feedback about people's concerns would be sometime this week.
Furthermore, Kevin isn't in control of whoever would most likely be making any decisions about what said feedback should be, so it isn't fair to Kevin for anyone here to try to hold him to a presumed hard deadline of this week based on when he said they would "probably" have some feedback to give us.
Also, given that ZOS is now owned by Microsoft, ZOS might conceivably need to get feedback from Microsoft before they can give any feedback to us, depending on who is ultimately in control of this particular issue.
SeaGtGruff wrote: »Also, given that ZOS is now owned by Microsoft, ZOS might conceivably need to get feedback from Microsoft before they can give any feedback to us, depending on who is ultimately in control of this particular issue.
LesserCircle wrote: »The annoying part for me is that these horrible things never get big on youtube or social media like they do when it happens with other games, we can complain all we want in the forums but it won't reach anyone outside of our ESO players and even then the forums are a minority too.
CrazyKitty wrote: »Can anecdotally say I was using some less than wholesome language with my friends while playing this weekend and haven't gotten banned, so hopefully we're just looking at some sort of overzealous moderation glitch
even amongst private chats you agreed to a TOS when you started the game, if you're throwing out the real world equivalents of n'wah/s'wit/etc then you're bound to get banned. keep it civil in-game even amongst friends.
That's ridiculous. Zos isn't some moral police over private citizens consenting private chat.
If you want a private chat, don't log into someone else's software in order to have it. When you sign up to the TOS then they apply to the whole game., not just the bits you think they should apply to.
Pardon me, but I thought ESO was rated "M" for mature. No?
Indeed, but all that refers to is mature content in the game, it doesn't give players the freedom to ignore the TOS.
It's not the TOS. It's the Code of Conduct that governs behaviour.
CrazyKitty wrote: »Can anecdotally say I was using some less than wholesome language with my friends while playing this weekend and haven't gotten banned, so hopefully we're just looking at some sort of overzealous moderation glitch
even amongst private chats you agreed to a TOS when you started the game, if you're throwing out the real world equivalents of n'wah/s'wit/etc then you're bound to get banned. keep it civil in-game even amongst friends.
That's ridiculous. Zos isn't some moral police over private citizens consenting private chat.
If you want a private chat, don't log into someone else's software in order to have it. When you sign up to the TOS then they apply to the whole game., not just the bits you think they should apply to.
Pardon me, but I thought ESO was rated "M" for mature. No?
Indeed, but all that refers to is mature content in the game, it doesn't give players the freedom to ignore the TOS.
It's not the TOS. It's the Code of Conduct that governs behaviour.
As I understand it, the Community Code of Conduct governs behaviour on the forums, and the Rules of Conduct are part of the Terms of Service which govern behaviour in the game.
This is actually a very good point no one else has made yet. What's the point of a filter if we aren't allowed to use ""naughty words"" to begin with?Carcamongus wrote: »If it's against the rules to say "vulgar" words, like that one starting with an F, then how come the profanity filter can be turned off? Perhaps we need some clarifications on more than just this rumored bot.
Oh, and by "that one starting with an F", I obviously meant "fadoodle".
Oh yeah I'm aware, I even brought up that point several pages back. A company claiming they have the right to do something in their ToS doesn't magically make illegal things legal. A company can't claim they'll come assault you for doing X thing because assault is illegal and law takes precedence over non-legal rules, such as a ToS.In that case I'd like to bring up something someone else did above me. Why wait going on 11 years to start handing out actions against accounts then? I will admit I've had my share of spicy RP. I've had my share of gruesome RP. No characters I've ever played use explicit language, or if they do, a lot of the time it's Old English like arse (which is used in-game) or something like "sod" in place of the f-word.ESO_player123 wrote: »@manukartofanu @jaws343
Are you guys being purposefully obtuse? You're saying that we should all be at risk of warning, suspension, or bans because of a computer program that can't tell context deciding private conversation between adults needs auto-moderating?
I'm sure, then, you'd be totally fine if your phone provider was monitoring every single private call you had and threatening to call the police on you if you ever said a bad word or something lewd?
I'm truly struggling to understand why y'all think it's cool for people to have accounts actioned for the HORRIBLE OFFENSE of using colorful language among friends in private instances and settings. So like...what if someone opens up about trauma they've dealt with and say they were S/A'd but use the word grape without the g, and the bot that can't tell the context that word was used in decides it must be a threat or an insult?
What about the issue that's been brought up multiple times with words in different languages being able to have innocent meanings in their native tongue but being offensive in another?
The report feature exists for a reason. No actions should be taken against accounts by a bot. Ever. And nothing anyone can say will change that.
I think your phone provider example is not a good one here. While subscribing to a phone service you do not sign any agreements about what you are allowed to say over the phone. People in this thread posted parts of the ToS that describe the topics/language that is not going to be tolerated. And we all signed it.
If this has always been in the ToS why have I never once, in like 9-ish years of RP, ever been hit by this bot/AI/whatever? Why is it only now being an issue? There are a few potential reasons.
This is all happening and it's a recent addition, thus it only really recently seeming to be such a huge issue.
This is all happening and it's because that computer program is glitching badly somehow. Maybe it existed this whole time and properly flagged things for human review, maybe it was recently added, but either way it's bugging out.
It's not actually happening and people are just stirring the pot trying to create drama where none exists. Though I feel if this were the case, the devs should have been able to quickly be like "Oh nah this isn't actually happening people are just out here spreading worry and chaos all's good" by now.
The fact that we don't have any sort of comment saying it's not an issue is very concerning. It likely means the devs are having to check things, maybe check other things depending on what they found, and potentially check even more branching stuff. Which means something either did go wrong or this was a deliberate decision and now it's a matter of them trying to work on a damage control response.
And I think we all know how long THAT can take.
Also I know we don't sign user agreements for phone service, and honestly I didn't phrase it right. I should have said they monitor all your private calls with an automatic system that can potentially flag you for use of "bad language". It was meant to highlight how ridiculous it is for anyone to think actions made by machines that can't consider context is justifiable. Do those people not realize they could be hit by this same system as anyone else, especially considering the whole accidentally triggering the filter thing? Nearly all of us should realize it really is too sensitive and nothing that trips that easily should be responsible for actioning accounts.
Nobody can break the law to enforce a rule. Whatever rules they have in the tos, they can't just do whatever they like to enforce them. So the usage of data needs to be within the terms of the tos and of the various privacy regimes they operate under.
This is actually a very good point no one else has made yet. What's the point of a filter if we aren't allowed to use ""naughty words"" to begin with?Carcamongus wrote: »If it's against the rules to say "vulgar" words, like that one starting with an F, then how come the profanity filter can be turned off? Perhaps we need some clarifications on more than just this rumored bot.
Oh, and by "that one starting with an F", I obviously meant "fadoodle".Oh yeah I'm aware, I even brought up that point several pages back. A company claiming they have the right to do something in their ToS doesn't magically make illegal things legal. A company can't claim they'll come assault you for doing X thing because assault is illegal and law takes precedence over non-legal rules, such as a ToS.In that case I'd like to bring up something someone else did above me. Why wait going on 11 years to start handing out actions against accounts then? I will admit I've had my share of spicy RP. I've had my share of gruesome RP. No characters I've ever played use explicit language, or if they do, a lot of the time it's Old English like arse (which is used in-game) or something like "sod" in place of the f-word.ESO_player123 wrote: »@manukartofanu @jaws343
Are you guys being purposefully obtuse? You're saying that we should all be at risk of warning, suspension, or bans because of a computer program that can't tell context deciding private conversation between adults needs auto-moderating?
I'm sure, then, you'd be totally fine if your phone provider was monitoring every single private call you had and threatening to call the police on you if you ever said a bad word or something lewd?
I'm truly struggling to understand why y'all think it's cool for people to have accounts actioned for the HORRIBLE OFFENSE of using colorful language among friends in private instances and settings. So like...what if someone opens up about trauma they've dealt with and say they were S/A'd but use the word grape without the g, and the bot that can't tell the context that word was used in decides it must be a threat or an insult?
What about the issue that's been brought up multiple times with words in different languages being able to have innocent meanings in their native tongue but being offensive in another?
The report feature exists for a reason. No actions should be taken against accounts by a bot. Ever. And nothing anyone can say will change that.
I think your phone provider example is not a good one here. While subscribing to a phone service you do not sign any agreements about what you are allowed to say over the phone. People in this thread posted parts of the ToS that describe the topics/language that is not going to be tolerated. And we all signed it.
If this has always been in the ToS why have I never once, in like 9-ish years of RP, ever been hit by this bot/AI/whatever? Why is it only now being an issue? There are a few potential reasons.
This is all happening and it's a recent addition, thus it only really recently seeming to be such a huge issue.
This is all happening and it's because that computer program is glitching badly somehow. Maybe it existed this whole time and properly flagged things for human review, maybe it was recently added, but either way it's bugging out.
It's not actually happening and people are just stirring the pot trying to create drama where none exists. Though I feel if this were the case, the devs should have been able to quickly be like "Oh nah this isn't actually happening people are just out here spreading worry and chaos all's good" by now.
The fact that we don't have any sort of comment saying it's not an issue is very concerning. It likely means the devs are having to check things, maybe check other things depending on what they found, and potentially check even more branching stuff. Which means something either did go wrong or this was a deliberate decision and now it's a matter of them trying to work on a damage control response.
And I think we all know how long THAT can take.
Also I know we don't sign user agreements for phone service, and honestly I didn't phrase it right. I should have said they monitor all your private calls with an automatic system that can potentially flag you for use of "bad language". It was meant to highlight how ridiculous it is for anyone to think actions made by machines that can't consider context is justifiable. Do those people not realize they could be hit by this same system as anyone else, especially considering the whole accidentally triggering the filter thing? Nearly all of us should realize it really is too sensitive and nothing that trips that easily should be responsible for actioning accounts.
Nobody can break the law to enforce a rule. Whatever rules they have in the tos, they can't just do whatever they like to enforce them. So the usage of data needs to be within the terms of the tos and of the various privacy regimes they operate under.
LesserCircle wrote: »The annoying part for me is that these horrible things never get big on youtube or social media like they do when it happens with other games, we can complain all we want in the forums but it won't reach anyone outside of our ESO players and even then the forums are a minority too.
That said, the likelihood that what they're doing is illegal is miniscule, given how many lawyers versed in the relevant international laws are likely to have written it.
Defending a TOS against claims that it's illegal is far more expensive than paying a lawyer to write a legal one.
@manukartofanu @jaws343
Are you guys being purposefully obtuse? You're saying that we should all be at risk of warning, suspension, or bans because of a computer program that can't tell context deciding private conversation between adults needs auto-moderating?
I'm sure, then, you'd be totally fine if your phone provider was monitoring every single private call you had and threatening to call the police on you if you ever said a bad word or something lewd?
I'm truly struggling to understand why y'all think it's cool for people to have accounts actioned for the HORRIBLE OFFENSE of using colorful language among friends in private instances and settings. So like...what if someone opens up about trauma they've dealt with and say they were S/A'd but use the word grape without the g, and the bot that can't tell the context that word was used in decides it must be a threat or an insult?
What about the issue that's been brought up multiple times with words in different languages being able to have innocent meanings in their native tongue but being offensive in another?
The report feature exists for a reason. No actions should be taken against accounts by a bot. Ever. And nothing anyone can say will change that.
Four_Fingers wrote: »Just a quick note, all phone providers have a TOS as well which nobody reads when they sign their contract or pay their first bill. The TOS usually contains info that your phone may be monitored for purposes of law enforcement, usually taking a court order. Not saying that is relevant to this subject but does exist.
karthrag_inak wrote: »Khajiit wagers the AI is ignorant of Ta'agra
To be honest, after reading this thread, I’m not so sure that absolutely everyone here wants a non-toxic community.I think no one here wants a toxic community.
I haven’t received any auto-bans either, although I often run into the filter. This especially happens with certain combinations of letters and numbers when I’m writing about prices in the game. I have no idea what they mean, but sometimes several messages in a row get filtered while I try to figure out how to phrase it so that the filter allows the correct number to be shown. If people are really getting banned because of the auto-filter, then that’s a problem, but it doesn’t seem to be the case here, since the complaints are about AI, and the auto-filter isn’t AI.That said, I have not been affected by an autoban yet, maybe it's because of my writing style, maybe because I'm in EU, I don't know, but I have seen the autocensor that blurrs certain words out. From what I've seen so far, it's so sensitive, it's like it was made for literal children! A censor filter that filters so strictly that people couldn't even talk about the content of a quest without the risk of a ban is a failure, plain and simple.
I’m working on training LLMs for specific work-related tasks. When it comes to content moderation, this is exactly where LLMs can perform much faster and better than any moderation team. LLMs are essentially designed to work with text. The only question is how they are set up. If what people here understand as AI is actually an LLM, then we should be glad that it's being fine-tuned. Those who deserve to be banned will be banned, and those who know how to communicate properly, even in roleplay, have nothing to worry about once the model is tuned correctly. The issue of context isn’t even relevant anymore. LLMs are now capable of recognizing any context and explaining to you in plain human language what was said and how. What is appropriate or inappropriate is a matter of how the model is set up, not about context.And yes, of course I understand that real mods cost money and and automatic moderation would be cheaper and easier (although I'd say that decent moderation is a part of running a game/service online that should certainly be considered a part of the normal game maintenance, and I'd think that ZOS, as a big company, would take that into consideration in their calculations beforehand). But what is it worth if it's not functioning correctly and angers the playerbase?
Indeed, I checked, and there is such a button. It’s possible that if you block the person first, the button disappears. I don’t remember exactly what the issue was now, I just remember that I wasn’t able to report someone during a card game. I remember the first part: to report, you have to surrender, and I didn’t want to do that, so I blocked the person instead. Why it became difficult to report after that, I don’t remember.I'm on PC and all I have to do is right-click on a person's name in the chat, then a menu pops up, including "whisper", "group invite", "spamlist", "report", etc.
I see. In other projects, if a report leads to a ban, there is an automatic notification. I think that’s the right approach, as it lets you know whether it’s even worth bothering with reporting and whether such tickets are actually being reviewed.You usually get a message that your report was received, but they won't tell you about the outcome. At least that's the way it had been a few years ago (I don't report often, only if people throw around racist, homophobic or other slurs in general chat - because I indeed don't want this toxicity in this game).
manukartofanu wrote: »If what people here understand as AI is actually an LLM, then we should be glad that it's being fine-tuned. Those who deserve to be banned will be banned, and those who know how to communicate properly, even in roleplay, have nothing to worry about once the model is tuned correctly. The issue of context isn’t even relevant anymore. LLMs are now capable of recognizing any context and explaining to you in plain human language what was said and how. What is appropriate or inappropriate is a matter of how the model is set up, not about context.