Mistral 3 family of models released
343 points
2 hours ago
| 22 comments
| mistral.ai
| HN
barrell
1 hour ago
[-]
I use large language models in http://phrasing.app to format data I can retrieve in a consistent skimmable manner. I switched to mistral-3-medium-0525 a few months back after struggling to get gpt-5 to stop producing gibberish. It's been insanely fast, cheap, reliable, and follows formatting instructions to the letter. I was (and still am) super super impressed. Even if it does not hold up in benchmarks, it still outperformed in practice.

I'm not sure how these new models compare to the biggest and baddest models, but if price, speed, and reliability are a concern for your use cases I cannot recommend Mistral enough.

Very excited to try out these new models! To be fair, mistral-3-medium-0525 still occasionally produces gibberish ~0.1% of my use cases (vs gpt-5's 15% failure rate). Will report back if that goes up or down with these new models

reply
mentalgear
1 minute ago
[-]
Thanks for sharing your use case of the mistral models, which are indeed top-notch ! I had a look at phrasing.app, and while a nice website, I found the copy of "Hand-crafted. Phrasing was designed & developed by humans, for humans." somewhat of a false virtue given your statements here of advanced lllm usage.
reply
mrtksn
50 minutes ago
[-]
Some time ago I canceled all my paid subscriptions to chatbots because they are interchangeable so I just rotate between Grok, ChatGPT, Gemini, Deepseek and Mistral.

On the API side of things my experience is that the model behaving as expected is the greatest feature.

There I also switched to Openrouter instead of paying directly so I can use whatever model fits best.

The recent buzz about ad-based chatbot services is probably because the companies no longer have an edge despite what the benchmarks say, users are noticing it and cancel paid plans. Just today OpenAI offered me 1 month free trial as if I wasn’t using it two months ago. I guess they hope I forget to cancel.

reply
barrell
38 minutes ago
[-]
Yep I spent 3 days optimizing my prompt trying to get gpt-5 to work. Tried a bunch of different models (some Azure some OpenRouter) and got a better success rate with several others without any tailoring of the prompt.

Was really plug and play. There are still small nuances to each one, but compared to a year ago prompts are much more portable

reply
barbazoo
40 minutes ago
[-]
> I guess they hope I forget to cancel.

Business model of most subscription based services.

reply
druskacik
23 minutes ago
[-]
This is my experience as well. Mistral models may not be the best according to benchmarks and I don't use them for personal chats or coding, but for simple tasks with pre-defined scope (such as categorization, summarization, etc.) they are the option I choose. I use mistral-small with batch API and it's probably the best cost-efficient option out there.
reply
metadat
58 minutes ago
[-]
Are you saying gpt-5 produces gibberish 15% of the time? Or are you comparing Mistral gibberish production rate to gpt-5.1's complex task failure rate?

Does Mistral even have a Tool Use model? That would be awesome to have a new coder entrant beyond OpenAI, Anthropic, Grok, and Qwen.

reply
barrell
42 minutes ago
[-]
Yes. I spent about 3 days trying to optimize the prompt to get gpt-5 to not produce gibberish, to no avail. Completions took several minutes, had an above 50% timeout rate (with a 6 minute timeout mind you), and after retrying they still would return gibberish about 15% of the time (12% on one task, 20% on another task).

I then tried multiple models, and they all failed in spectacular ways. Only Grok and Mistral had an acceptable success rate, although Grok did not follow the formatting instructions as well as Mistral.

Phrasing is a language learning application, so the formatting is very complicated, with multiple languages and multiple scripts intertwined with markdown formatting. I do include dozens of examples in the prompts, but it's something many models struggle with.

This was a few months ago, so to be fair, it's possible gpt-5.1 or gemini-3 or the new deepseek model may have caught up. I have not had the time or need to compare, as Mistral has been sufficient for my use cases.

I mean, I'd love to get that 0.1% error rate down, but there have always more pressing issues XD

reply
data-ottawa
1 minute ago
[-]
With gpt5 did you try adjusting the reasoning level to "minimal"?

I tried using it for a very small and quick summarization task that needed low latency and any level above that took several seconds to get a response. Using minimal brought that down significantly.

Weirdly gpt5's reasoning levels don't map to the OpenAI api level reasoning effort levels.

reply
barbazoo
39 minutes ago
[-]
Hard to gauge what gibberish is without an example of the data and what you prompted the LLM with.
reply
barrell
30 minutes ago
[-]
If you wanted examples, you needed only ask :)

These are screenshots from that week: https://x.com/barrelltech/status/1995900100174880806

I'm not going to share the prompt because (1) it's very long (2) there were dozens of variations and (3) it seems like poor business practices to share the most indefensible part of your business online XD

reply
sandblast
22 minutes ago
[-]
XD XD
reply
msp26
56 minutes ago
[-]
The new large model uses DeepseekV2 architecture. 0 mention on the page lol.

It's a good thing that open source models use the best arch available. K2 does the same but at least mentions "Kimi K2 was designed to further scale up Moonlight, which employs an architecture similar to DeepSeek-V3".

---

vllm/model_executor/models/mistral_large_3.py

```

from vllm.model_executor.models.deepseek_v2 import DeepseekV3ForCausalLM

class MistralLarge3ForCausalLM(DeepseekV3ForCausalLM):

```

"Science has always thrived on openness and shared discovery." btw

Okay I'll stop being snarky now and try the 14B model at home. Vision is good additional functionality on Large.

reply
nullbio
31 minutes ago
[-]
Anyone else find that despite Gemini performing best on benches, it's actually still far worse than ChatGPT and Claude? It seems to hallucinate nonsense far more frequently than any of the others. Feels like Google just bench maxes all day every day. As for Mistral, hopefully OSS can eat all of their lunch soon enough.
reply
mrtksn
29 seconds ago
[-]
Yep, Gemini is my least favorite and I’m convinced that the hype around it isn’t organic because I don’t see the claimed “superiority”, quite the opposite.
reply
mvkel
27 minutes ago
[-]
Open weight LLMs aren't supposed to "beat" closed models, and they never will. That isn’t their purpose. Their value is as a structural check on the power of proprietary systems; they guarantee a competitive floor. They’re essential to the ecosystem, but they’re not chasing SOTA.
reply
barrell
23 minutes ago
[-]
I can attest to Mistral beating OpenAI in my use cases pretty definitively :)
reply
re-thc
7 minutes ago
[-]
> Open weight LLMs aren't supposed to "beat" closed models, and they never will. That isn’t their purpose.

Do things ever work that way? What if Google did Open source Gemini. Would you say the same? You never know. There's never "supposed" and "purpose" like that.

reply
apexalpha
29 minutes ago
[-]
No, I've been using Gemini for help while learning / building my onprem k8s cluster and it has been almost spotless.

Granted, this is a subject that is very well present in the training data but still.

reply
Synthetic7346
12 minutes ago
[-]
I found gemini 3 to be pretty lackluster for setting up an onprem k8s cluster - sonnet 4.5 was more accurate from the get go, required less handholding
reply
alfalfasprout
28 minutes ago
[-]
If anything it's a testament to human intelligence that benchmarks haven't really been a good measure of a model's competence for some time now. They provide a relative sorting to some degree, within model families, but it feels like we've hit an AI winter.
reply
mythz
1 hour ago
[-]
Europe's bright star has been quiet for a while, great to see them back and good to see them come back to Open Source light with Apache 2.0 licenses - they're too far from the SOTA pack that exclusive/proprietary models would work in their favor.

Mistral had the best small models on consumer GPUs for a while, hopefully Ministral 14B lives up to their benchmarks.

reply
rvz
1 hour ago
[-]
All thanks to the US VCs that acutally have money to fund Mistral's entire business.

Had they gone to the EU, Mistral would have gotten a miniscule grant from the EU to train their AI models.

reply
amarcheschi
35 minutes ago
[-]
Mistral biggest investor is asml, although it became so later than other vcs
reply
crimsoneer
1 hour ago
[-]
I mean, one is a government, the other are VCs (also, I would be shocked if there isn't some French gov funding somewhere in the massive mistral pile).
reply
whiplash451
1 hour ago
[-]
1. so what 2. asml
reply
rvz
53 minutes ago
[-]
1. It matters.

2. Did ASML invest in Mistral in their first round of venture funding or was it US VCs all along that took that early risk and backed them from the very start?

Risk aversion is in the DNA and in almost every plot of land in Europe such that US VCs saw something in Mistral before even the european giants like ASML did.

ASML would have passed on Mistral from the start and Mistral would have instead begged to the EU for a grant.

reply
apexalpha
59 minutes ago
[-]
1. Big problem

2. ASML was propped up by ASM and Philips, stepping in as "VCs"

reply
didibus
53 minutes ago
[-]
For VC don't you need a lot of capital and people with too much money?

Isn't that then a chicken and egg?

reply
timpera
2 hours ago
[-]
Extremely cool! I just wish they would also include comparisons to SOTA models from OpenAI, Google, and Anthropic in the press release, so it's easier to know how it fares in the grand scheme of things.
reply
Youden
1 hour ago
[-]
They mentioned LMArena, you can get the results for that here: https://lmarena.ai/leaderboard/text

Mistral Large 3 is ranked 28, behind all the other major SOTA models. The delta between Mistral and the leader is only 1418 vs. 1491 though. I *think* that means the difference is relatively small.

reply
jampekka
47 minutes ago
[-]
1491 vs 1418 ELO means the stronger model wins about 60% of the time.
reply
supermatt
39 minutes ago
[-]
Probably naive questions:

Does that also mean that Gemini-3 (the top ranked model) loses to mistral 3 40% of the time?

Does that make Gemini 1.5x better, or mistral 2/3rd as good as Gemini, or can we not quantify the difference like that?

reply
esafak
35 minutes ago
[-]
Yes, of course.
reply
qznc
1 hour ago
[-]
I guess that could be considered comparative advertising then and companies generally try to avoid that scrutiny.
reply
rvz
1 hour ago
[-]
> I just wish they would also include comparisons to SOTA models from OpenAI, Google, and Anthropic in the press release,

Why would they? They know they can't compete against the heavily closed-source models.

They are not even comparing against GPT-OSS.

That is absolutely and shockingly bearish.

reply
constantcrying
2 hours ago
[-]
The lack of the comparison (which absolutely was done), tells you exactly what you need to know.
reply
bildung
41 minutes ago
[-]
I think people from the US often aren't aware how many companies from the EU simply won't risk losing their data to the providers you have in mind, OpenAI, Anthropic and Google. They simply are no option at all.

The company I work for for example, a mid-sized tech business, currently investigates their local hosting options for LLMs. So Mistral certainly will be an option, among the Qwen familiy and Deepseek.

Mistral is positioning themselves for that market, not the one you have in mind. Comparing their models with Claude etc. would mean associating themselves with the data leeches, which they probably try to avoid.

reply
popinman322
1 hour ago
[-]
They're comparing against open weights models that are roughly a month away from the frontier. Likely there's an implicit open-weights political stance here.

There are also plenty of reasons not to use proprietary US models for comparison: The major US models haven't been living up to their benchmarks; their releases rarely include training & architectural details; they're not terribly cost effective; they often fail to compare with non-US models; and the performance delta between model releases has plateaued.

A decent number of users in r/LocalLlama have reported that they've switched back from Opus 4.5 to Sonnet 4.5 because Opus' real world performance was worse. From my vantage point it seems like trust in OpenAI, Anthropic, and Google is waning and this lack of comparison is another symptom.

reply
kalkin
30 minutes ago
[-]
Scale AI wrote a paper a year ago comparing various models performance on benchmarks to performance on similar but held-out questions. Generally the closed source models performed better, and Mistral came out looking pretty badly: https://arxiv.org/pdf/2405.00332
reply
extr
48 minutes ago
[-]
??? Closed US frontier models are vastly more effective than anything OSS right now, the reason they didn’t compare is because they’re a different weight class (and therefore product) and it’s a bit unfair.

We’re actually at a unique point right now where the gap is larger than it has been in some time. Consensus since the latest batch of releases is that we haven’t found the wall yet. 5.1 Max, Opus 4.5, and G3 are absolutely astounding models and unless you have unique requirements some way down the price/perf curve I would not even look at this release (which is fine!)

reply
tarruda
1 hour ago
[-]
Here's what I understood from the blog post:

- Mistral Large 3 is comparable with the previous Deepseek release.

- Ministral 3 LLMs are comparable with older open LLMs of similar sizes.

reply
constantcrying
1 hour ago
[-]
And implicit in this is that it compares very poorly to SOTA models. Do you disagree with that? Do you think these Models are beating SOTA and they did not include the benchmarks, because they forgot?
reply
saubeidl
57 minutes ago
[-]
Those are SOTA for open models. It's a separate league from closed models entirely.
reply
supermatt
20 minutes ago
[-]
> It's a separate league from closed models entirely.

To be fair, we don’t even know if the closed models are even LLMs. They could be doing all manner of tool use behind the scenes.

reply
tarruda
1 hour ago
[-]
> Do you disagree with that?

I think that Qwen3 8B and 4B are SOTA for their size. The GPQA Diamond accuracy chart is weird: Both Qwen3 8B and 4B have higher scores, so they used this weid chart where "x" axis shows the number of output tokens. I missed the point of this.

reply
crimsoneer
1 hour ago
[-]
If someone is using these models, they probably can't or won't use the existing SOTA models, so not sure how useful those comparisons actually are. "Here is a benchmark that makes us look bad from a model you can't use on a task you won't be undertaking" isn't actually helpful (and definitely not in a press release).
reply
constantcrying
1 hour ago
[-]
Completely agree, that there are legitimate reasons to prefer comparison to e.g. deepeek models. But that doesn't change my point, we both agree that the comparisons would be extremely unfavorable.
reply
Lapel2742
1 hour ago
[-]
> that the comparisons would be extremely unfavorable.

Why should they compare apples to oranges? Ministral3 Large costs ~1/10th of Sonnet 4.5. They clearly target different users. If you want a coding assistant you probably wouldn't choose this model for various reasons. There is place for more than only the benchmark king.

reply
constantcrying
1 hour ago
[-]
Come on. Do you just not read posts at all?
reply
esafak
58 minutes ago
[-]
Which lightweight models do these compare unfavorably with?
reply
yvoschaap
2 hours ago
[-]
Upvoting for Europe's best efforts.
reply
sebzim4500
1 hour ago
[-]
That's unfair to Europe. A bunch of AI work is done in London (Deepmind is based here for a start)
reply
Glemkloksdjf
1 hour ago
[-]
Thats not the point.

Deepmind is not an UK company, its google aka US.

Mistral is a real EU based company.

reply
gishh
55 minutes ago
[-]
Using US VC dollars. Where their desks are isn’t really important.
reply
vintermann
36 minutes ago
[-]
Currency is interchangeable. Location might not be.
reply
p2detar
1 hour ago
[-]
That's ok. How could they know that there are companies like Alph Alpha, Helsing or the famous DeepL. European companies are just not that vocal, but that doesn't mean there are not making progress.
reply
colesantiago
1 hour ago
[-]
Deepmind doesn't exist anymore.

Google DeepMind does exist.

reply
GaggiX
1 hour ago
[-]
London is not part of Europe anymore since Brexit /s
reply
ot
1 hour ago
[-]
Is it so hard for people to understand that Europe is a continent, EU is a federation of European countries, and the two are not the same?
reply
denysvitali
12 minutes ago
[-]
I honestly think it is. The amount of people who thinks Europe and EU are the same thing is really concerning.

And no, it's not only americans. I keep hearing this thing from people living in Europe as well (or better, in the EU). I also very often hear phrases like "Switzerland is not in Europe" to indicate that the country is not part of the European Union.

reply
usrnm
1 hour ago
[-]
Europe isn't even a continent and has no real definition (none that would make any sense, anyway), so the whole thing is confusing by design
reply
lostmsu
21 minutes ago
[-]
Isn't London on an island, mr. Pedantic?
reply
GaggiX
1 hour ago
[-]
I think you missed the joke
reply
LunaSea
35 minutes ago
[-]
Upvoting Windows 11 as the US's best effort at Operating Systems development.
reply
DarmokJalad1701
8 minutes ago
[-]
Wouldn't that be macOS? Or BSD? Or Unix? CentOS?
reply
dmezzetti
1 minute ago
[-]
Looking forward to trying them out. Great to see they are Apache 2.0...always good to have easy-to-understand licensing.
reply
simgt
2 hours ago
[-]
I still don't understand what the incentive is for releasing genuinely good model weights. What makes sense however is OpenAI releasing a somewhat generic model like gpt-oss that games the benchmarks just for PR. Or some Chinese companies doing the same to cut the ground from under the feet of American big tech. Are we really hopeful we'll still get decent open weights models in the future?
reply
mirekrusin
1 hour ago
[-]
Because there is no money in making them closed.

Open weight means secondary sales channels like their fine tuning service for enterprises [0].

They can't compete with large proprietary providers but they can erode and potentially collapse them.

Open weights and research builds on itself advancing its participants creating environment that has a shot at proprietary services.

Transparency, control, privacy, cost etc. do matter to people and corporations.

[0] https://mistral.ai/solutions/custom-model-training

reply
NitpickLawyer
1 hour ago
[-]
> gpt-oss that games the benchmarks just for PR.

gpt-oss is killing the ongoing AIME3 competition on kaggle. They're using a hidden, new set of problems, IMO level, handcrafted to be "AI hardened". And gpt-oss submissions are at ~33/50 right now, two weeks into the competition. The benchmarks (at least for math) were not gamed at all. They are really good at math.

reply
lostmsu
18 minutes ago
[-]
Are they ahead of all other recent open models? Is there a leaderboard?
reply
NitpickLawyer
7 minutes ago
[-]
There is a leaderboard [1] but we'll have to wait till april for the competition to end to know what models they're using. The current number 3 on there (34/50) has mentioned in discussions that they're using gpt-oss-120b. There were also some scores shared for gpt-oss-20b, in the 25/50 range.

The next "public" model is qwen30b-thinking at 23/50.

Competition is limited to 1 H100 (80GB) and 5h runtime for 50 problems. So larger open models (deepseek, larger qwens) don't fit.

[1] https://www.kaggle.com/competitions/ai-mathematical-olympiad...

reply
talliman
1 hour ago
[-]
Until there is a sustainable, profitable and moat-building business model for generative AI, the competition is not to have the best proprietary model, but rather to raise the most VC money to be well positioned when that business model does arise.

Releasing a near stat-of-the-art open model instanly catapults companies to a valuation of several billion dollars, making it possible raise money to acquire GPUs and train more SOTA models.

Now, what happens if such a business model does not emerge? I hope we won't find out!

reply
mirekrusin
1 hour ago
[-]
Explained well in this documentary [0].

[0] https://www.youtube.com/watch?v=BzAdXyPYKQo

reply
simgt
1 hour ago
[-]
I was fully expecting that but it doesn't get old ;)
reply
memming
1 hour ago
[-]
It’s funny how future money drive the world. Fortunately it’s fueling progress this time around.
reply
nullbio
30 minutes ago
[-]
Google games benchmarks more than anyone, hence Gemini's strong bench lead. In reality though, it's still garbage for general usage.
reply
prodigycorp
1 hour ago
[-]
gpt-oss are really solid models. by far the best at tool calling, and performant.
reply
arnaudsm
1 hour ago
[-]
Geometric mean of MMMLU + GPQA-Diamond + SimpleQA + LiveCodeBench :

- Gemini 3.0 Pro : 84.8

- DeepSeek 3.2 : 83.6

- GPT-5.1 : 69.2

- Claude Opus 4.5 : 67.4

- Kimi-K2 (1.2T) : 42.0

- Mistral Large 3 (675B) : 41.9

- Deepseek-3.1 (670B) : 39.7

The 14B 8B & 3B models are SOTA though, and do not have chinese censorship like Qwen3.

reply
jasonjmcghee
1 hour ago
[-]
How is there such a gap between Gemini 3 vs GPT 5.1/Opus 4.5? What is Gemini 3 crushing the others on?
reply
arnaudsm
39 minutes ago
[-]
Could be optimized for benchmarks, but Gemini 3 has been stellar for my tasks so far.

Maybe an architectural leap?

reply
gishh
54 minutes ago
[-]
Gamed tests?
reply
rdtsc
46 minutes ago
[-]
I always joke that Google pays for a dedicated developer to spend their full time just to make pelicans on bicycles look good. They certainly have the cash to do it.
reply
hnuser123456
1 hour ago
[-]
Looks like their own HF link is broken or the collection hasn't been made public yet. The 14B instruct model is here:

https://huggingface.co/mistralai/Ministral-3-14B-Instruct-25...

The unsloth quants are here:

https://huggingface.co/unsloth/Ministral-3-14B-Instruct-2512...

reply
janpio
1 hour ago
[-]
reply
lalassu
1 hour ago
[-]
It's sad that they only compare to open weight models. I feel most users don't care much about OSS/not OSS. The value proposition is the quality of the generation for some use case.

I guess it says a bit about the state of European AI

reply
para_parolu
1 hour ago
[-]
It’s not for users but for businesses. There is demand for inhouse use with data privacy. Regular users can’t even run large model due to lack of compute.
reply
hopelite
1 hour ago
[-]
It seems to be a reasonable comparison since that is the primary/differentiating characteristic of the model. It’s really common to also and seemingly only ever see the comparison of closed weight/proprietary models in a way that seems to act as if all of the non-American and open weight models don’t even exist.

I also think most people do not consider open weights as OSS.

reply
esafak
1 hour ago
[-]
Well done to the France's Mistral team for closing the gap. If the benchmarks are to be believed, this is a viable model, especially at the edge.
reply
nullbio
28 minutes ago
[-]
Benchmarks are never to be believed, and that has been the case since day 1.
reply
andhuman
1 hour ago
[-]
This is big. The first really big open weights model that understands images.
reply
yoavm
1 hour ago
[-]
How is this different from Llama 3.2 "vision capabilities"?

https://www.llama.com/docs/how-to-guides/vision-capabilities...

reply
Havoc
1 hour ago
[-]
Guessing GP commenter considers Apache more "open" than Meta's license. Which to be fair isn't terrible but also not quite as clean as straight apache
reply
tootyskooty
37 minutes ago
[-]
Since no one has mentioned it yet: note that the benchmarks for large are for the base model, not for the instruct model available in the API.

Most likely reason is that the instruct model underperforms compared to the open competition (even among non-reasoners like Kimi K2).

reply
trvz
29 minutes ago
[-]
Sad to see they've apparently fully given up on releasing their models via torrent magnet URLs shared on Twitter; those will stay around long after Hugging Face is dead.
reply
jasonjmcghee
1 hour ago
[-]
I wish they showed how they compared to models larger/better and what the gap is, rather than only models they're better than.

Like how does 14B compare to Qwen30B-A3B?

(Which I think is a lot of people's goto or it's instruct/coding variant, from what I've seen in local model circles)

reply
another_twist
1 hour ago
[-]
I am not sure why Meta paid 13B+ to hire some kid vs just hiring back or acquiring these folks. They'll easily catch up.
reply
Rastonbury
57 minutes ago
[-]
Age aside, not sure what Zuck was thinking, seeing as Scale AI was in data labelling and not training models, perhaps he thought he was a good operator? Then again the talent scarcity is in scientists, there are many operators, let alone one worth 14B. Back to age, the people he is managing are likely all several years older than him and Meta long timers, which would make it even more challenging
reply
Tiberium
1 hour ago
[-]
A bit interesting that they used Deepseek 3's architecture for their Large model :)
reply
codybontecou
2 hours ago
[-]
Do all of these models, regardless of parameters, support tool use and structured output?
reply
Y_Y
1 hour ago
[-]
In principle any model can do these. Tool use is just detecting something like "I should run a db query for pattern X" and structured output is even easier, just reject output tokens that don't match the grammar. The only question is how well they're trained, and how well your inference environment takes advantage.
reply
tucnak
1 hour ago
[-]
If the claims on multilingual and pretraining performance are accurate, this is huge! This may be the best-in-class multilingual stuff since the more recent Gemma's, where they used to be unmatched. I know Americans don't care much about the rest of the world, but we're still using our native tongues thank you very much; there is a huge issue with i.e. Ukrainian (as opposed to Russian) being underrepresented in many open-weight and weight-available models. Gemma used to be a notable exception, I wonder if it's still the case. On a different note: I wonder why scores on TriviaQA vis-a-vis 14b model lags behind Gemma 12b so much; that one is not a formatting-heavy benchmark.
reply
NitpickLawyer
1 hour ago
[-]
> I wonder why scores on TriviaQA vis-a-vis 14b model lags behind Gemma 12b so much; that one is not a formatting-heavy benchmark.

My guess is the vast scale of google data. They've been hoovering data for decades now, and have had curation pipelines (guided by real human interactions) since forever.

reply
GaggiX
1 hour ago
[-]
The small dense model seems particularly good for their small sizes, I can't wait to test them out.
reply
s_dev
1 hour ago
[-]
I was subscribing to these guys purely to support the EU tech scene. So I was on Pro for about 2 years while using ChatGPT and Claude.

Went to actually use it, got a message saying that I missed a payment 8 months previously and thus wasn't allowed to use Pro despite having paid for Pro for the previous 8 months. The lady I contacted in support simply told me to pay the outstanding balance. You would think if you missed a payment it would relate to simply that month that was missed not all subsequent months.

Utterly ridiculous that one missed payment can justify not providing the service (otherwise paid for in full) at all.

Basically if you find yourself in this situation you're actually better of deleting the account and resigning up again under a different email.

We really need to get our shit together in the EU on this sort of stuff, I was a paying customer purely out of sympathy but that sympathy dried up pretty quick with hostile customer service.

reply
shlomo_z
16 minutes ago
[-]
This seems like a legitimate complaint... I wonder why it's downvoted
reply