Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

$6 billion dollars raised in 2 months for their series C is blowing my mind. What does Anthropic have that OpenAI or other LLM startups don't? What other companies have raised that much in a single round?


They managed to build something equally good (if not better) than GPT4, without losing 50% of their cap table to Microsoft. That alone is worth a lot.

The investments are absurdly large but if you believe LLMs will be fundamentally transformative then I can understand the logic.


Claude 2 is good and better than 3.5 but nothing close to 4 in general.


I’m not saying your experience is wrong, but for my use case (code, mainly), I’ve found Claude to be far more reliable (not to mention faster) than GPT4.


The larger context window might be nice, but it's nowhere close to GPT4's reasoning ability. GPT4 is still the best for coding.


I’m now almost always using Claude over GPT4 in my day to day and rarely have moments where I’m not happy with the output and fallback to ChatGPT

The big context window is so nice, I can dump in huge files or conversation exports and it can handle them without problems


How do you guys access Claude?


Claude is available without API on Poe and with API on Amazon Bedrock.


http://Claude.AI

Am I the same only one who uses the web interface?

I keep thing simple (KISS) because I teach for a living, and only things my student can pick up in a class are useful to me.


I use the web interface as well. Is there even any other way? I also pay for Claude Pro


Google announced you'll be able to use Claude via GCP Vertex AI soon.


claude.ai


In my experience Claude 2 is better than GPT 4 in many non programming tasks.


It's such a small playing field with too few players.

I have a slide in a presentation on the topic that's an inverted pyramid, as pretty much the entire LLM field rests on only a few companies, who aren't necessarily even doing everything correctly.

The fact that they even have a seat at such a small table with so much in the pot already and much more forecasted means they command a large buy in regardless of their tech. They don't ever need to be in the lead, they just need to maintain their seat at the table and they'll still have money being thrown their way.

The threat of missing the next wave or letting a competitor gain exclusive access is too high at this point.

Of course, FOMO driving investments is also the very well known pattern of a bubble, and we may have a bit of a generative AI bubble among the top firms where large sums of money are going to go down the drain on investing into overvalued promises because the cost of missing a promise that will actually come to fruition is considered too high.

Ironically the real payoff is probably in focusing on integration layers at this point, particularly given the gains in performance over the past year in research by developing improved interfacing with SotA models.

LLMs at a foundational pretrained layer are in a race towards parity. Having access to model A isn't going to be much more interesting than having access to model B. But if you have a plug and play intermediate product that can hook into either model A or B and deliver improved results to direct access to either - that's where the money is going to be for cloud providers in the next 18 months.


> What does Anthropic have that OpenAI or other LLM startups don't?

The big context window is pretty magical for some use cases. There are lots of things RAG with limited context can't do.


The "big context window" talk reminds me of high wattage speakers in the 80s. Yes, it's loud. "Does it sound good?", is the question.

Having a large context window is pointless unless the model is able to attend to attention on the document submitted. As RAG is basically search, this helps set attention regardless of the model's context window size.

Stuffing random thing into a prompt doesn't improve things, so RAG is always required, even if it's just loading the history of the conversation into the window.


RAG is not always required. If you can fit a whole document/story/etc in the context length, it often performs better, especially for complex questions (not just basic retrieval).


Exactly. RAG is great if you need certain paragraphs to answer specific questions but there are limits.

Assuming a new unheard of play called "Romeo and Juliet " RAG can answer "Who is Rosaline?".

But a large context window can answer "How does Shakespeare use foreshadowing to build tension and anticipation throughout the play?" "How are gender roles and societal expectations depicted in the play" or "What does the play suggest about the power of love to overcome adversity?".

In other words, RAG doesn't help you answer questions that pertain to the whole document and aren't keyword driven retrieval queries. It's a pretty big limitation if you aren't just looking for a specific fact.

RAG limits you to answers where the information can be contained in your chunk size.


Gotta be the talent. Started by ex-OpenAI and probably attracted some AI geniuses who buy in to their safety approach too.


joking, but maybe they have achieved agi (internally) ;)


OpenAI is said to have done so by a guy who leaked some very specific information that turned out to be accurate.


> What does Anthropic have that OpenAI or other LLM startups don't

I dunno I kind of get the perception Microsoft's locked up OpenAI's funding so Google couldn't throw money at them even if they wanted to?


Yeah I believe so as well. They are probably going after the second best thing that is out there that has a chance of overtaking GPT 4


I also rather like Claude's capability of using XML tags as scrap-pads of sorts to see the inner workings of the AI. Makes for better prompt design.

https://docs.anthropic.com/claude/docs/give-claude-room-to-t...

You can see how the AI arrived at the response.


This is just a prompting hack you can use with any LLM, not exclusive to Claude. But I do like the fact that they include these tricks in their documentation.


The large(100k tokens) context window together with the fact that it can actually use the information in that context window. From personal experience other models including open ai fail to properly answer when provided large(more than 5k tokens) inputs as context even when the model officially accepts much larger contexts. But Claude 2 models are uncannily good at taking all that context into consideration.


It's absolutely good, have you tried their Claude 2 AI?


It’s not as good as GPT-4, the only feature I used often was their PDF QA. The rest ChatGPT plus ( GPT-4 ) still miles ahead


I regularly compare queries with both GPT-4 and Claude 2 and I find I prefer Claude's answers most often. The 100k token context is a major plus too.

Claude is more creative and that comes with a higher rate of hallucinations. I hope I'm not misremembering but GPT-4 was also initially more prone to hallucinations but also more capable in some ways.


Massive context length


Competition.


[flagged]


How is that relevant?


If you ran a successful lemonade stand, and suddenly I discovered that Al Capone gave you the money to buy lemons, I would find it relevant, although not sure how I would act on the information.

Also, possibly there is some money laundering aspect to these investments that we don't fully see through.


Because it was a wasteful investment back then, Anthropic has more equity to spare now in a Series C.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: