Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

WSJ is wrong. Autocomplete code models are small (5B param) and very cheap. A single inference is roughly $0.000001.


Copilot uses GPT4 now


I mentioned autocomplete, not the chat service.


GPT4 isn’t a chat service, friend.


I think the point that ipsum2 is trying to make is that Copilot's chat service and its code completion service could be using different models, which is not uncommon for coding assistants.

Continue[0] for example can use up to 3 different models in a session: a large model such as GPT-4 Turbo for chat and code QA, a smaller low latency model such as StarCoder2-3B for code completion, and yet another model such as all-MiniLM-L6-v2 for generating embeddings.

[0] https://www.continue.dev/




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: