Developer and refugee from Reddit

  • 13 Posts
  • 319 Comments
Joined 2 years ago
cake
Cake day: July 2nd, 2023

help-circle




  • Not really. I’m using Copilot for my job because I’m required to, but I limit it to creating basic scripts that accomplish very specific tasks, which it’s okay at. I never let it touch my actual codebase.

    Every experiment I’ve run to test how it does with real code from a large, production application has failed miserably. It introduces mocks, it duplicates stuff everywhere, it fakes shit when real APIs are available… I have no fucking idea how these “AI” companies have managed to convince so many businesses that producing catastrophically shitty code very quickly is a good thing.










  • Venture capital drying up.

    Here’s the thing… No LLM provider’s business is making a profit. None of them. Not OpenAI. Not Anthropic. Not even Google (they’re profitable in other areas, obviously). OpenAI optimistically believes it might start being profitable in 2029.

    What’s keeping them afloat? Venture capital. And what happens when those investors decide to stop throwing good money after bad?

    BOOM.




  • There are tricks to getting better output from it, especially if you’re using Copilot in VS Code and your employer is paying for access to models, but it’s still asking for trouble if you’re not extremely careful, extremely detailed, and extremely precise with your prompts.

    And even then it absolutely will fuck up. If it actually succeeds at building something that technically works, you’ll spend considerable time afterwards going through its output and removing unnecessary crap it added, fixing duplications, securing insecure garbage, removing mocks (God… So many fucking mocks), and so on.

    I think about what my employer is spending on it a lot. It can’t possibly be worth it.