• flamingo_pinyata@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    4 months ago

    Yes actually, I can imagine it getting microcontroller code wrong. My niche is general backend services. I’ve been using Github copilot a lot and it served me well for generating unit tests. Write test description and it pops out the code with ~ 80% accuracy

    • hendrik@palaver.p3x.de
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      4 months ago

      Sure. There are lots of tedious tasks in a programmers life that don’t require a great amount of intelligence. I suppose writing some comments, docstrings, unit tests, “glue” and boilerplate code that connects things and probably several other things that now escape my mind are good tasks for an AI to assist a proper programmer and make them more effective and get things done faster.

      I just wouldn’t call that programming software. I think assisting with some narrow tasks is more exact.

      Maybe I should try doing some backend stuff. Or give it an API definition and see what it does 😅 Maybe I was a bit blinded by ChatGPT having read the Wikipedia and claiming it understands robotics concepts. But it really doesn’t seem to have any proper knowledge. Same probably applies to engineering and other nighboring fields that might need software.

      • flamingo_pinyata@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 months ago

        It might also have to do with specialized vs general models. Copilot is good at generating code but ask it to write prose text and it fails completely. In contrast ChatGPT is awful at code but handles human readable text decently.