• 0 Posts
  • 387 Comments
Joined 1 year ago
cake
Cake day: June 30th, 2023

help-circle
  • That’s fair. I see what I see at an engineering and architecture level. You see what you see at the business level.

    That said. I stand by my statement because I and most of my colleagues in similar roles get continued, repeated and expanded-scope engagements. Definitely in LLMs and genAI in general especially over the last 3-5 years or so, but definitely not just in LLMs.

    “AI” is an incredibly wide and deep field; much more so than the common perception of what it is and does.

    Perhaps I’m just not as jaded in my tech career.

    operations research, and conventional software which never makes mistakes if it’s programmed correctly.

    Now this is where I push back. I spent the first decade of my tech career doing ops research/industrial engineering (in parallel with process engineering). You’d shit a brick if you knew how much “fudge-factoring” and “completely disconnected from reality—aka we have no fucking clue” assumptions go into the “conventional” models that inform supply-chain analytics, business process engineering, etc. To state that they “never make mistakes” is laughable.


  • Absolutely not true. Disclaimer, I do work for NVIDIA as a forward deployed AI Engineer/Solutions Architect—meaning I don’t build AI software internally for NVIDIA but I embed with their customers’ engineering teams to help them build their AI software and deploy and run their models on NVIDIA hardware and software.

    To state this as simply as possible: I wouldn’t have a job if our customers weren’t seeing tremendous benefit from AI technology. The companies I work with typically are very sensitive to CapX and OpX costs of AI—they self-serve in private clouds. If it doesn’t help them make money (revenue growth) or save money (efficiency), then it’s gone—and so am I. I’ve seen it happen; entire engineering teams laid off because a technology just couldn’t be implemented in a cost-effective way.

    LLMs are a small subset of AI and Accelerated-Compute workflows in general.






  • We’re looking at this from opposite sides of the same coin.

    The NN graph is written at a high-level in Python using frameworks (PyTorch, Tensorflow—man I really don’t miss TF after jumping to Torch :) ).

    But the calculations don’t execute on the Python kernel—sure you could write it to do so but it would be sloooow. The actual network of calculations happen within the framework internals; C++. Then depending on the hardware you want to run it on, you go down to BLAS or CUDA, etc. all of which are written in low-level languages like Fortran or C.

    Numpy fits into places all throughout this stack and its performant pieces are mostly implemented in C.

    Any way you slice it: the post I was responding to is to argue that AI IS CODE. No two ways about that. It’s also the weights and biases and activations of the models that have been trained.










  • Oh they fucking know. Say it with me:

    Wealth Inequity
    

    I don’t think anyone really hates Jo Millionaire. Jo, the master electrician that lives down the street and employs 5-10 electricians from apprentice to employee-master is a millionaire and contributes positively to their local community. Creating jobs through helping people with their electrical projects, spending in the local economy, etc. And that’s a realistic goal for their apprentices to aspire and work towards.

    Unfortunately that’s who republican voters think they’re voting to support.

    But they’ve been duped; they’re actually voting to support the Billionaire Aristocrats of the world who pull up the ladder behind them through monetary influence of politics and not paying a damn dime on their ‘income’ (because they’re “borrowing against” their unfathomable hoard).

    “They” know why the voters and disenfranchised and that’s their fucking plan—because it keeps them employed and wining and dining fancy with their Aristocrat puppet masters.





  • No. Strictly and technically speaking, LLMs absolutely fall under the category of AI. You’re thinking of AGI, which is a subset of AI, and which LLMs will be a necessary but insufficient component of.

    I’m an AI Engineer; I’ve taken to, in my circles, calling AI “Algorithmic Intelligence” rather than “Artificial Intelligence.” It’s far more fitting term for what is happening. But until the Yanns and Ngs and Hintons of the field start calling it that, we’re stuck with it.