top of page

#

4

Weekly Digest: 09th Aug 2024

Headlines: OpenAI founder dropouts, Compute mania in Generative AI, Google competitor Perplexity scaling, How big tech is circumventing regulation in M&A, New architecture in Neural networks, Long read - Generative AI industry dynamic, Deep Learning Foundations book, Explainer - Attention is all you need


AI in Businesses

 


Technology updates from AI

 

  • A new architecture in deep learning neural networks. The new networks are called Kolmogorov-Arnold Networks (KANs), after two mathematicians who studied how functions could be combined.

     

    "In the new architecture, the synapses play a more complex role. Instead of simply learning how strong the connection between two neurons is, they learn the full nature of that connection—the function that maps input to output. Unlike the activation function used by neurons in the traditional architecture, this function could be more complex—, a “spline” or combination of several functions—and is different in each instance."

     

  •  Why would we care - the resulting network is several orders of magnitude smaller compared to similarly performing MLP models (depending on the task performed - mathematical problems, image recognition). In one case it was as much as 100 times.

  • Not just that, the resulting networks are a lot more explainable as now there is a way to inspect the functions rather than absolute 'weights' itself.

  • However, The architecture won't go mainstream yet given that it is not inherently parallelisable as the MLP networks and takes lot longer to train. A disadvantage while training LLMs. 

     

     (Matthew Hutson, IEEE Spectrum)


Long read - Generative AI Industry Dynamics

 

  • It's just not clear if there will be a long-term, winner-take-all dynamic in generative AI. Who owns the generative AI platform is an insightful investigation into competitive advantage and long-term revenue stability at various layers of the generative AI business. The below paragraph summarises it well.

  • "There don’t appear to be any systemic moats in generative AI today. As a first-order approximation, applications lack strong product differentiation because they use similar models; models face unclear long-term differentiation because they are trained on similar datasets with similar architectures; cloud providers lack deep technical differentiation because they run the same GPUs; and even the hardware companies manufacture their chips at the same fabs."

  • "There are, of course, the standard moats: scale moats (“I have or can raise more money than you!”), supply-chain moats (“I have the GPUs, you don’t!”), ecosystem moats (“Everyone uses my software already!”), algorithmic moats (“We’re more clever than you!”), distribution moats (“I already have a sales team and more customers than you!”) and data pipeline moats (“I’ve crawled more of the internet than you!”). But none of these moats tend to be durable over the long term. And it’s too early to tell if strong, direct network effects are taking hold in any layer of the stack."

     

    (Matt Bornstein, Guido Appenzeller, Martin Casado - Andreessen Horowitz)



Resources

 

Weekly Digest: 09th Aug 2024

Follow

  • X
  • LinkedIn

©2024 Collationist.

bottom of page