AI SOFTWARE CODING

Welcome to the Transformers’ Era

Transformers may reshape deeply the way applications will be developed in the coming years. In the near future, most of the low-value code produced by a developer might be generated by AI and Transformers.

A Transformer is a deep learning model relying on self-attention mechanism to compute representations without using sequence-aligned Recurrent Neural Networks (RNNs).

Recently, OpenAI released the largest model of its Generative Pre-trained Transformer called GPT-3, having a huge capacity of around 175 billion machine learning parameters. It has a remarkable capability to leverage deep learning and generate human understandable texts that could be stories, poems but also source code.

Read More

Share This