Apoio Machine Learning-2
How businesses are actually using AI
Today AI is most prevalent in big companies (with more than 250 employees), which can afford to enlist dedicated AI teams and to pay for necessary investments.A poll of large firms by Morgan Stanley, a bank, found that between the start and end of 2023 the share with pilot AI projects rose from 9% to 23%.
This “use-case sprawl”, as one consultant calls it, can be divided into three big categories: window-dressing, tools for workers with low to middling skills, and those for a firm’s most valuable employees.
Of these, window-dressing is by far the most common. Many firms are rebranding run-of-the-mill digitisation efforts as “gen AI programmes” to sound more sophisticated, says Kristina McElheran of the University of Toronto
Tools for lower-skilled workers could be more immediately useful. Some simple applications for things like customer service involve off-the-shelf AI. Routine administrative tasks likewise look ripe for AI disruption. At Nasdaq, a financial-services firm, it helps financial-crime sleuths gather evidence to assess suspicious bank transactions. According to the company, this cuts a process which can take 30-60 minutes to three minutes.
Giving AI tools to a firm’s most valuable workers, whose needs are complex, is less widespread so far. But it, too, is increasingly visible. Lawyers have been among the earliest adopters. Allen & Overy, a big law firm, teamed up with Harvey, an AI startup, to develop a system that its lawyers use to help with everything from due diligence to contract analysis.
AI´s Long Tail (customization) Problem
É necessário lembrar que o "AI´s Long Tail Problem" é um produto off-the-shelf, não somente para low-skilled workers, mas também, e principalmente, para pequenas empresas.
E o mais interessante é que depois que você faz um off-the-shelf para uma pizzaria fica muito fáil adaptar para outra pizzaria, apesar de serem dois off-the-shelf diferentes.


Inteligência Artificial na Medicina: Casos Impressionantes
10 de dezembro de 2024
Itemização da Apresentação:
1 - Tipos de Aplicações em Geral
2 - Casos Práticos em Cirurgias, UTIs, Identificação de câncer, etc.
3 - Aplicações em hospitais brasileiros
4 - Aprendizado de Máquina (Machine Learning) e Base de Dados
5 - Os problemas éticos e casos reais na Justiça
Essa Apresentação é uma colaboração do colega Daniel Cavalcanti (dcavalcanti08@gmail.com)
Relatório Atualizado das Ferramentas GenAI
15 de dezembro de 2024
O IMD Business School, em Lausanne na Suíça, através do Centro de Pesquisas em Transformação Digital, publicou esta semana, o relatório atualizado das ferramentas GenAI categorizados e com a indicação (linha pontilhada) de qual é a ferramenta ou serviço de IA que possue melhor qualidade naquela categoria (dentro da metodologia de avaliação deles, claro). Serve como um bom guia, dentro da diversidade e variedade que temos. Um ótimo indicativo como guia.

Esta é uma colaboração do colega Fernando José (***@gmail.com)
Killer Robots
- NETFLIX -
2023


Machine Learning e Pizza???
- YouTube Originals - The Age of A.I.
em 15/jan/2020

Assitir a partir de 27min e 31seg. No total dá uns 9 min aproximadamente.
Section 1.6 - Reinforcement Learning and Supervised Learning
At the core of deep reinforcement learning is function approximation. This is something it shares with supervised learning (SL). However, reinforcement learning is unlike supervised learning in a number of ways. There are three main differences:
a) Lack of an oracle;
b) Sparsity of feedback; and
c) Data generated during training
Subsection 1.6.1 - Lack of an Oracle
A major difference between reinforcement learning and supervised learning is that for reinforcement learning problems, the “correct” answer for each model input is not available, whereas in supervised learning we have access to the correct or optimal answer for each example. In reinforcement learning, the equivalent of the correct answer would be access to an “oracle” which tells us the optimal action to take at every time step so as to maximize the objective.
In reinforcement learning, after an agent takes action a in state s, it only has access to the reward it receives. The agent is not told what the best action to take was. Instead, it is only given an indication, via the reward, of how good or bad a was. Not only does this convey less information than the right answer would have provided, but the agent only learns about rewards for the states it experiences. To learn about (s; a; r), an agent must experience the transition (s; a; r; s’). An agent may have no knowledge about important parts of the state and action spaces because it hasn’t experienced them.
Laura Graesser and Wah Loon Keng,
“Foundations of Deep Reinforcement Learning: Theory and Practice in Python”,
Publisher: Addison-Wesley Professional; 1st edition, 2020.
Chapter 19 - Reinforcement Learning for Decision Making in Complex Environments
The theoretical foundations of RL

Sebastian Raschka, Yuxi (Hayden) Liu, Vahid Mirjalili,
“Machine Learning with PyTorch and Scikit-Learn:
Develop machine learning and deep learning models with Python”,
Packt Publishing, February 25, 2022.
Reinforcement Learning (RL) Problems Are Strategic
Reinforcement learning (RL) is an old subject; it’s decades old. Only recently has it gained enough prominence to raise its head outside of academia. I think that this is partly because there isn’t enough disseminated industrial knowledge yet. The vast majority of the literature talks about algorithms and contrived simulations, until now.
Researchers and industrialists are beginning to realize the potential of RL. This brings a wealth of experience that wasn’t available in 2015. Frameworks and libraries are following suit, which is increasing awareness and lowering the barrier to entry.
I like to think of software engineering as a way of automating processes and machine learning can automate decisions. RL can automate strategies.
Figure 9-3 hammers this idea home. Modern businesses are made of three core functions. Businesses have a plethora of processes, from performance reviews to how to use the printer. Enacting of any one of these processes is rarely valuable, but they tend to occur often. Because of the frequency and the fact it takes time and therefore money for a human to perform this process, it makes sense to automate the process in software.

Figure 9-3. A depiction of the different functions of a modern business, the value of each action in that function, and the technology that maps to the function.
Business decisions are rarer than business processes, but they still happen fairly often. Making the decision to perform CPR (Cardiopulmonary Resuscitation), phone a warm prospect, or block a suspicious looking transaction can save lives/be profitable/prevent fraud. All of these challenges can be solved using machine learning. These actions are quantifiably valuable, but like I said previously, they have a limited life span. They intentionally focus on the single transitive event. Decisions need processes in place to make them possible.
Business strategies form the heart of the company. They happen rarely, primarily because there are limited numbers of people (or one) designing and implementing the strategy. They are closely held secrets and direct a multitude of future people, decisions, and processes. They have wide-reaching consequences, like the healthcare for a nation or the annual profits of the company. This is why optimal strategies are so valuable and why RL is so important.
Imagine being able to entrust the running of your company/healthcare/country to RL. How does that sound? What do you think?
You might suggest that this is a pipe dream, but I assure you it is not. Taxation is being tackled with MARL. Healthcare use cases are obvious and there are thousands of references. And companies must automate to become more competitive, which is nothing new - companies have been automating for over a century now.
The future, therefore, is in the hands of the engineers, the people who know how to build, adapt, and constrain algorithms to be robust, useful, and above all, safe.
Phil Winder Ph.D., “Reinforcement Learning: Industrial Applications of Intelligent Agents”,
Publisher: O'Reilly Media; 1st edition (December 15, 2020)
An overview of the major Deep Reinforcement Learning algorithms in each family and how they are related.

Laura Graesser and Wah Loon Keng,
“Foundations of Deep Reinforcement Learning: Theory and Practice in Python”,
Publisher: Addison-Wesley Professional; 1st edition, 2020.
Generative AI in a Nutshell
- how to survive and thrive in the age of AI

AI Is Dangerous, but Not for the Reasons You Think | Sasha Luccioni | TED
https://www.youtube.com/watch?v=eXdVDhOGqoE
6 de nov. de 2023

A meta de decifrar todo o conectoma humano é alcançável???
Números em Revisão
- Recontagem de neurônios põe em xeque ideias da neurociência


Build a Large Language Model (From Scratch)
https://www.manning.com/books/build-a-large-language-model-from-scratch
Acesado em 24 de abril de 2024

As illustrated in Figure 1.3, the first step in creating an LLM is to train it in on a large corpus of text data, sometimes referred to as raw text. Here, "raw" refers to the fact that this data is just regular text without any labeling information[1]. (Filtering may be applied, such as removing formatting characters or documents in unknown languages.)
This first training stage of an LLM is also known as pretraining, creating an initial pretrained LLM, often called a base or foundation model. A typical example of such a model is the GPT-3 model (the precursor of the original model offered in ChatGPT). This model is capable of text completion, that is, finishing a half-written sentence provided by a user. It also has limited few-shot capabilities, which means it can learn to perform new tasks based on only a few examples instead of needing extensive training data. This is further illustrated in the next section, Using transformers for different tasks.
After obtaining a pretrained LLM from training on large text datasets, where the LLM is trained to predict the next word in the text, we can further train the LLM on labeled data, also known as finetuning.
The two most popular categories of finetuning LLMs include instruction-finetuning and finetuning for classification tasks. In instruction-finetuning, the labeled dataset consists of instruction and answer pairs, such as a query to translate a text accompanied by the correctly translated text. In classification finetuning, the labeled dataset consists of texts and associated class labels, for example, emails associated with spam and non-spam labels.
In this book, we will cover both code implementations for pretraining and finetuning LLM, and we will delve deeper into the specifics of instruction-finetuning and finetuning for classification later in this book after pretraining a base LLM.