Merge pull request #267 from bhamail/bhamail-patch-typo-a-canary

Add missing "a"
This commit is contained in:
Pikachú 2024-05-29 18:16:45 +03:00 коммит произвёл GitHub
Родитель fc75b316bb 41639bf8ae
Коммит 028369d569
Не найден ключ, соответствующий данной подписи
Идентификатор ключа GPG: B5690EEEBB952194
2 изменённых файлов: 2 добавлений и 2 удалений

Просмотреть файл

@ -58,7 +58,7 @@ Untyped-Language | doesn't have | type definitions
> ✅ Think how triplets can be used to represent other types of knowledge.
2. **Hierarchical representations** emphasize the fact that we often create a hierarchy of objects inside our head. For example, we know that canary is a bird, and all birds have wings. We also have some idea about what colour canary usually is, and what is their flight speed.
2. **Hierarchical representations** emphasize the fact that we often create a hierarchy of objects inside our head. For example, we know that canary is a bird, and all birds have wings. We also have some idea about what colour a canary usually is, and what is their flight speed.
- **Frame representation** is based on representing each object or class of objects as a **frame** which contains **slots**. Slots have possible default values, value restrictions, or stored procedures that can be called to obtain the value of a slot. All frames form a hierarchy similar to an object hierarchy in object-oriented programming languages.
- **Scenarios** are special kind of frames that represent complex situations that can unfold in time.

Просмотреть файл

@ -29,7 +29,7 @@ In the definition of *f* above, *w* and *b* are called **parameters** θ=
## Gradient Descent Optimization
There is a well-known method of function optimization called **gradient descent**. The idea is that we can compute a derivative (in multi-dimensional case call **gradient**) of loss function with respect to parameters, and vary parameters in such a way that the error would decrease. This can be formalized as follows:
There is a well-known method of function optimization called **gradient descent**. The idea is that we can compute a derivative (in multi-dimensional case called **gradient**) of loss function with respect to parameters, and vary parameters in such a way that the error would decrease. This can be formalized as follows:
* Initialize parameters by some random values w<sup>(0)</sup>, b<sup>(0)</sup>
* Repeat the following step many times: