Skip to content

Commit ea5d6ea

Browse files
QiJuneabhinavarora
authored andcommitted
update support new device docs (#6963)
* update docs * follow comments * fix typo
1 parent 97ec6f8 commit ea5d6ea

File tree

1 file changed

+9
-7
lines changed

1 file changed

+9
-7
lines changed

doc/design/support_new_device.md

Lines changed: 9 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -25,13 +25,14 @@ There are mainly three parts that we have to consider while integrating a new de
2525

2626
### Place and DeviceContext
2727

28+
Please remind that device and computing library are not one-to-one corresponding. A device can have a lot of computing libraries and a computing library can also support several devices.
2829

2930
#### Place
30-
Fluid uses class [Place](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/platform/place.h#L55) to represent different devices and computing libraries. There are inheritance relationships between different kinds of `Place`.
31+
Fluid uses class [Place](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/platform/place.h#L55) to represent the device memory where data is located. If we add another device, we have to add corresponding `DevicePlace`.
3132

3233
```
33-
| CPUPlace --> MKLDNNPlace
34-
Place --| CUDAPlace --> CUDNNPlace
34+
| CPUPlace
35+
Place --| CUDAPlace
3536
| FPGAPlace
3637
```
3738

@@ -43,7 +44,7 @@ typedef boost::variant<CUDAPlace, CPUPlace, FPGAPlace> Place;
4344

4445
#### DeviceContext
4546

46-
Fluid uses class [DeviceContext](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/platform/device_context.h#L30) to manage the resources in different hardwares, such as CUDA stream in `CDUADeviceContext`. There are also inheritance relationships between different kinds of `DeviceContext`.
47+
Fluid uses class [DeviceContext](https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/platform/device_context.h#L30) to manage the resources in different libraries, such as CUDA stream in `CDUADeviceContext`. There are also inheritance relationships between different kinds of `DeviceContext`.
4748

4849

4950
```
@@ -106,7 +107,7 @@ template <typename Place>
106107
size_t Used(Place place);
107108
```
108109

109-
To implementing these interfaces, we have to implement MemoryAllocator for different Devices
110+
To implement these interfaces, we have to implement MemoryAllocator for different Devices.
110111

111112

112113
#### Tensor
@@ -243,6 +244,7 @@ REGISTER_OP_CUDA_KERNEL(
243244
Generally, we will impelement OpKernel for all Device/Library of an Operator. We can easily train a Convolutional Neural Network in GPU. However, some OpKernel is not sutibale on a specific Device. For example, crf operator can only run on CPU, whereas most other operators can run at GPU. To achieve high performance in such circumstance, we have to switch between different Device/Library.
244245

245246

246-
We will discuss how to implement an efficient OpKernel switch policy.
247+
For more details, please refer to following docs:
247248

248-
- TBD
249+
- operator kernel type [doc](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/design/operator_kernel_type.md)
250+
- switch kernel [doc](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/design/switch_kernel.md)

0 commit comments

Comments
 (0)