Inside TensorFlow: tf.data + tf.distribute

  Рет қаралды 22,117

TensorFlow

TensorFlow

Күн бұрын

Пікірлер: 8
@bapireddy5790
@bapireddy5790 4 жыл бұрын
Great talk on tf.data kudos to the engineers for making these high level api's
@zytriesthings4540
@zytriesthings4540 2 ай бұрын
My mind just exploded
@haneulkim4902
@haneulkim4902 Жыл бұрын
This is an amazing talk! Thank you. I have two questions. 1. @9:22 I understand that it reads all elements tfrecords files sequentially. but since map function is applied before batch, how many elements are loaded? -> one by one? 2. @27:25 so after gradients from different GPU have been for example averaged it then gets passed back to each GPU to update the model? and this process is repeated?
@prathameshdinkar2966
@prathameshdinkar2966 Жыл бұрын
Is synchronous training with 2 devices is similar to just doubling the batch size on single device?
@shaunsun9076
@shaunsun9076 4 жыл бұрын
In fact, a little confused about the pipeline. Sometimes we consider it a kind of inputting data method, but sometimes it shows a way of training acceleration in distributed parallel strategies. How to think about it? Thx
@RobertLugg
@RobertLugg 4 жыл бұрын
I like to think about the purpose: Read data from somewhere. Perhaps perform some transformation. Make the data available to keras.fit(). The "transform" is going from the original format to the correct format that the model needs. One common example is to convert RGB images, where each pixel is represented by three numbers with a range of 0 to 255. The transform might be a division by 255 which will convert each pixel to be three numbers with a range of 0 and 1. This function is specified within the pipeline (the function .map) calls. Now, when reading and making the data available there is a lot of wasted CPU time. So, instead, while the system is waiting, it does the transform for another batch. It doesn't really accelerate training. However, it makes sure that the data in process isn't delaying the training process. I think the best way to think about it is to call it the data input pipeline. Also during inference, you run data through the same input data pipeline so that the same transforms are applied.
@tanyachutani7272
@tanyachutani7272 4 жыл бұрын
Informative talk
@jijie133
@jijie133 4 жыл бұрын
Great!
Inside TensorFlow: TF-Agents
49:59
TensorFlow
Рет қаралды 20 М.
Inside TensorFlow: TensorFlow Lite
37:31
TensorFlow
Рет қаралды 22 М.
Who’s the Real Dad Doll Squid? Can You Guess in 60 Seconds? | Roblox 3D
00:34
pumpkins #shorts
00:39
Mr DegrEE
Рет қаралды 106 МЛН
How I Turned a Lolipop Into A New One 🤯🍭
00:19
Wian
Рет қаралды 10 МЛН
Ray Data Streaming for Large-Scale ML Training and Inference
30:19
Inside TensorFlow: TF Debugging
38:26
TensorFlow
Рет қаралды 15 М.
Inside TensorFlow: tf.distribute.Strategy
54:33
TensorFlow
Рет қаралды 21 М.
A friendly introduction to distributed training (ML Tech Talks)
24:19
Deploying Agentic AI to Navigate Industrial Processes: A Case Study from RHI Magnesita
26:24
Inside TensorFlow: AutoGraph
53:55
TensorFlow
Рет қаралды 10 М.
Speed up your TensorFlow code using TFRecords and dataset pipelines
26:07
Inside TensorFlow: Quantization aware training
30:35
TensorFlow
Рет қаралды 14 М.
All Rust string types explained
22:13
Let's Get Rusty
Рет қаралды 176 М.
Inside TensorFlow: tf.Keras (Part 1)
53:37
TensorFlow
Рет қаралды 50 М.
Who’s the Real Dad Doll Squid? Can You Guess in 60 Seconds? | Roblox 3D
00:34