Skip to content

Commit 36b1f23

Browse files
committed
Merge remote-tracking branch 'origin/main' into main
2 parents 22ce8e2 + 56475a7 commit 36b1f23

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

_posts/2021-03-08-introducing-sql-delta-import.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ Yes we could have ran our sqoop jobs on EMR clusters but we wanted to run everyt
4242
avoid additional technology footprint. But even if we drop that restriction...
4343

4444
#### 2. `sqoop` does not support writing data directly to Delta Lake
45-
`scoop` can only import data as text or parquet. Writing to delta directly allows us to
45+
`sqoop` can only import data as text or parquet. Writing to delta directly allows us to
4646
optimize data storage for best performance on reads by just adding a couple of configuration options
4747

4848
```shell script
@@ -57,7 +57,7 @@ spark-submit /
5757
```
5858

5959
#### 3. `--num-mappers` just not good enough to control parallelism when working with a database
60-
`sqooop` uses map-reduce under the hood. We can specify `--num-mappers` parameter that controls how many
60+
`sqoop` uses map-reduce under the hood. We can specify `--num-mappers` parameter that controls how many
6161
mappers will be used to import data. Small number of mappers can result in large volume
6262
of data per import and long running transactions. Large number of mappers will result in many connections
6363
to database potentially overloading it especially when there are a lot of `sqoop` jobs running in parallel.

0 commit comments

Comments
 (0)