Pyspark Advanced interview questions part 1

  Рет қаралды 59,758

TechLake

TechLake

Күн бұрын

Pyspark Advanced interview questions part 1
How to create Databricks Free Community Edition.
• Databricks Tutorial 3 ...
Complete Databricks Tutorial
• Databricks Tutorial 2 ...
Databricks Delta Lake Tutorials
• introduction To Delta ...
Pyspark Tutorials
• Pyspark Tutorial 3, fi...
Top 30 PySpark Interview Questions and Answers
PySpark Interview Questions,
PySpark Interview and Questions,
PySpark Interview Questions for freshers,
PySpark Interview Questions for experienced ,
Top 40 Apache Spark Interview Questions and Answers,
Most Common PySpark Interview Questions & Answers,
PySpark Interview Questions and Answers,
Top Apache Spark Interview Questions You Should Prepare In 2021,
Apache Spark Interview Questions And Answers,
Best PySpark Interview Questions and Answers
PySpark Interview Questions and Answers for beginners and experts. List of frequently asked PySpark Interview Questions with Answers by Besant Technologies. We hope these PySpark Interview Questions and Answers are useful and will help you to get the best job in the networking industry. This PySpark interview questions and answers are prepared by PySpark Professionals based on MNC Companies’ expectations. Stay tune we will update New PySpark Interview questions with Answers Frequently
Top 25 Pyspark Interview Questions & Answers
Top 40 Apache Spark Interview Questions and Answers in 2021
Top 10 Spark Interview Questions and Answers in 2021
Top Spark Interview Questions
Top 50 Spark Interview Questions and Answers for 2021
Best Pyspark Interview Questions and Answers
10 Essential Spark Interview Questions
Top 75 Apache Spark Interview Questions - Completely Covered With Answers
SPARK SQL PROGRAMMING INTERVIEW QUESTIONS & ANSWERS

Пікірлер: 31
@abhilash0410
@abhilash0410 3 жыл бұрын
Bro bring more real-time interview questions like these thank you so much !
@vedanthasm2659
@vedanthasm2659 3 жыл бұрын
One of the best explanation. Bro..Please make more videos on Pyspark
@sjitghosh
@sjitghosh 2 жыл бұрын
You are doing an excellent work. Helping a lot!!
@saachinileshpatil
@saachinileshpatil 7 ай бұрын
Thanks for sharing 👍, very informative
@janardhanreddy3267
@janardhanreddy3267 6 ай бұрын
nice explanation ,please attach csv file or json in description to practice
@rocku4evr
@rocku4evr 2 жыл бұрын
Great......fortunate to be your subscriber
@seshuseshu4106
@seshuseshu4106 3 жыл бұрын
Very good detailed explanation, thanks for your efforts, keep continue ..
@akashpb4044
@akashpb4044 2 жыл бұрын
Awesome video... Cleared my doubts 👍👍👍
@achintamondal1494
@achintamondal1494 Жыл бұрын
Awesome video. Could you please share the notebook, it will really help.
@nsrchndshkh
@nsrchndshkh 3 жыл бұрын
Thanks Man. This was some detailed explanation. Kudos
@TRRaveendra
@TRRaveendra 3 жыл бұрын
Ur welcome 👍
@fratkalkan7850
@fratkalkan7850 2 жыл бұрын
very clean explanation thank you sir
@rajanib9057
@rajanib9057 11 ай бұрын
can you pleaae explain how did spark filter those 2 colums as bad data? I don't see any where condition mentioned for the corrupt column
@sanooosai
@sanooosai 5 ай бұрын
great thank you
@janardhanreddy3267
@janardhanreddy3267 6 ай бұрын
please upload all pyspark interview questions videos
@balajia8376
@balajia8376 2 жыл бұрын
seems querying _corrupt_record is not working. I tried it today and not allowing me to query with the column name.cust_df.filter("_corrupt_record is not null"). AnalysisException: Since Spark 2.3, the queries from raw JSON/CSV files are disallowed when the referenced columns only include the internal corrupt record column (named _corrupt_record by default). For example: spark.read.schema(schema).csv(file).filter($"_corrupt_record".isNotNull).count() and spark.read.schema(schema).csv(file).select("_corrupt_record").show(). Instead, you can cache or save the parsed results and then send the same query. For example, val df = spark.read.schema(schema).csv(file).cache() and then df.filter($"_corrupt_record".isNotNull).count().
@TRRaveendra
@TRRaveendra 2 жыл бұрын
cust_df.cache() Cache dataframe and it's won't raise exception
@balajia8376
@balajia8376 2 жыл бұрын
@@TRRaveendra Yes I did, even after that also not allowing to write a query on _corrupt_record is null or not null.
@balajia8376
@balajia8376 2 жыл бұрын
seems badRecordsPath is only the solution.
@varuns4472
@varuns4472 2 жыл бұрын
Nice one
@balajia8376
@balajia8376 2 жыл бұрын
cust_df.select("_corrupt_record").show() is working but not allowing is null or not null. cust_df.select("_corrupt_record is null").show(). let me know if this is working for you. thank you.
@shreekrishnavani7868
@shreekrishnavani7868 2 жыл бұрын
Nice explanation 👌 thanks
@johnsonrajendran6194
@johnsonrajendran6194 3 жыл бұрын
are any such mode options available while reading parquet files?
@rahulyeole6411
@rahulyeole6411 2 жыл бұрын
Please share basic big data video
@swagatikatripathy4917
@swagatikatripathy4917 2 жыл бұрын
Why do we write inferschema= true
@TRRaveendra
@TRRaveendra 2 жыл бұрын
InferSchema =True Creating datatypes based on data. Header = True creating columns from file first line
@naveendayyala1484
@naveendayyala1484 11 ай бұрын
plz share the notebook in .dbc format
@sachintiwari6846
@sachintiwari6846 Жыл бұрын
Woah what a explanation
@srikanthbachina7764
@srikanthbachina7764 Жыл бұрын
Hi pls share ur contact details I am looking for python, pyspark, databricks training
@balajia8376
@balajia8376 2 жыл бұрын
root |-- cust_id: integer (nullable = true) |-- cust_name: string (nullable = true) |-- manager: string (nullable = true) |-- city: string (nullable = true) |-- phno: long (nullable = true) |-- _corrupt_record: string (nullable = true) . display(cust_df.filter("_corrupt_record is not null")). FileReadException: Error while reading file dbfs:/FileStore/tables/csv_with_bad_records.csv. Caused by: IllegalArgumentException: _corrupt_record does not exist. Available: cust_id, cust_name, manager, city, phno
Cute kitty gadgets 💛
00:24
TheSoul Music Family
Рет қаралды 22 МЛН
나랑 아빠가 아이스크림 먹을 때
00:15
진영민yeongmin
Рет қаралды 16 МЛН
10 recently asked Pyspark Interview Questions | Big Data Interview
28:36
Cute kitty gadgets 💛
00:24
TheSoul Music Family
Рет қаралды 22 МЛН