Pyspark Advanced interview questions part 1

Поділитися
Вставка
  • Опубліковано 11 вер 2024

КОМЕНТАРІ • 31

  • @abhilash0410
    @abhilash0410 3 роки тому +8

    Bro bring more real-time interview questions like these thank you so much !

  • @vedanthasm2659
    @vedanthasm2659 3 роки тому +3

    One of the best explanation. Bro..Please make more videos on Pyspark

  • @sjitghosh
    @sjitghosh 2 роки тому +3

    You are doing an excellent work. Helping a lot!!

  • @saachinileshpatil
    @saachinileshpatil 7 місяців тому +1

    Thanks for sharing 👍, very informative

  • @rocku4evr
    @rocku4evr 2 роки тому +1

    Great......fortunate to be your subscriber

  • @seshuseshu4106
    @seshuseshu4106 3 роки тому +1

    Very good detailed explanation, thanks for your efforts, keep continue ..

  • @janardhanreddy3267
    @janardhanreddy3267 6 місяців тому

    nice explanation ,please attach csv file or json in description to practice

  • @nsrchndshkh
    @nsrchndshkh 3 роки тому +1

    Thanks Man. This was some detailed explanation. Kudos

  • @akashpb4044
    @akashpb4044 2 роки тому +1

    Awesome video... Cleared my doubts 👍👍👍

  • @varuns4472
    @varuns4472 2 роки тому

    Nice one

  • @sanooosai
    @sanooosai 5 місяців тому

    great thank you

  • @achintamondal1494
    @achintamondal1494 Рік тому +1

    Awesome video.
    Could you please share the notebook, it will really help.

  • @janardhanreddy3267
    @janardhanreddy3267 6 місяців тому

    please upload all pyspark interview questions videos

  • @fratkalkan7850
    @fratkalkan7850 2 роки тому

    very clean explanation thank you sir

  • @shreekrishnavani7868
    @shreekrishnavani7868 2 роки тому

    Nice explanation 👌 thanks

  • @rajanib9057
    @rajanib9057 11 місяців тому

    can you pleaae explain how did spark filter those 2 colums as bad data? I don't see any where condition mentioned for the corrupt column

  • @rahulyeole6411
    @rahulyeole6411 2 роки тому

    Please share basic big data video

  • @johnsonrajendran6194
    @johnsonrajendran6194 3 роки тому

    are any such mode options available while reading parquet files?

  • @balajia8376
    @balajia8376 2 роки тому

    seems querying _corrupt_record is not working. I tried it today and not allowing me to query with the column name.cust_df.filter("_corrupt_record is not null"). AnalysisException: Since Spark 2.3, the queries from raw JSON/CSV files are disallowed when the
    referenced columns only include the internal corrupt record column
    (named _corrupt_record by default). For example:
    spark.read.schema(schema).csv(file).filter($"_corrupt_record".isNotNull).count()
    and spark.read.schema(schema).csv(file).select("_corrupt_record").show().
    Instead, you can cache or save the parsed results and then send the same query.
    For example, val df = spark.read.schema(schema).csv(file).cache() and then
    df.filter($"_corrupt_record".isNotNull).count().

    • @TRRaveendra
      @TRRaveendra  2 роки тому

      cust_df.cache()
      Cache dataframe and it's won't raise exception

    • @balajia8376
      @balajia8376 2 роки тому

      @@TRRaveendra Yes I did, even after that also not allowing to write a query on _corrupt_record is null or not null.

    • @balajia8376
      @balajia8376 2 роки тому

      seems badRecordsPath is only the solution.

  • @sachintiwari6846
    @sachintiwari6846 Рік тому

    Woah what a explanation

  • @balajia8376
    @balajia8376 2 роки тому

    cust_df.select("_corrupt_record").show() is working but not allowing is null or not null. cust_df.select("_corrupt_record is null").show(). let me know if this is working for you. thank you.

  • @naveendayyala1484
    @naveendayyala1484 11 місяців тому

    plz share the notebook in .dbc format

  • @swagatikatripathy4917
    @swagatikatripathy4917 2 роки тому +1

    Why do we write inferschema= true

    • @TRRaveendra
      @TRRaveendra  2 роки тому +2

      InferSchema =True Creating datatypes based on data.
      Header = True creating columns from file first line

  • @srikanthbachina7764
    @srikanthbachina7764 Рік тому

    Hi pls share ur contact details I am looking for python, pyspark, databricks training

  • @balajia8376
    @balajia8376 2 роки тому

    root
    |-- cust_id: integer (nullable = true)
    |-- cust_name: string (nullable = true)
    |-- manager: string (nullable = true)
    |-- city: string (nullable = true)
    |-- phno: long (nullable = true)
    |-- _corrupt_record: string (nullable = true) . display(cust_df.filter("_corrupt_record is not null")). FileReadException: Error while reading file dbfs:/FileStore/tables/csv_with_bad_records.csv.
    Caused by: IllegalArgumentException: _corrupt_record does not exist. Available: cust_id, cust_name, manager, city, phno